LLVM 23.0.0git
SelectionDAG.cpp
Go to the documentation of this file.
1//===- SelectionDAG.cpp - Implement the SelectionDAG data structures ------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This implements the SelectionDAG class.
10//
11//===----------------------------------------------------------------------===//
12
14#include "SDNodeDbgValue.h"
15#include "llvm/ADT/APFloat.h"
16#include "llvm/ADT/APInt.h"
17#include "llvm/ADT/APSInt.h"
18#include "llvm/ADT/ArrayRef.h"
19#include "llvm/ADT/BitVector.h"
20#include "llvm/ADT/DenseSet.h"
21#include "llvm/ADT/FoldingSet.h"
22#include "llvm/ADT/STLExtras.h"
25#include "llvm/ADT/Twine.h"
51#include "llvm/IR/Constant.h"
52#include "llvm/IR/Constants.h"
53#include "llvm/IR/DataLayout.h"
55#include "llvm/IR/DebugLoc.h"
57#include "llvm/IR/Function.h"
58#include "llvm/IR/GlobalValue.h"
59#include "llvm/IR/Metadata.h"
60#include "llvm/IR/Type.h"
64#include "llvm/Support/Debug.h"
73#include <algorithm>
74#include <cassert>
75#include <cstdint>
76#include <cstdlib>
77#include <limits>
78#include <optional>
79#include <string>
80#include <utility>
81#include <vector>
82
83using namespace llvm;
84using namespace llvm::SDPatternMatch;
85
86/// makeVTList - Return an instance of the SDVTList struct initialized with the
87/// specified members.
88static SDVTList makeVTList(const EVT *VTs, unsigned NumVTs) {
89 SDVTList Res = {VTs, NumVTs};
90 return Res;
91}
92
93// Default null implementations of the callbacks.
97
98void SelectionDAG::DAGNodeDeletedListener::anchor() {}
99void SelectionDAG::DAGNodeInsertedListener::anchor() {}
100
101#define DEBUG_TYPE "selectiondag"
102
103static cl::opt<bool> EnableMemCpyDAGOpt("enable-memcpy-dag-opt",
104 cl::Hidden, cl::init(true),
105 cl::desc("Gang up loads and stores generated by inlining of memcpy"));
106
107static cl::opt<int> MaxLdStGlue("ldstmemcpy-glue-max",
108 cl::desc("Number limit for gluing ld/st of memcpy."),
109 cl::Hidden, cl::init(0));
110
112 MaxSteps("has-predecessor-max-steps", cl::Hidden, cl::init(8192),
113 cl::desc("DAG combiner limit number of steps when searching DAG "
114 "for predecessor nodes"));
115
117 LLVM_DEBUG(dbgs() << Msg; V.getNode()->dump(G););
118}
119
121
122//===----------------------------------------------------------------------===//
123// ConstantFPSDNode Class
124//===----------------------------------------------------------------------===//
125
126/// isExactlyValue - We don't rely on operator== working on double values, as
127/// it returns true for things that are clearly not equal, like -0.0 and 0.0.
128/// As such, this method can be used to do an exact bit-for-bit comparison of
129/// two floating point values.
131 return getValueAPF().bitwiseIsEqual(V);
132}
133
135 const APFloat& Val) {
136 assert(VT.isFloatingPoint() && "Can only convert between FP types");
137
138 // convert modifies in place, so make a copy.
139 APFloat Val2 = APFloat(Val);
140 bool losesInfo;
142 &losesInfo);
143 return !losesInfo;
144}
145
146//===----------------------------------------------------------------------===//
147// ISD Namespace
148//===----------------------------------------------------------------------===//
149
150bool ISD::isConstantSplatVector(const SDNode *N, APInt &SplatVal) {
151 if (N->getOpcode() == ISD::SPLAT_VECTOR) {
152 if (auto OptAPInt = N->getOperand(0)->bitcastToAPInt()) {
153 unsigned EltSize =
154 N->getValueType(0).getVectorElementType().getSizeInBits();
155 SplatVal = OptAPInt->trunc(EltSize);
156 return true;
157 }
158 }
159
160 auto *BV = dyn_cast<BuildVectorSDNode>(N);
161 if (!BV)
162 return false;
163
164 APInt SplatUndef;
165 unsigned SplatBitSize;
166 bool HasUndefs;
167 unsigned EltSize = N->getValueType(0).getVectorElementType().getSizeInBits();
168 // Endianness does not matter here. We are checking for a splat given the
169 // element size of the vector, and if we find such a splat for little endian
170 // layout, then that should be valid also for big endian (as the full vector
171 // size is known to be a multiple of the element size).
172 const bool IsBigEndian = false;
173 return BV->isConstantSplat(SplatVal, SplatUndef, SplatBitSize, HasUndefs,
174 EltSize, IsBigEndian) &&
175 EltSize == SplatBitSize;
176}
177
178// FIXME: AllOnes and AllZeros duplicate a lot of code. Could these be
179// specializations of the more general isConstantSplatVector()?
180
181bool ISD::isConstantSplatVectorAllOnes(const SDNode *N, bool BuildVectorOnly) {
182 // Look through a bit convert.
183 while (N->getOpcode() == ISD::BITCAST)
184 N = N->getOperand(0).getNode();
185
186 if (!BuildVectorOnly && N->getOpcode() == ISD::SPLAT_VECTOR) {
187 APInt SplatVal;
188 return isConstantSplatVector(N, SplatVal) && SplatVal.isAllOnes();
189 }
190
191 if (N->getOpcode() != ISD::BUILD_VECTOR) return false;
192
193 unsigned i = 0, e = N->getNumOperands();
194
195 // Skip over all of the undef values.
196 while (i != e && N->getOperand(i).isUndef())
197 ++i;
198
199 // Do not accept an all-undef vector.
200 if (i == e) return false;
201
202 // Do not accept build_vectors that aren't all constants or which have non-~0
203 // elements. We have to be a bit careful here, as the type of the constant
204 // may not be the same as the type of the vector elements due to type
205 // legalization (the elements are promoted to a legal type for the target and
206 // a vector of a type may be legal when the base element type is not).
207 // We only want to check enough bits to cover the vector elements, because
208 // we care if the resultant vector is all ones, not whether the individual
209 // constants are.
210 SDValue NotZero = N->getOperand(i);
211 if (auto OptAPInt = NotZero->bitcastToAPInt()) {
212 unsigned EltSize = N->getValueType(0).getScalarSizeInBits();
213 if (OptAPInt->countr_one() < EltSize)
214 return false;
215 } else
216 return false;
217
218 // Okay, we have at least one ~0 value, check to see if the rest match or are
219 // undefs. Even with the above element type twiddling, this should be OK, as
220 // the same type legalization should have applied to all the elements.
221 for (++i; i != e; ++i)
222 if (N->getOperand(i) != NotZero && !N->getOperand(i).isUndef())
223 return false;
224 return true;
225}
226
227bool ISD::isConstantSplatVectorAllZeros(const SDNode *N, bool BuildVectorOnly) {
228 // Look through a bit convert.
229 while (N->getOpcode() == ISD::BITCAST)
230 N = N->getOperand(0).getNode();
231
232 if (!BuildVectorOnly && N->getOpcode() == ISD::SPLAT_VECTOR) {
233 APInt SplatVal;
234 return isConstantSplatVector(N, SplatVal) && SplatVal.isZero();
235 }
236
237 if (N->getOpcode() != ISD::BUILD_VECTOR) return false;
238
239 bool IsAllUndef = true;
240 for (const SDValue &Op : N->op_values()) {
241 if (Op.isUndef())
242 continue;
243 IsAllUndef = false;
244 // Do not accept build_vectors that aren't all constants or which have non-0
245 // elements. We have to be a bit careful here, as the type of the constant
246 // may not be the same as the type of the vector elements due to type
247 // legalization (the elements are promoted to a legal type for the target
248 // and a vector of a type may be legal when the base element type is not).
249 // We only want to check enough bits to cover the vector elements, because
250 // we care if the resultant vector is all zeros, not whether the individual
251 // constants are.
252 if (auto OptAPInt = Op->bitcastToAPInt()) {
253 unsigned EltSize = N->getValueType(0).getScalarSizeInBits();
254 if (OptAPInt->countr_zero() < EltSize)
255 return false;
256 } else
257 return false;
258 }
259
260 // Do not accept an all-undef vector.
261 if (IsAllUndef)
262 return false;
263 return true;
264}
265
267 return isConstantSplatVectorAllOnes(N, /*BuildVectorOnly*/ true);
268}
269
271 return isConstantSplatVectorAllZeros(N, /*BuildVectorOnly*/ true);
272}
273
275 if (N->getOpcode() != ISD::BUILD_VECTOR)
276 return false;
277
278 for (const SDValue &Op : N->op_values()) {
279 if (Op.isUndef())
280 continue;
282 return false;
283 }
284 return true;
285}
286
288 if (N->getOpcode() != ISD::BUILD_VECTOR)
289 return false;
290
291 for (const SDValue &Op : N->op_values()) {
292 if (Op.isUndef())
293 continue;
295 return false;
296 }
297 return true;
298}
299
300bool ISD::isVectorShrinkable(const SDNode *N, unsigned NewEltSize,
301 bool Signed) {
302 assert(N->getValueType(0).isVector() && "Expected a vector!");
303
304 unsigned EltSize = N->getValueType(0).getScalarSizeInBits();
305 if (EltSize <= NewEltSize)
306 return false;
307
308 if (N->getOpcode() == ISD::ZERO_EXTEND) {
309 return (N->getOperand(0).getValueType().getScalarSizeInBits() <=
310 NewEltSize) &&
311 !Signed;
312 }
313 if (N->getOpcode() == ISD::SIGN_EXTEND) {
314 return (N->getOperand(0).getValueType().getScalarSizeInBits() <=
315 NewEltSize) &&
316 Signed;
317 }
318 if (N->getOpcode() != ISD::BUILD_VECTOR)
319 return false;
320
321 for (const SDValue &Op : N->op_values()) {
322 if (Op.isUndef())
323 continue;
325 return false;
326
327 APInt C = Op->getAsAPIntVal().trunc(EltSize);
328 if (Signed && C.trunc(NewEltSize).sext(EltSize) != C)
329 return false;
330 if (!Signed && C.trunc(NewEltSize).zext(EltSize) != C)
331 return false;
332 }
333
334 return true;
335}
336
338 // Return false if the node has no operands.
339 // This is "logically inconsistent" with the definition of "all" but
340 // is probably the desired behavior.
341 if (N->getNumOperands() == 0)
342 return false;
343 return all_of(N->op_values(), [](SDValue Op) { return Op.isUndef(); });
344}
345
347 return N->getOpcode() == ISD::FREEZE && N->getOperand(0).isUndef();
348}
349
350template <typename ConstNodeType>
352 std::function<bool(ConstNodeType *)> Match,
353 bool AllowUndefs, bool AllowTruncation) {
354 // FIXME: Add support for scalar UNDEF cases?
355 if (auto *C = dyn_cast<ConstNodeType>(Op))
356 return Match(C);
357
358 // FIXME: Add support for vector UNDEF cases?
359 if (ISD::BUILD_VECTOR != Op.getOpcode() &&
360 ISD::SPLAT_VECTOR != Op.getOpcode())
361 return false;
362
363 EVT SVT = Op.getValueType().getScalarType();
364 for (unsigned i = 0, e = Op.getNumOperands(); i != e; ++i) {
365 if (AllowUndefs && Op.getOperand(i).isUndef()) {
366 if (!Match(nullptr))
367 return false;
368 continue;
369 }
370
371 auto *Cst = dyn_cast<ConstNodeType>(Op.getOperand(i));
372 if (!Cst || (!AllowTruncation && Cst->getValueType(0) != SVT) ||
373 !Match(Cst))
374 return false;
375 }
376 return true;
377}
378// Build used template types.
380 SDValue, std::function<bool(ConstantSDNode *)>, bool, bool);
382 SDValue, std::function<bool(ConstantFPSDNode *)>, bool, bool);
383
385 SDValue LHS, SDValue RHS,
386 std::function<bool(ConstantSDNode *, ConstantSDNode *)> Match,
387 bool AllowUndefs, bool AllowTypeMismatch) {
388 if (!AllowTypeMismatch && LHS.getValueType() != RHS.getValueType())
389 return false;
390
391 // TODO: Add support for scalar UNDEF cases?
392 if (auto *LHSCst = dyn_cast<ConstantSDNode>(LHS))
393 if (auto *RHSCst = dyn_cast<ConstantSDNode>(RHS))
394 return Match(LHSCst, RHSCst);
395
396 // TODO: Add support for vector UNDEF cases?
397 if (LHS.getOpcode() != RHS.getOpcode() ||
398 (LHS.getOpcode() != ISD::BUILD_VECTOR &&
399 LHS.getOpcode() != ISD::SPLAT_VECTOR))
400 return false;
401
402 EVT SVT = LHS.getValueType().getScalarType();
403 for (unsigned i = 0, e = LHS.getNumOperands(); i != e; ++i) {
404 SDValue LHSOp = LHS.getOperand(i);
405 SDValue RHSOp = RHS.getOperand(i);
406 bool LHSUndef = AllowUndefs && LHSOp.isUndef();
407 bool RHSUndef = AllowUndefs && RHSOp.isUndef();
408 auto *LHSCst = dyn_cast<ConstantSDNode>(LHSOp);
409 auto *RHSCst = dyn_cast<ConstantSDNode>(RHSOp);
410 if ((!LHSCst && !LHSUndef) || (!RHSCst && !RHSUndef))
411 return false;
412 if (!AllowTypeMismatch && (LHSOp.getValueType() != SVT ||
413 LHSOp.getValueType() != RHSOp.getValueType()))
414 return false;
415 if (!Match(LHSCst, RHSCst))
416 return false;
417 }
418 return true;
419}
420
422 switch (MinMaxOpc) {
423 default:
424 llvm_unreachable("unrecognized opcode");
425 case ISD::UMIN:
426 return ISD::UMAX;
427 case ISD::UMAX:
428 return ISD::UMIN;
429 case ISD::SMIN:
430 return ISD::SMAX;
431 case ISD::SMAX:
432 return ISD::SMIN;
433 }
434}
435
437 switch (MinMaxOpc) {
438 default:
439 llvm_unreachable("unrecognized min/max opcode");
440 case ISD::SMIN:
441 return ISD::UMIN;
442 case ISD::SMAX:
443 return ISD::UMAX;
444 case ISD::UMIN:
445 return ISD::SMIN;
446 case ISD::UMAX:
447 return ISD::SMAX;
448 }
449}
450
452 switch (VecReduceOpcode) {
453 default:
454 llvm_unreachable("Expected VECREDUCE opcode");
457 case ISD::VP_REDUCE_FADD:
458 case ISD::VP_REDUCE_SEQ_FADD:
459 return ISD::FADD;
462 case ISD::VP_REDUCE_FMUL:
463 case ISD::VP_REDUCE_SEQ_FMUL:
464 return ISD::FMUL;
466 case ISD::VP_REDUCE_ADD:
467 return ISD::ADD;
469 case ISD::VP_REDUCE_MUL:
470 return ISD::MUL;
472 case ISD::VP_REDUCE_AND:
473 return ISD::AND;
475 case ISD::VP_REDUCE_OR:
476 return ISD::OR;
478 case ISD::VP_REDUCE_XOR:
479 return ISD::XOR;
481 case ISD::VP_REDUCE_SMAX:
482 return ISD::SMAX;
484 case ISD::VP_REDUCE_SMIN:
485 return ISD::SMIN;
487 case ISD::VP_REDUCE_UMAX:
488 return ISD::UMAX;
490 case ISD::VP_REDUCE_UMIN:
491 return ISD::UMIN;
493 case ISD::VP_REDUCE_FMAX:
494 return ISD::FMAXNUM;
496 case ISD::VP_REDUCE_FMIN:
497 return ISD::FMINNUM;
499 case ISD::VP_REDUCE_FMAXIMUM:
500 return ISD::FMAXIMUM;
502 case ISD::VP_REDUCE_FMINIMUM:
503 return ISD::FMINIMUM;
504 }
505}
506
507bool ISD::isVPOpcode(unsigned Opcode) {
508 switch (Opcode) {
509 default:
510 return false;
511#define BEGIN_REGISTER_VP_SDNODE(VPSD, ...) \
512 case ISD::VPSD: \
513 return true;
514#include "llvm/IR/VPIntrinsics.def"
515 }
516}
517
518bool ISD::isVPBinaryOp(unsigned Opcode) {
519 switch (Opcode) {
520 default:
521 break;
522#define BEGIN_REGISTER_VP_SDNODE(VPSD, ...) case ISD::VPSD:
523#define VP_PROPERTY_BINARYOP return true;
524#define END_REGISTER_VP_SDNODE(VPSD) break;
525#include "llvm/IR/VPIntrinsics.def"
526 }
527 return false;
528}
529
530bool ISD::isVPReduction(unsigned Opcode) {
531 switch (Opcode) {
532 default:
533 return false;
534 case ISD::VP_REDUCE_ADD:
535 case ISD::VP_REDUCE_MUL:
536 case ISD::VP_REDUCE_AND:
537 case ISD::VP_REDUCE_OR:
538 case ISD::VP_REDUCE_XOR:
539 case ISD::VP_REDUCE_SMAX:
540 case ISD::VP_REDUCE_SMIN:
541 case ISD::VP_REDUCE_UMAX:
542 case ISD::VP_REDUCE_UMIN:
543 case ISD::VP_REDUCE_FMAX:
544 case ISD::VP_REDUCE_FMIN:
545 case ISD::VP_REDUCE_FMAXIMUM:
546 case ISD::VP_REDUCE_FMINIMUM:
547 case ISD::VP_REDUCE_FADD:
548 case ISD::VP_REDUCE_FMUL:
549 case ISD::VP_REDUCE_SEQ_FADD:
550 case ISD::VP_REDUCE_SEQ_FMUL:
551 return true;
552 }
553}
554
555/// The operand position of the vector mask.
556std::optional<unsigned> ISD::getVPMaskIdx(unsigned Opcode) {
557 switch (Opcode) {
558 default:
559 return std::nullopt;
560#define BEGIN_REGISTER_VP_SDNODE(VPSD, LEGALPOS, TDNAME, MASKPOS, ...) \
561 case ISD::VPSD: \
562 return MASKPOS;
563#include "llvm/IR/VPIntrinsics.def"
564 }
565}
566
567/// The operand position of the explicit vector length parameter.
568std::optional<unsigned> ISD::getVPExplicitVectorLengthIdx(unsigned Opcode) {
569 switch (Opcode) {
570 default:
571 return std::nullopt;
572#define BEGIN_REGISTER_VP_SDNODE(VPSD, LEGALPOS, TDNAME, MASKPOS, EVLPOS) \
573 case ISD::VPSD: \
574 return EVLPOS;
575#include "llvm/IR/VPIntrinsics.def"
576 }
577}
578
579std::optional<unsigned> ISD::getBaseOpcodeForVP(unsigned VPOpcode,
580 bool hasFPExcept) {
581 // FIXME: Return strict opcodes in case of fp exceptions.
582 switch (VPOpcode) {
583 default:
584 return std::nullopt;
585#define BEGIN_REGISTER_VP_SDNODE(VPOPC, ...) case ISD::VPOPC:
586#define VP_PROPERTY_FUNCTIONAL_SDOPC(SDOPC) return ISD::SDOPC;
587#define END_REGISTER_VP_SDNODE(VPOPC) break;
588#include "llvm/IR/VPIntrinsics.def"
589 }
590 return std::nullopt;
591}
592
593std::optional<unsigned> ISD::getVPForBaseOpcode(unsigned Opcode) {
594 switch (Opcode) {
595 default:
596 return std::nullopt;
597#define BEGIN_REGISTER_VP_SDNODE(VPOPC, ...) break;
598#define VP_PROPERTY_FUNCTIONAL_SDOPC(SDOPC) case ISD::SDOPC:
599#define END_REGISTER_VP_SDNODE(VPOPC) return ISD::VPOPC;
600#include "llvm/IR/VPIntrinsics.def"
601 }
602}
603
605 switch (ExtType) {
606 case ISD::EXTLOAD:
607 return IsFP ? ISD::FP_EXTEND : ISD::ANY_EXTEND;
608 case ISD::SEXTLOAD:
609 return ISD::SIGN_EXTEND;
610 case ISD::ZEXTLOAD:
611 return ISD::ZERO_EXTEND;
612 default:
613 break;
614 }
615
616 llvm_unreachable("Invalid LoadExtType");
617}
618
620 // To perform this operation, we just need to swap the L and G bits of the
621 // operation.
622 unsigned OldL = (Operation >> 2) & 1;
623 unsigned OldG = (Operation >> 1) & 1;
624 return ISD::CondCode((Operation & ~6) | // Keep the N, U, E bits
625 (OldL << 1) | // New G bit
626 (OldG << 2)); // New L bit.
627}
628
630 unsigned Operation = Op;
631 if (isIntegerLike)
632 Operation ^= 7; // Flip L, G, E bits, but not U.
633 else
634 Operation ^= 15; // Flip all of the condition bits.
635
637 Operation &= ~8; // Don't let N and U bits get set.
638
639 return ISD::CondCode(Operation);
640}
641
645
647 bool isIntegerLike) {
648 return getSetCCInverseImpl(Op, isIntegerLike);
649}
650
651/// For an integer comparison, return 1 if the comparison is a signed operation
652/// and 2 if the result is an unsigned comparison. Return zero if the operation
653/// does not depend on the sign of the input (setne and seteq).
654static int isSignedOp(ISD::CondCode Opcode) {
655 switch (Opcode) {
656 default: llvm_unreachable("Illegal integer setcc operation!");
657 case ISD::SETEQ:
658 case ISD::SETNE: return 0;
659 case ISD::SETLT:
660 case ISD::SETLE:
661 case ISD::SETGT:
662 case ISD::SETGE: return 1;
663 case ISD::SETULT:
664 case ISD::SETULE:
665 case ISD::SETUGT:
666 case ISD::SETUGE: return 2;
667 }
668}
669
671 EVT Type) {
672 bool IsInteger = Type.isInteger();
673 if (IsInteger && (isSignedOp(Op1) | isSignedOp(Op2)) == 3)
674 // Cannot fold a signed integer setcc with an unsigned integer setcc.
675 return ISD::SETCC_INVALID;
676
677 unsigned Op = Op1 | Op2; // Combine all of the condition bits.
678
679 // If the N and U bits get set, then the resultant comparison DOES suddenly
680 // care about orderedness, and it is true when ordered.
681 if (Op > ISD::SETTRUE2)
682 Op &= ~16; // Clear the U bit if the N bit is set.
683
684 // Canonicalize illegal integer setcc's.
685 if (IsInteger && Op == ISD::SETUNE) // e.g. SETUGT | SETULT
686 Op = ISD::SETNE;
687
688 return ISD::CondCode(Op);
689}
690
692 EVT Type) {
693 bool IsInteger = Type.isInteger();
694 if (IsInteger && (isSignedOp(Op1) | isSignedOp(Op2)) == 3)
695 // Cannot fold a signed setcc with an unsigned setcc.
696 return ISD::SETCC_INVALID;
697
698 // Combine all of the condition bits.
699 ISD::CondCode Result = ISD::CondCode(Op1 & Op2);
700
701 // Canonicalize illegal integer setcc's.
702 if (IsInteger) {
703 switch (Result) {
704 default: break;
705 case ISD::SETUO : Result = ISD::SETFALSE; break; // SETUGT & SETULT
706 case ISD::SETOEQ: // SETEQ & SETU[LG]E
707 case ISD::SETUEQ: Result = ISD::SETEQ ; break; // SETUGE & SETULE
708 case ISD::SETOLT: Result = ISD::SETULT ; break; // SETULT & SETNE
709 case ISD::SETOGT: Result = ISD::SETUGT ; break; // SETUGT & SETNE
710 }
711 }
712
713 return Result;
714}
715
716//===----------------------------------------------------------------------===//
717// SDNode Profile Support
718//===----------------------------------------------------------------------===//
719
720/// AddNodeIDOpcode - Add the node opcode to the NodeID data.
721static void AddNodeIDOpcode(FoldingSetNodeID &ID, unsigned OpC) {
722 ID.AddInteger(OpC);
723}
724
725/// AddNodeIDValueTypes - Value type lists are intern'd so we can represent them
726/// solely with their pointer.
728 ID.AddPointer(VTList.VTs);
729}
730
731/// AddNodeIDOperands - Various routines for adding operands to the NodeID data.
734 for (const auto &Op : Ops) {
735 ID.AddPointer(Op.getNode());
736 ID.AddInteger(Op.getResNo());
737 }
738}
739
740/// AddNodeIDOperands - Various routines for adding operands to the NodeID data.
743 for (const auto &Op : Ops) {
744 ID.AddPointer(Op.getNode());
745 ID.AddInteger(Op.getResNo());
746 }
747}
748
749static void AddNodeIDNode(FoldingSetNodeID &ID, unsigned OpC,
750 SDVTList VTList, ArrayRef<SDValue> OpList) {
751 AddNodeIDOpcode(ID, OpC);
752 AddNodeIDValueTypes(ID, VTList);
753 AddNodeIDOperands(ID, OpList);
754}
755
756/// If this is an SDNode with special info, add this info to the NodeID data.
758 switch (N->getOpcode()) {
761 case ISD::MCSymbol:
762 llvm_unreachable("Should only be used on nodes with operands");
763 default: break; // Normal nodes don't need extra info.
765 case ISD::Constant: {
767 ID.AddPointer(C->getConstantIntValue());
768 ID.AddBoolean(C->isOpaque());
769 break;
770 }
772 case ISD::ConstantFP:
773 ID.AddPointer(cast<ConstantFPSDNode>(N)->getConstantFPValue());
774 break;
780 ID.AddPointer(GA->getGlobal());
781 ID.AddInteger(GA->getOffset());
782 ID.AddInteger(GA->getTargetFlags());
783 break;
784 }
785 case ISD::BasicBlock:
787 break;
788 case ISD::Register:
789 ID.AddInteger(cast<RegisterSDNode>(N)->getReg().id());
790 break;
792 ID.AddPointer(cast<RegisterMaskSDNode>(N)->getRegMask());
793 break;
794 case ISD::SRCVALUE:
795 ID.AddPointer(cast<SrcValueSDNode>(N)->getValue());
796 break;
797 case ISD::FrameIndex:
799 ID.AddInteger(cast<FrameIndexSDNode>(N)->getIndex());
800 break;
802 ID.AddInteger(cast<PseudoProbeSDNode>(N)->getGuid());
803 ID.AddInteger(cast<PseudoProbeSDNode>(N)->getIndex());
804 ID.AddInteger(cast<PseudoProbeSDNode>(N)->getAttributes());
805 break;
806 case ISD::JumpTable:
808 ID.AddInteger(cast<JumpTableSDNode>(N)->getIndex());
809 ID.AddInteger(cast<JumpTableSDNode>(N)->getTargetFlags());
810 break;
814 ID.AddInteger(CP->getAlign().value());
815 ID.AddInteger(CP->getOffset());
818 else
819 ID.AddPointer(CP->getConstVal());
820 ID.AddInteger(CP->getTargetFlags());
821 break;
822 }
823 case ISD::TargetIndex: {
825 ID.AddInteger(TI->getIndex());
826 ID.AddInteger(TI->getOffset());
827 ID.AddInteger(TI->getTargetFlags());
828 break;
829 }
830 case ISD::LOAD: {
831 const LoadSDNode *LD = cast<LoadSDNode>(N);
832 ID.AddInteger(LD->getMemoryVT().getRawBits());
833 ID.AddInteger(LD->getRawSubclassData());
834 ID.AddInteger(LD->getPointerInfo().getAddrSpace());
835 ID.AddInteger(LD->getMemOperand()->getFlags());
836 break;
837 }
838 case ISD::STORE: {
839 const StoreSDNode *ST = cast<StoreSDNode>(N);
840 ID.AddInteger(ST->getMemoryVT().getRawBits());
841 ID.AddInteger(ST->getRawSubclassData());
842 ID.AddInteger(ST->getPointerInfo().getAddrSpace());
843 ID.AddInteger(ST->getMemOperand()->getFlags());
844 break;
845 }
846 case ISD::VP_LOAD: {
847 const VPLoadSDNode *ELD = cast<VPLoadSDNode>(N);
848 ID.AddInteger(ELD->getMemoryVT().getRawBits());
849 ID.AddInteger(ELD->getRawSubclassData());
850 ID.AddInteger(ELD->getPointerInfo().getAddrSpace());
851 ID.AddInteger(ELD->getMemOperand()->getFlags());
852 break;
853 }
854 case ISD::VP_LOAD_FF: {
855 const auto *LD = cast<VPLoadFFSDNode>(N);
856 ID.AddInteger(LD->getMemoryVT().getRawBits());
857 ID.AddInteger(LD->getRawSubclassData());
858 ID.AddInteger(LD->getPointerInfo().getAddrSpace());
859 ID.AddInteger(LD->getMemOperand()->getFlags());
860 break;
861 }
862 case ISD::VP_STORE: {
863 const VPStoreSDNode *EST = cast<VPStoreSDNode>(N);
864 ID.AddInteger(EST->getMemoryVT().getRawBits());
865 ID.AddInteger(EST->getRawSubclassData());
866 ID.AddInteger(EST->getPointerInfo().getAddrSpace());
867 ID.AddInteger(EST->getMemOperand()->getFlags());
868 break;
869 }
870 case ISD::EXPERIMENTAL_VP_STRIDED_LOAD: {
872 ID.AddInteger(SLD->getMemoryVT().getRawBits());
873 ID.AddInteger(SLD->getRawSubclassData());
874 ID.AddInteger(SLD->getPointerInfo().getAddrSpace());
875 break;
876 }
877 case ISD::EXPERIMENTAL_VP_STRIDED_STORE: {
879 ID.AddInteger(SST->getMemoryVT().getRawBits());
880 ID.AddInteger(SST->getRawSubclassData());
881 ID.AddInteger(SST->getPointerInfo().getAddrSpace());
882 break;
883 }
884 case ISD::VP_GATHER: {
886 ID.AddInteger(EG->getMemoryVT().getRawBits());
887 ID.AddInteger(EG->getRawSubclassData());
888 ID.AddInteger(EG->getPointerInfo().getAddrSpace());
889 ID.AddInteger(EG->getMemOperand()->getFlags());
890 break;
891 }
892 case ISD::VP_SCATTER: {
894 ID.AddInteger(ES->getMemoryVT().getRawBits());
895 ID.AddInteger(ES->getRawSubclassData());
896 ID.AddInteger(ES->getPointerInfo().getAddrSpace());
897 ID.AddInteger(ES->getMemOperand()->getFlags());
898 break;
899 }
900 case ISD::MLOAD: {
902 ID.AddInteger(MLD->getMemoryVT().getRawBits());
903 ID.AddInteger(MLD->getRawSubclassData());
904 ID.AddInteger(MLD->getPointerInfo().getAddrSpace());
905 ID.AddInteger(MLD->getMemOperand()->getFlags());
906 break;
907 }
908 case ISD::MSTORE: {
910 ID.AddInteger(MST->getMemoryVT().getRawBits());
911 ID.AddInteger(MST->getRawSubclassData());
912 ID.AddInteger(MST->getPointerInfo().getAddrSpace());
913 ID.AddInteger(MST->getMemOperand()->getFlags());
914 break;
915 }
916 case ISD::MGATHER: {
918 ID.AddInteger(MG->getMemoryVT().getRawBits());
919 ID.AddInteger(MG->getRawSubclassData());
920 ID.AddInteger(MG->getPointerInfo().getAddrSpace());
921 ID.AddInteger(MG->getMemOperand()->getFlags());
922 break;
923 }
924 case ISD::MSCATTER: {
926 ID.AddInteger(MS->getMemoryVT().getRawBits());
927 ID.AddInteger(MS->getRawSubclassData());
928 ID.AddInteger(MS->getPointerInfo().getAddrSpace());
929 ID.AddInteger(MS->getMemOperand()->getFlags());
930 break;
931 }
934 case ISD::ATOMIC_SWAP:
946 case ISD::ATOMIC_LOAD:
947 case ISD::ATOMIC_STORE: {
948 const AtomicSDNode *AT = cast<AtomicSDNode>(N);
949 ID.AddInteger(AT->getMemoryVT().getRawBits());
950 ID.AddInteger(AT->getRawSubclassData());
951 ID.AddInteger(AT->getPointerInfo().getAddrSpace());
952 ID.AddInteger(AT->getMemOperand()->getFlags());
953 break;
954 }
955 case ISD::VECTOR_SHUFFLE: {
956 ArrayRef<int> Mask = cast<ShuffleVectorSDNode>(N)->getMask();
957 for (int M : Mask)
958 ID.AddInteger(M);
959 break;
960 }
961 case ISD::ADDRSPACECAST: {
963 ID.AddInteger(ASC->getSrcAddressSpace());
964 ID.AddInteger(ASC->getDestAddressSpace());
965 break;
966 }
968 case ISD::BlockAddress: {
970 ID.AddPointer(BA->getBlockAddress());
971 ID.AddInteger(BA->getOffset());
972 ID.AddInteger(BA->getTargetFlags());
973 break;
974 }
975 case ISD::AssertAlign:
976 ID.AddInteger(cast<AssertAlignSDNode>(N)->getAlign().value());
977 break;
978 case ISD::PREFETCH:
981 // Handled by MemIntrinsicSDNode check after the switch.
982 break;
984 ID.AddPointer(cast<MDNodeSDNode>(N)->getMD());
985 break;
986 } // end switch (N->getOpcode())
987
988 // MemIntrinsic nodes could also have subclass data, address spaces, and flags
989 // to check.
990 if (auto *MN = dyn_cast<MemIntrinsicSDNode>(N)) {
991 ID.AddInteger(MN->getRawSubclassData());
992 ID.AddInteger(MN->getMemoryVT().getRawBits());
993 for (const MachineMemOperand *MMO : MN->memoperands()) {
994 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
995 ID.AddInteger(MMO->getFlags());
996 }
997 }
998}
999
1000/// AddNodeIDNode - Generic routine for adding a nodes info to the NodeID
1001/// data.
1003 AddNodeIDOpcode(ID, N->getOpcode());
1004 // Add the return value info.
1005 AddNodeIDValueTypes(ID, N->getVTList());
1006 // Add the operand info.
1007 AddNodeIDOperands(ID, N->ops());
1008
1009 // Handle SDNode leafs with special info.
1011}
1012
1013//===----------------------------------------------------------------------===//
1014// SelectionDAG Class
1015//===----------------------------------------------------------------------===//
1016
1017/// doNotCSE - Return true if CSE should not be performed for this node.
1018static bool doNotCSE(SDNode *N) {
1019 if (N->getValueType(0) == MVT::Glue)
1020 return true; // Never CSE anything that produces a glue result.
1021
1022 switch (N->getOpcode()) {
1023 default: break;
1024 case ISD::HANDLENODE:
1025 case ISD::EH_LABEL:
1026 return true; // Never CSE these nodes.
1027 }
1028
1029 // Check that remaining values produced are not flags.
1030 for (unsigned i = 1, e = N->getNumValues(); i != e; ++i)
1031 if (N->getValueType(i) == MVT::Glue)
1032 return true; // Never CSE anything that produces a glue result.
1033
1034 return false;
1035}
1036
1037/// RemoveDeadNodes - This method deletes all unreachable nodes in the
1038/// SelectionDAG.
1040 // Create a dummy node (which is not added to allnodes), that adds a reference
1041 // to the root node, preventing it from being deleted.
1042 HandleSDNode Dummy(getRoot());
1043
1044 SmallVector<SDNode*, 128> DeadNodes;
1045
1046 // Add all obviously-dead nodes to the DeadNodes worklist.
1047 for (SDNode &Node : allnodes())
1048 if (Node.use_empty())
1049 DeadNodes.push_back(&Node);
1050
1051 RemoveDeadNodes(DeadNodes);
1052
1053 // If the root changed (e.g. it was a dead load, update the root).
1054 setRoot(Dummy.getValue());
1055}
1056
1057/// RemoveDeadNodes - This method deletes the unreachable nodes in the
1058/// given list, and any nodes that become unreachable as a result.
1060
1061 // Process the worklist, deleting the nodes and adding their uses to the
1062 // worklist.
1063 while (!DeadNodes.empty()) {
1064 SDNode *N = DeadNodes.pop_back_val();
1065 // Skip to next node if we've already managed to delete the node. This could
1066 // happen if replacing a node causes a node previously added to the node to
1067 // be deleted.
1068 if (N->getOpcode() == ISD::DELETED_NODE)
1069 continue;
1070
1071 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1072 DUL->NodeDeleted(N, nullptr);
1073
1074 // Take the node out of the appropriate CSE map.
1075 RemoveNodeFromCSEMaps(N);
1076
1077 // Next, brutally remove the operand list. This is safe to do, as there are
1078 // no cycles in the graph.
1079 for (SDNode::op_iterator I = N->op_begin(), E = N->op_end(); I != E; ) {
1080 SDUse &Use = *I++;
1081 SDNode *Operand = Use.getNode();
1082 Use.set(SDValue());
1083
1084 // Now that we removed this operand, see if there are no uses of it left.
1085 if (Operand->use_empty())
1086 DeadNodes.push_back(Operand);
1087 }
1088
1089 DeallocateNode(N);
1090 }
1091}
1092
1094 SmallVector<SDNode*, 16> DeadNodes(1, N);
1095
1096 // Create a dummy node that adds a reference to the root node, preventing
1097 // it from being deleted. (This matters if the root is an operand of the
1098 // dead node.)
1099 HandleSDNode Dummy(getRoot());
1100
1101 RemoveDeadNodes(DeadNodes);
1102}
1103
1105 // First take this out of the appropriate CSE map.
1106 RemoveNodeFromCSEMaps(N);
1107
1108 // Finally, remove uses due to operands of this node, remove from the
1109 // AllNodes list, and delete the node.
1110 DeleteNodeNotInCSEMaps(N);
1111}
1112
1113void SelectionDAG::DeleteNodeNotInCSEMaps(SDNode *N) {
1114 assert(N->getIterator() != AllNodes.begin() &&
1115 "Cannot delete the entry node!");
1116 assert(N->use_empty() && "Cannot delete a node that is not dead!");
1117
1118 // Drop all of the operands and decrement used node's use counts.
1119 N->DropOperands();
1120
1121 DeallocateNode(N);
1122}
1123
1124void SDDbgInfo::add(SDDbgValue *V, bool isParameter) {
1125 assert(!(V->isVariadic() && isParameter));
1126 if (isParameter)
1127 ByvalParmDbgValues.push_back(V);
1128 else
1129 DbgValues.push_back(V);
1130 for (const SDNode *Node : V->getSDNodes())
1131 if (Node)
1132 DbgValMap[Node].push_back(V);
1133}
1134
1136 DbgValMapType::iterator I = DbgValMap.find(Node);
1137 if (I == DbgValMap.end())
1138 return;
1139 for (auto &Val: I->second)
1140 Val->setIsInvalidated();
1141 DbgValMap.erase(I);
1142}
1143
1144void SelectionDAG::DeallocateNode(SDNode *N) {
1145 // If we have operands, deallocate them.
1147
1148 NodeAllocator.Deallocate(AllNodes.remove(N));
1149
1150 // Set the opcode to DELETED_NODE to help catch bugs when node
1151 // memory is reallocated.
1152 // FIXME: There are places in SDag that have grown a dependency on the opcode
1153 // value in the released node.
1154 __asan_unpoison_memory_region(&N->NodeType, sizeof(N->NodeType));
1155 N->NodeType = ISD::DELETED_NODE;
1156
1157 // If any of the SDDbgValue nodes refer to this SDNode, invalidate
1158 // them and forget about that node.
1159 DbgInfo->erase(N);
1160
1161 // Invalidate extra info.
1162 SDEI.erase(N);
1163}
1164
1165#ifndef NDEBUG
1166/// VerifySDNode - Check the given SDNode. Aborts if it is invalid.
1167void SelectionDAG::verifyNode(SDNode *N) const {
1168 switch (N->getOpcode()) {
1169 default:
1170 if (N->isTargetOpcode())
1172 break;
1173 case ISD::BUILD_PAIR: {
1174 EVT VT = N->getValueType(0);
1175 assert(N->getNumValues() == 1 && "Too many results!");
1176 assert(!VT.isVector() && (VT.isInteger() || VT.isFloatingPoint()) &&
1177 "Wrong return type!");
1178 assert(N->getNumOperands() == 2 && "Wrong number of operands!");
1179 assert(N->getOperand(0).getValueType() == N->getOperand(1).getValueType() &&
1180 "Mismatched operand types!");
1181 assert(N->getOperand(0).getValueType().isInteger() == VT.isInteger() &&
1182 "Wrong operand type!");
1183 assert(VT.getSizeInBits() == 2 * N->getOperand(0).getValueSizeInBits() &&
1184 "Wrong return type size");
1185 break;
1186 }
1187 case ISD::BUILD_VECTOR: {
1188 assert(N->getNumValues() == 1 && "Too many results!");
1189 assert(N->getValueType(0).isVector() && "Wrong return type!");
1190 assert(N->getNumOperands() == N->getValueType(0).getVectorNumElements() &&
1191 "Wrong number of operands!");
1192 EVT EltVT = N->getValueType(0).getVectorElementType();
1193 for (const SDUse &Op : N->ops()) {
1194 assert((Op.getValueType() == EltVT ||
1195 (EltVT.isInteger() && Op.getValueType().isInteger() &&
1196 EltVT.bitsLE(Op.getValueType()))) &&
1197 "Wrong operand type!");
1198 assert(Op.getValueType() == N->getOperand(0).getValueType() &&
1199 "Operands must all have the same type");
1200 }
1201 break;
1202 }
1203 case ISD::SADDO:
1204 case ISD::UADDO:
1205 case ISD::SSUBO:
1206 case ISD::USUBO:
1207 assert(N->getNumValues() == 2 && "Wrong number of results!");
1208 assert(N->getVTList().NumVTs == 2 && N->getNumOperands() == 2 &&
1209 "Invalid add/sub overflow op!");
1210 assert(N->getVTList().VTs[0].isInteger() &&
1211 N->getVTList().VTs[1].isInteger() &&
1212 N->getOperand(0).getValueType() == N->getOperand(1).getValueType() &&
1213 N->getOperand(0).getValueType() == N->getVTList().VTs[0] &&
1214 "Binary operator types must match!");
1215 break;
1216 }
1217}
1218#endif // NDEBUG
1219
1220/// Insert a newly allocated node into the DAG.
1221///
1222/// Handles insertion into the all nodes list and CSE map, as well as
1223/// verification and other common operations when a new node is allocated.
1224void SelectionDAG::InsertNode(SDNode *N) {
1225 AllNodes.push_back(N);
1226#ifndef NDEBUG
1227 N->PersistentId = NextPersistentId++;
1228 verifyNode(N);
1229#endif
1230 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1231 DUL->NodeInserted(N);
1232}
1233
1234/// RemoveNodeFromCSEMaps - Take the specified node out of the CSE map that
1235/// correspond to it. This is useful when we're about to delete or repurpose
1236/// the node. We don't want future request for structurally identical nodes
1237/// to return N anymore.
1238bool SelectionDAG::RemoveNodeFromCSEMaps(SDNode *N) {
1239 bool Erased = false;
1240 switch (N->getOpcode()) {
1241 case ISD::HANDLENODE: return false; // noop.
1242 case ISD::CONDCODE:
1243 assert(CondCodeNodes[cast<CondCodeSDNode>(N)->get()] &&
1244 "Cond code doesn't exist!");
1245 Erased = CondCodeNodes[cast<CondCodeSDNode>(N)->get()] != nullptr;
1246 CondCodeNodes[cast<CondCodeSDNode>(N)->get()] = nullptr;
1247 break;
1249 Erased = ExternalSymbols.erase(cast<ExternalSymbolSDNode>(N)->getSymbol());
1250 break;
1252 ExternalSymbolSDNode *ESN = cast<ExternalSymbolSDNode>(N);
1253 Erased = TargetExternalSymbols.erase(std::pair<std::string, unsigned>(
1254 ESN->getSymbol(), ESN->getTargetFlags()));
1255 break;
1256 }
1257 case ISD::MCSymbol: {
1258 auto *MCSN = cast<MCSymbolSDNode>(N);
1259 Erased = MCSymbols.erase(MCSN->getMCSymbol());
1260 break;
1261 }
1262 case ISD::VALUETYPE: {
1263 EVT VT = cast<VTSDNode>(N)->getVT();
1264 if (VT.isExtended()) {
1265 Erased = ExtendedValueTypeNodes.erase(VT);
1266 } else {
1267 Erased = ValueTypeNodes[VT.getSimpleVT().SimpleTy] != nullptr;
1268 ValueTypeNodes[VT.getSimpleVT().SimpleTy] = nullptr;
1269 }
1270 break;
1271 }
1272 default:
1273 // Remove it from the CSE Map.
1274 assert(N->getOpcode() != ISD::DELETED_NODE && "DELETED_NODE in CSEMap!");
1275 assert(N->getOpcode() != ISD::EntryToken && "EntryToken in CSEMap!");
1276 Erased = CSEMap.RemoveNode(N);
1277 break;
1278 }
1279#ifndef NDEBUG
1280 // Verify that the node was actually in one of the CSE maps, unless it has a
1281 // glue result (which cannot be CSE'd) or is one of the special cases that are
1282 // not subject to CSE.
1283 if (!Erased && N->getValueType(N->getNumValues()-1) != MVT::Glue &&
1284 !N->isMachineOpcode() && !doNotCSE(N)) {
1285 N->dump(this);
1286 dbgs() << "\n";
1287 llvm_unreachable("Node is not in map!");
1288 }
1289#endif
1290 return Erased;
1291}
1292
1293/// AddModifiedNodeToCSEMaps - The specified node has been removed from the CSE
1294/// maps and modified in place. Add it back to the CSE maps, unless an identical
1295/// node already exists, in which case transfer all its users to the existing
1296/// node. This transfer can potentially trigger recursive merging.
1297void
1298SelectionDAG::AddModifiedNodeToCSEMaps(SDNode *N) {
1299 // For node types that aren't CSE'd, just act as if no identical node
1300 // already exists.
1301 if (!doNotCSE(N)) {
1302 SDNode *Existing = CSEMap.GetOrInsertNode(N);
1303 if (Existing != N) {
1304 // If there was already an existing matching node, use ReplaceAllUsesWith
1305 // to replace the dead one with the existing one. This can cause
1306 // recursive merging of other unrelated nodes down the line.
1307 Existing->intersectFlagsWith(N->getFlags());
1308 if (auto *MemNode = dyn_cast<MemSDNode>(Existing))
1309 MemNode->refineRanges(cast<MemSDNode>(N)->memoperands());
1310 ReplaceAllUsesWith(N, Existing);
1311
1312 // N is now dead. Inform the listeners and delete it.
1313 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1314 DUL->NodeDeleted(N, Existing);
1315 DeleteNodeNotInCSEMaps(N);
1316 return;
1317 }
1318 }
1319
1320 // If the node doesn't already exist, we updated it. Inform listeners.
1321 for (DAGUpdateListener *DUL = UpdateListeners; DUL; DUL = DUL->Next)
1322 DUL->NodeUpdated(N);
1323}
1324
1325/// FindModifiedNodeSlot - Find a slot for the specified node if its operands
1326/// were replaced with those specified. If this node is never memoized,
1327/// return null, otherwise return a pointer to the slot it would take. If a
1328/// node already exists with these operands, the slot will be non-null.
1329SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N, SDValue Op,
1330 void *&InsertPos) {
1331 if (doNotCSE(N))
1332 return nullptr;
1333
1334 SDValue Ops[] = { Op };
1335 FoldingSetNodeID ID;
1336 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops);
1338 SDNode *Node = FindNodeOrInsertPos(ID, SDLoc(N), InsertPos);
1339 if (Node)
1340 Node->intersectFlagsWith(N->getFlags());
1341 return Node;
1342}
1343
1344/// FindModifiedNodeSlot - Find a slot for the specified node if its operands
1345/// were replaced with those specified. If this node is never memoized,
1346/// return null, otherwise return a pointer to the slot it would take. If a
1347/// node already exists with these operands, the slot will be non-null.
1348SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N,
1349 SDValue Op1, SDValue Op2,
1350 void *&InsertPos) {
1351 if (doNotCSE(N))
1352 return nullptr;
1353
1354 SDValue Ops[] = { Op1, Op2 };
1355 FoldingSetNodeID ID;
1356 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops);
1358 SDNode *Node = FindNodeOrInsertPos(ID, SDLoc(N), InsertPos);
1359 if (Node)
1360 Node->intersectFlagsWith(N->getFlags());
1361 return Node;
1362}
1363
1364/// FindModifiedNodeSlot - Find a slot for the specified node if its operands
1365/// were replaced with those specified. If this node is never memoized,
1366/// return null, otherwise return a pointer to the slot it would take. If a
1367/// node already exists with these operands, the slot will be non-null.
1368SDNode *SelectionDAG::FindModifiedNodeSlot(SDNode *N, ArrayRef<SDValue> Ops,
1369 void *&InsertPos) {
1370 if (doNotCSE(N))
1371 return nullptr;
1372
1373 FoldingSetNodeID ID;
1374 AddNodeIDNode(ID, N->getOpcode(), N->getVTList(), Ops);
1376 SDNode *Node = FindNodeOrInsertPos(ID, SDLoc(N), InsertPos);
1377 if (Node)
1378 Node->intersectFlagsWith(N->getFlags());
1379 return Node;
1380}
1381
1383 Type *Ty = VT == MVT::iPTR ? PointerType::get(*getContext(), 0)
1384 : VT.getTypeForEVT(*getContext());
1385
1386 return getDataLayout().getABITypeAlign(Ty);
1387}
1388
1389// EntryNode could meaningfully have debug info if we can find it...
1391 : TM(tm), OptLevel(OL), EntryNode(ISD::EntryToken, 0, DebugLoc(),
1392 getVTList(MVT::Other, MVT::Glue)),
1393 Root(getEntryNode()) {
1394 InsertNode(&EntryNode);
1395 DbgInfo = new SDDbgInfo();
1396}
1397
1399 OptimizationRemarkEmitter &NewORE, Pass *PassPtr,
1400 const TargetLibraryInfo *LibraryInfo,
1401 const LibcallLoweringInfo *LibcallsInfo,
1402 UniformityInfo *NewUA, ProfileSummaryInfo *PSIin,
1404 FunctionVarLocs const *VarLocs) {
1405 MF = &NewMF;
1406 SDAGISelPass = PassPtr;
1407 ORE = &NewORE;
1410 LibInfo = LibraryInfo;
1411 Libcalls = LibcallsInfo;
1412 Context = &MF->getFunction().getContext();
1413 UA = NewUA;
1414 PSI = PSIin;
1415 BFI = BFIin;
1416 MMI = &MMIin;
1417 FnVarLocs = VarLocs;
1418}
1419
1421 assert(!UpdateListeners && "Dangling registered DAGUpdateListeners");
1422 allnodes_clear();
1423 OperandRecycler.clear(OperandAllocator);
1424 delete DbgInfo;
1425}
1426
1428 return llvm::shouldOptimizeForSize(FLI->MBB->getBasicBlock(), PSI, BFI);
1429}
1430
1431void SelectionDAG::allnodes_clear() {
1432 assert(&*AllNodes.begin() == &EntryNode);
1433 AllNodes.remove(AllNodes.begin());
1434 while (!AllNodes.empty())
1435 DeallocateNode(&AllNodes.front());
1436#ifndef NDEBUG
1437 NextPersistentId = 0;
1438#endif
1439}
1440
1441SDNode *SelectionDAG::FindNodeOrInsertPos(const FoldingSetNodeID &ID,
1442 void *&InsertPos) {
1443 SDNode *N = CSEMap.FindNodeOrInsertPos(ID, InsertPos);
1444 if (N) {
1445 switch (N->getOpcode()) {
1446 default: break;
1447 case ISD::Constant:
1448 case ISD::ConstantFP:
1449 llvm_unreachable("Querying for Constant and ConstantFP nodes requires "
1450 "debug location. Use another overload.");
1451 }
1452 }
1453 return N;
1454}
1455
1456SDNode *SelectionDAG::FindNodeOrInsertPos(const FoldingSetNodeID &ID,
1457 const SDLoc &DL, void *&InsertPos) {
1458 SDNode *N = CSEMap.FindNodeOrInsertPos(ID, InsertPos);
1459 if (N) {
1460 switch (N->getOpcode()) {
1461 case ISD::Constant:
1462 case ISD::ConstantFP:
1463 // Erase debug location from the node if the node is used at several
1464 // different places. Do not propagate one location to all uses as it
1465 // will cause a worse single stepping debugging experience.
1466 if (N->getDebugLoc() != DL.getDebugLoc())
1467 N->setDebugLoc(DebugLoc());
1468 break;
1469 default:
1470 // When the node's point of use is located earlier in the instruction
1471 // sequence than its prior point of use, update its debug info to the
1472 // earlier location.
1473 if (DL.getIROrder() && DL.getIROrder() < N->getIROrder())
1474 N->setDebugLoc(DL.getDebugLoc());
1475 break;
1476 }
1477 }
1478 return N;
1479}
1480
1482 allnodes_clear();
1483 OperandRecycler.clear(OperandAllocator);
1484 OperandAllocator.Reset();
1485 CSEMap.clear();
1486
1487 ExtendedValueTypeNodes.clear();
1488 ExternalSymbols.clear();
1489 TargetExternalSymbols.clear();
1490 MCSymbols.clear();
1491 SDEI.clear();
1492 llvm::fill(CondCodeNodes, nullptr);
1493 llvm::fill(ValueTypeNodes, nullptr);
1494
1495 EntryNode.UseList = nullptr;
1496 InsertNode(&EntryNode);
1497 Root = getEntryNode();
1498 DbgInfo->clear();
1499}
1500
1502 return VT.bitsGT(Op.getValueType())
1503 ? getNode(ISD::FP_EXTEND, DL, VT, Op)
1504 : getNode(ISD::FP_ROUND, DL, VT, Op,
1505 getIntPtrConstant(0, DL, /*isTarget=*/true));
1506}
1507
1508std::pair<SDValue, SDValue>
1510 const SDLoc &DL, EVT VT) {
1511 assert(!VT.bitsEq(Op.getValueType()) &&
1512 "Strict no-op FP extend/round not allowed.");
1513 SDValue Res =
1514 VT.bitsGT(Op.getValueType())
1515 ? getNode(ISD::STRICT_FP_EXTEND, DL, {VT, MVT::Other}, {Chain, Op})
1516 : getNode(ISD::STRICT_FP_ROUND, DL, {VT, MVT::Other},
1517 {Chain, Op, getIntPtrConstant(0, DL, /*isTarget=*/true)});
1518
1519 return std::pair<SDValue, SDValue>(Res, SDValue(Res.getNode(), 1));
1520}
1521
1523 return VT.bitsGT(Op.getValueType()) ?
1524 getNode(ISD::ANY_EXTEND, DL, VT, Op) :
1525 getNode(ISD::TRUNCATE, DL, VT, Op);
1526}
1527
1529 return VT.bitsGT(Op.getValueType()) ?
1530 getNode(ISD::SIGN_EXTEND, DL, VT, Op) :
1531 getNode(ISD::TRUNCATE, DL, VT, Op);
1532}
1533
1535 return VT.bitsGT(Op.getValueType()) ?
1536 getNode(ISD::ZERO_EXTEND, DL, VT, Op) :
1537 getNode(ISD::TRUNCATE, DL, VT, Op);
1538}
1539
1541 EVT VT) {
1542 assert(!VT.isVector());
1543 auto Type = Op.getValueType();
1544 SDValue DestOp;
1545 if (Type == VT)
1546 return Op;
1547 auto Size = Op.getValueSizeInBits();
1548 DestOp = getBitcast(EVT::getIntegerVT(*Context, Size), Op);
1549 if (DestOp.getValueType() == VT)
1550 return DestOp;
1551
1552 return getAnyExtOrTrunc(DestOp, DL, VT);
1553}
1554
1556 EVT VT) {
1557 assert(!VT.isVector());
1558 auto Type = Op.getValueType();
1559 SDValue DestOp;
1560 if (Type == VT)
1561 return Op;
1562 auto Size = Op.getValueSizeInBits();
1563 DestOp = getBitcast(MVT::getIntegerVT(Size), Op);
1564 if (DestOp.getValueType() == VT)
1565 return DestOp;
1566
1567 return getSExtOrTrunc(DestOp, DL, VT);
1568}
1569
1571 EVT VT) {
1572 assert(!VT.isVector());
1573 auto Type = Op.getValueType();
1574 SDValue DestOp;
1575 if (Type == VT)
1576 return Op;
1577 auto Size = Op.getValueSizeInBits();
1578 DestOp = getBitcast(MVT::getIntegerVT(Size), Op);
1579 if (DestOp.getValueType() == VT)
1580 return DestOp;
1581
1582 return getZExtOrTrunc(DestOp, DL, VT);
1583}
1584
1586 EVT OpVT) {
1587 if (VT.bitsLE(Op.getValueType()))
1588 return getNode(ISD::TRUNCATE, SL, VT, Op);
1589
1590 TargetLowering::BooleanContent BType = TLI->getBooleanContents(OpVT);
1591 return getNode(TLI->getExtendForContent(BType), SL, VT, Op);
1592}
1593
1595 EVT OpVT = Op.getValueType();
1596 assert(VT.isInteger() && OpVT.isInteger() &&
1597 "Cannot getZeroExtendInReg FP types");
1598 assert(VT.isVector() == OpVT.isVector() &&
1599 "getZeroExtendInReg type should be vector iff the operand "
1600 "type is vector!");
1601 assert((!VT.isVector() ||
1603 "Vector element counts must match in getZeroExtendInReg");
1604 assert(VT.getScalarType().bitsLE(OpVT.getScalarType()) && "Not extending!");
1605 if (OpVT == VT)
1606 return Op;
1607 // TODO: Use computeKnownBits instead of AssertZext.
1608 if (Op.getOpcode() == ISD::AssertZext && cast<VTSDNode>(Op.getOperand(1))
1609 ->getVT()
1610 .getScalarType()
1611 .bitsLE(VT.getScalarType()))
1612 return Op;
1614 VT.getScalarSizeInBits());
1615 return getNode(ISD::AND, DL, OpVT, Op, getConstant(Imm, DL, OpVT));
1616}
1617
1619 SDValue EVL, const SDLoc &DL,
1620 EVT VT) {
1621 EVT OpVT = Op.getValueType();
1622 assert(VT.isInteger() && OpVT.isInteger() &&
1623 "Cannot getVPZeroExtendInReg FP types");
1624 assert(VT.isVector() && OpVT.isVector() &&
1625 "getVPZeroExtendInReg type and operand type should be vector!");
1627 "Vector element counts must match in getZeroExtendInReg");
1628 assert(VT.getScalarType().bitsLE(OpVT.getScalarType()) && "Not extending!");
1629 if (OpVT == VT)
1630 return Op;
1632 VT.getScalarSizeInBits());
1633 return getNode(ISD::VP_AND, DL, OpVT, Op, getConstant(Imm, DL, OpVT), Mask,
1634 EVL);
1635}
1636
1638 // Only unsigned pointer semantics are supported right now. In the future this
1639 // might delegate to TLI to check pointer signedness.
1640 return getZExtOrTrunc(Op, DL, VT);
1641}
1642
1644 // Only unsigned pointer semantics are supported right now. In the future this
1645 // might delegate to TLI to check pointer signedness.
1646 return getZeroExtendInReg(Op, DL, VT);
1647}
1648
1650 return getNode(ISD::SUB, DL, VT, getConstant(0, DL, VT), Val);
1651}
1652
1653/// getNOT - Create a bitwise NOT operation as (XOR Val, -1).
1655 return getNode(ISD::XOR, DL, VT, Val, getAllOnesConstant(DL, VT));
1656}
1657
1659 SDValue TrueValue = getBoolConstant(true, DL, VT, VT);
1660 return getNode(ISD::XOR, DL, VT, Val, TrueValue);
1661}
1662
1664 SDValue Mask, SDValue EVL, EVT VT) {
1665 SDValue TrueValue = getBoolConstant(true, DL, VT, VT);
1666 return getNode(ISD::VP_XOR, DL, VT, Val, TrueValue, Mask, EVL);
1667}
1668
1670 SDValue Mask, SDValue EVL) {
1671 return getVPZExtOrTrunc(DL, VT, Op, Mask, EVL);
1672}
1673
1675 SDValue Mask, SDValue EVL) {
1676 if (VT.bitsGT(Op.getValueType()))
1677 return getNode(ISD::VP_ZERO_EXTEND, DL, VT, Op, Mask, EVL);
1678 if (VT.bitsLT(Op.getValueType()))
1679 return getNode(ISD::VP_TRUNCATE, DL, VT, Op, Mask, EVL);
1680 return Op;
1681}
1682
1684 EVT OpVT) {
1685 if (!V)
1686 return getConstant(0, DL, VT);
1687
1688 switch (TLI->getBooleanContents(OpVT)) {
1691 return getConstant(1, DL, VT);
1693 return getAllOnesConstant(DL, VT);
1694 }
1695 llvm_unreachable("Unexpected boolean content enum!");
1696}
1697
1699 bool isT, bool isO) {
1700 return getConstant(APInt(VT.getScalarSizeInBits(), Val, /*isSigned=*/false),
1701 DL, VT, isT, isO);
1702}
1703
1705 bool isT, bool isO) {
1706 return getConstant(*ConstantInt::get(*Context, Val), DL, VT, isT, isO);
1707}
1708
1710 EVT VT, bool isT, bool isO) {
1711 assert(VT.isInteger() && "Cannot create FP integer constant!");
1712
1713 EVT EltVT = VT.getScalarType();
1714 const ConstantInt *Elt = &Val;
1715
1716 // Vector splats are explicit within the DAG, with ConstantSDNode holding the
1717 // to-be-splatted scalar ConstantInt.
1718 if (isa<VectorType>(Elt->getType()))
1719 Elt = ConstantInt::get(*getContext(), Elt->getValue());
1720
1721 // In some cases the vector type is legal but the element type is illegal and
1722 // needs to be promoted, for example v8i8 on ARM. In this case, promote the
1723 // inserted value (the type does not need to match the vector element type).
1724 // Any extra bits introduced will be truncated away.
1725 if (VT.isVector() && TLI->getTypeAction(*getContext(), EltVT) ==
1727 EltVT = TLI->getTypeToTransformTo(*getContext(), EltVT);
1728 APInt NewVal;
1729 if (TLI->isSExtCheaperThanZExt(VT.getScalarType(), EltVT))
1730 NewVal = Elt->getValue().sextOrTrunc(EltVT.getSizeInBits());
1731 else
1732 NewVal = Elt->getValue().zextOrTrunc(EltVT.getSizeInBits());
1733 Elt = ConstantInt::get(*getContext(), NewVal);
1734 }
1735 // In other cases the element type is illegal and needs to be expanded, for
1736 // example v2i64 on MIPS32. In this case, find the nearest legal type, split
1737 // the value into n parts and use a vector type with n-times the elements.
1738 // Then bitcast to the type requested.
1739 // Legalizing constants too early makes the DAGCombiner's job harder so we
1740 // only legalize if the DAG tells us we must produce legal types.
1741 else if (NewNodesMustHaveLegalTypes && VT.isVector() &&
1742 TLI->getTypeAction(*getContext(), EltVT) ==
1744 const APInt &NewVal = Elt->getValue();
1745 EVT ViaEltVT = TLI->getTypeToTransformTo(*getContext(), EltVT);
1746 unsigned ViaEltSizeInBits = ViaEltVT.getSizeInBits();
1747
1748 // For scalable vectors, try to use a SPLAT_VECTOR_PARTS node.
1749 if (VT.isScalableVector() ||
1750 TLI->isOperationLegal(ISD::SPLAT_VECTOR, VT)) {
1751 assert(EltVT.getSizeInBits() % ViaEltSizeInBits == 0 &&
1752 "Can only handle an even split!");
1753 unsigned Parts = EltVT.getSizeInBits() / ViaEltSizeInBits;
1754
1755 SmallVector<SDValue, 2> ScalarParts;
1756 for (unsigned i = 0; i != Parts; ++i)
1757 ScalarParts.push_back(getConstant(
1758 NewVal.extractBits(ViaEltSizeInBits, i * ViaEltSizeInBits), DL,
1759 ViaEltVT, isT, isO));
1760
1761 return getNode(ISD::SPLAT_VECTOR_PARTS, DL, VT, ScalarParts);
1762 }
1763
1764 unsigned ViaVecNumElts = VT.getSizeInBits() / ViaEltSizeInBits;
1765 EVT ViaVecVT = EVT::getVectorVT(*getContext(), ViaEltVT, ViaVecNumElts);
1766
1767 // Check the temporary vector is the correct size. If this fails then
1768 // getTypeToTransformTo() probably returned a type whose size (in bits)
1769 // isn't a power-of-2 factor of the requested type size.
1770 assert(ViaVecVT.getSizeInBits() == VT.getSizeInBits());
1771
1772 SmallVector<SDValue, 2> EltParts;
1773 for (unsigned i = 0; i < ViaVecNumElts / VT.getVectorNumElements(); ++i)
1774 EltParts.push_back(getConstant(
1775 NewVal.extractBits(ViaEltSizeInBits, i * ViaEltSizeInBits), DL,
1776 ViaEltVT, isT, isO));
1777
1778 // EltParts is currently in little endian order. If we actually want
1779 // big-endian order then reverse it now.
1780 if (getDataLayout().isBigEndian())
1781 std::reverse(EltParts.begin(), EltParts.end());
1782
1783 // The elements must be reversed when the element order is different
1784 // to the endianness of the elements (because the BITCAST is itself a
1785 // vector shuffle in this situation). However, we do not need any code to
1786 // perform this reversal because getConstant() is producing a vector
1787 // splat.
1788 // This situation occurs in MIPS MSA.
1789
1791 for (unsigned i = 0, e = VT.getVectorNumElements(); i != e; ++i)
1792 llvm::append_range(Ops, EltParts);
1793
1794 SDValue V =
1795 getNode(ISD::BITCAST, DL, VT, getBuildVector(ViaVecVT, DL, Ops));
1796 return V;
1797 }
1798
1799 assert(Elt->getBitWidth() == EltVT.getSizeInBits() &&
1800 "APInt size does not match type size!");
1801 unsigned Opc = isT ? ISD::TargetConstant : ISD::Constant;
1802 SDVTList VTs = getVTList(EltVT);
1804 AddNodeIDNode(ID, Opc, VTs, {});
1805 ID.AddPointer(Elt);
1806 ID.AddBoolean(isO);
1807 void *IP = nullptr;
1808 SDNode *N = nullptr;
1809 if ((N = FindNodeOrInsertPos(ID, DL, IP)))
1810 if (!VT.isVector())
1811 return SDValue(N, 0);
1812
1813 if (!N) {
1814 N = newSDNode<ConstantSDNode>(isT, isO, Elt, VTs);
1815 CSEMap.InsertNode(N, IP);
1816 InsertNode(N);
1817 NewSDValueDbgMsg(SDValue(N, 0), "Creating constant: ", this);
1818 }
1819
1820 SDValue Result(N, 0);
1821 if (VT.isVector())
1822 Result = getSplat(VT, DL, Result);
1823 return Result;
1824}
1825
1827 bool isT, bool isO) {
1828 unsigned Size = VT.getScalarSizeInBits();
1829 return getConstant(APInt(Size, Val, /*isSigned=*/true), DL, VT, isT, isO);
1830}
1831
1833 bool IsOpaque) {
1835 IsTarget, IsOpaque);
1836}
1837
1839 bool isTarget) {
1840 return getConstant(Val, DL, TLI->getPointerTy(getDataLayout()), isTarget);
1841}
1842
1844 const SDLoc &DL) {
1845 assert(VT.isInteger() && "Shift amount is not an integer type!");
1846 EVT ShiftVT = TLI->getShiftAmountTy(VT, getDataLayout());
1847 return getConstant(Val, DL, ShiftVT);
1848}
1849
1851 const SDLoc &DL) {
1852 assert(Val.ult(VT.getScalarSizeInBits()) && "Out of range shift");
1853 return getShiftAmountConstant(Val.getZExtValue(), VT, DL);
1854}
1855
1857 bool isTarget) {
1858 return getConstant(Val, DL, TLI->getVectorIdxTy(getDataLayout()), isTarget);
1859}
1860
1862 bool isTarget) {
1863 return getConstantFP(*ConstantFP::get(*getContext(), V), DL, VT, isTarget);
1864}
1865
1867 EVT VT, bool isTarget) {
1868 assert(VT.isFloatingPoint() && "Cannot create integer FP constant!");
1869
1870 EVT EltVT = VT.getScalarType();
1871 const ConstantFP *Elt = &V;
1872
1873 // Vector splats are explicit within the DAG, with ConstantFPSDNode holding
1874 // the to-be-splatted scalar ConstantFP.
1875 if (isa<VectorType>(Elt->getType()))
1876 Elt = ConstantFP::get(*getContext(), Elt->getValue());
1877
1878 // Do the map lookup using the actual bit pattern for the floating point
1879 // value, so that we don't have problems with 0.0 comparing equal to -0.0, and
1880 // we don't have issues with SNANs.
1881 unsigned Opc = isTarget ? ISD::TargetConstantFP : ISD::ConstantFP;
1882 SDVTList VTs = getVTList(EltVT);
1884 AddNodeIDNode(ID, Opc, VTs, {});
1885 ID.AddPointer(Elt);
1886 void *IP = nullptr;
1887 SDNode *N = nullptr;
1888 if ((N = FindNodeOrInsertPos(ID, DL, IP)))
1889 if (!VT.isVector())
1890 return SDValue(N, 0);
1891
1892 if (!N) {
1893 N = newSDNode<ConstantFPSDNode>(isTarget, Elt, VTs);
1894 CSEMap.InsertNode(N, IP);
1895 InsertNode(N);
1896 }
1897
1898 SDValue Result(N, 0);
1899 if (VT.isVector())
1900 Result = getSplat(VT, DL, Result);
1901 NewSDValueDbgMsg(Result, "Creating fp constant: ", this);
1902 return Result;
1903}
1904
1906 bool isTarget) {
1907 EVT EltVT = VT.getScalarType();
1908 if (EltVT == MVT::f32)
1909 return getConstantFP(APFloat((float)Val), DL, VT, isTarget);
1910 if (EltVT == MVT::f64)
1911 return getConstantFP(APFloat(Val), DL, VT, isTarget);
1912 if (EltVT == MVT::f80 || EltVT == MVT::f128 || EltVT == MVT::ppcf128 ||
1913 EltVT == MVT::f16 || EltVT == MVT::bf16) {
1914 bool Ignored;
1915 APFloat APF = APFloat(Val);
1917 &Ignored);
1918 return getConstantFP(APF, DL, VT, isTarget);
1919 }
1920 llvm_unreachable("Unsupported type in getConstantFP");
1921}
1922
1924 EVT VT, int64_t Offset, bool isTargetGA,
1925 unsigned TargetFlags) {
1926 assert((TargetFlags == 0 || isTargetGA) &&
1927 "Cannot set target flags on target-independent globals");
1928
1929 // Truncate (with sign-extension) the offset value to the pointer size.
1931 if (BitWidth < 64)
1933
1934 unsigned Opc;
1935 if (GV->isThreadLocal())
1937 else
1939
1940 SDVTList VTs = getVTList(VT);
1942 AddNodeIDNode(ID, Opc, VTs, {});
1943 ID.AddPointer(GV);
1944 ID.AddInteger(Offset);
1945 ID.AddInteger(TargetFlags);
1946 void *IP = nullptr;
1947 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP))
1948 return SDValue(E, 0);
1949
1950 auto *N = newSDNode<GlobalAddressSDNode>(
1951 Opc, DL.getIROrder(), DL.getDebugLoc(), GV, VTs, Offset, TargetFlags);
1952 CSEMap.InsertNode(N, IP);
1953 InsertNode(N);
1954 return SDValue(N, 0);
1955}
1956
1958 SDVTList VTs = getVTList(MVT::Untyped);
1961 ID.AddPointer(GV);
1962 void *IP = nullptr;
1963 if (SDNode *E = FindNodeOrInsertPos(ID, SDLoc(), IP))
1964 return SDValue(E, 0);
1965
1966 auto *N = newSDNode<DeactivationSymbolSDNode>(GV, VTs);
1967 CSEMap.InsertNode(N, IP);
1968 InsertNode(N);
1969 return SDValue(N, 0);
1970}
1971
1972SDValue SelectionDAG::getFrameIndex(int FI, EVT VT, bool isTarget) {
1973 unsigned Opc = isTarget ? ISD::TargetFrameIndex : ISD::FrameIndex;
1974 SDVTList VTs = getVTList(VT);
1976 AddNodeIDNode(ID, Opc, VTs, {});
1977 ID.AddInteger(FI);
1978 void *IP = nullptr;
1979 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
1980 return SDValue(E, 0);
1981
1982 auto *N = newSDNode<FrameIndexSDNode>(FI, VTs, isTarget);
1983 CSEMap.InsertNode(N, IP);
1984 InsertNode(N);
1985 return SDValue(N, 0);
1986}
1987
1988SDValue SelectionDAG::getJumpTable(int JTI, EVT VT, bool isTarget,
1989 unsigned TargetFlags) {
1990 assert((TargetFlags == 0 || isTarget) &&
1991 "Cannot set target flags on target-independent jump tables");
1992 unsigned Opc = isTarget ? ISD::TargetJumpTable : ISD::JumpTable;
1993 SDVTList VTs = getVTList(VT);
1995 AddNodeIDNode(ID, Opc, VTs, {});
1996 ID.AddInteger(JTI);
1997 ID.AddInteger(TargetFlags);
1998 void *IP = nullptr;
1999 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2000 return SDValue(E, 0);
2001
2002 auto *N = newSDNode<JumpTableSDNode>(JTI, VTs, isTarget, TargetFlags);
2003 CSEMap.InsertNode(N, IP);
2004 InsertNode(N);
2005 return SDValue(N, 0);
2006}
2007
2009 const SDLoc &DL) {
2011 return getNode(ISD::JUMP_TABLE_DEBUG_INFO, DL, MVT::Other, Chain,
2012 getTargetConstant(static_cast<uint64_t>(JTI), DL, PTy, true));
2013}
2014
2016 MaybeAlign Alignment, int Offset,
2017 bool isTarget, unsigned TargetFlags) {
2018 assert((TargetFlags == 0 || isTarget) &&
2019 "Cannot set target flags on target-independent globals");
2020 if (!Alignment)
2021 Alignment = shouldOptForSize()
2022 ? getDataLayout().getABITypeAlign(C->getType())
2023 : getDataLayout().getPrefTypeAlign(C->getType());
2024 unsigned Opc = isTarget ? ISD::TargetConstantPool : ISD::ConstantPool;
2025 SDVTList VTs = getVTList(VT);
2027 AddNodeIDNode(ID, Opc, VTs, {});
2028 ID.AddInteger(Alignment->value());
2029 ID.AddInteger(Offset);
2030 ID.AddPointer(C);
2031 ID.AddInteger(TargetFlags);
2032 void *IP = nullptr;
2033 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2034 return SDValue(E, 0);
2035
2036 auto *N = newSDNode<ConstantPoolSDNode>(isTarget, C, VTs, Offset, *Alignment,
2037 TargetFlags);
2038 CSEMap.InsertNode(N, IP);
2039 InsertNode(N);
2040 SDValue V = SDValue(N, 0);
2041 NewSDValueDbgMsg(V, "Creating new constant pool: ", this);
2042 return V;
2043}
2044
2046 MaybeAlign Alignment, int Offset,
2047 bool isTarget, unsigned TargetFlags) {
2048 assert((TargetFlags == 0 || isTarget) &&
2049 "Cannot set target flags on target-independent globals");
2050 if (!Alignment)
2051 Alignment = getDataLayout().getPrefTypeAlign(C->getType());
2052 unsigned Opc = isTarget ? ISD::TargetConstantPool : ISD::ConstantPool;
2053 SDVTList VTs = getVTList(VT);
2055 AddNodeIDNode(ID, Opc, VTs, {});
2056 ID.AddInteger(Alignment->value());
2057 ID.AddInteger(Offset);
2058 C->addSelectionDAGCSEId(ID);
2059 ID.AddInteger(TargetFlags);
2060 void *IP = nullptr;
2061 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2062 return SDValue(E, 0);
2063
2064 auto *N = newSDNode<ConstantPoolSDNode>(isTarget, C, VTs, Offset, *Alignment,
2065 TargetFlags);
2066 CSEMap.InsertNode(N, IP);
2067 InsertNode(N);
2068 return SDValue(N, 0);
2069}
2070
2073 AddNodeIDNode(ID, ISD::BasicBlock, getVTList(MVT::Other), {});
2074 ID.AddPointer(MBB);
2075 void *IP = nullptr;
2076 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2077 return SDValue(E, 0);
2078
2079 auto *N = newSDNode<BasicBlockSDNode>(MBB);
2080 CSEMap.InsertNode(N, IP);
2081 InsertNode(N);
2082 return SDValue(N, 0);
2083}
2084
2086 if (VT.isSimple() && (unsigned)VT.getSimpleVT().SimpleTy >=
2087 ValueTypeNodes.size())
2088 ValueTypeNodes.resize(VT.getSimpleVT().SimpleTy+1);
2089
2090 SDNode *&N = VT.isExtended() ?
2091 ExtendedValueTypeNodes[VT] : ValueTypeNodes[VT.getSimpleVT().SimpleTy];
2092
2093 if (N) return SDValue(N, 0);
2094 N = newSDNode<VTSDNode>(VT);
2095 InsertNode(N);
2096 return SDValue(N, 0);
2097}
2098
2100 SDNode *&N = ExternalSymbols[Sym];
2101 if (N) return SDValue(N, 0);
2102 N = newSDNode<ExternalSymbolSDNode>(false, Sym, 0, getVTList(VT));
2103 InsertNode(N);
2104 return SDValue(N, 0);
2105}
2106
2107SDValue SelectionDAG::getExternalSymbol(RTLIB::LibcallImpl Libcall, EVT VT) {
2109 return getExternalSymbol(SymName.data(), VT);
2110}
2111
2113 SDNode *&N = MCSymbols[Sym];
2114 if (N)
2115 return SDValue(N, 0);
2116 N = newSDNode<MCSymbolSDNode>(Sym, getVTList(VT));
2117 InsertNode(N);
2118 return SDValue(N, 0);
2119}
2120
2122 unsigned TargetFlags) {
2123 SDNode *&N =
2124 TargetExternalSymbols[std::pair<std::string, unsigned>(Sym, TargetFlags)];
2125 if (N) return SDValue(N, 0);
2126 N = newSDNode<ExternalSymbolSDNode>(true, Sym, TargetFlags, getVTList(VT));
2127 InsertNode(N);
2128 return SDValue(N, 0);
2129}
2130
2132 EVT VT, unsigned TargetFlags) {
2134 return getTargetExternalSymbol(SymName.data(), VT, TargetFlags);
2135}
2136
2138 if ((unsigned)Cond >= CondCodeNodes.size())
2139 CondCodeNodes.resize(Cond+1);
2140
2141 if (!CondCodeNodes[Cond]) {
2142 auto *N = newSDNode<CondCodeSDNode>(Cond);
2143 CondCodeNodes[Cond] = N;
2144 InsertNode(N);
2145 }
2146
2147 return SDValue(CondCodeNodes[Cond], 0);
2148}
2149
2151 assert(MulImm.getBitWidth() == VT.getSizeInBits() &&
2152 "APInt size does not match type size!");
2153
2154 if (MulImm == 0)
2155 return getConstant(0, DL, VT);
2156
2157 const MachineFunction &MF = getMachineFunction();
2158 const Function &F = MF.getFunction();
2159 ConstantRange CR = getVScaleRange(&F, 64);
2160 if (const APInt *C = CR.getSingleElement())
2161 return getConstant(MulImm * C->getZExtValue(), DL, VT);
2162
2163 return getNode(ISD::VSCALE, DL, VT, getConstant(MulImm, DL, VT));
2164}
2165
2166/// \returns a value of type \p VT that represents the runtime value of \p
2167/// Quantity, i.e. scaled by vscale if it's scalable, or a fixed constant
2168/// otherwise. Quantity should be a FixedOrScalableQuantity, i.e. ElementCount
2169/// or TypeSize.
2170template <typename Ty>
2172 EVT VT, Ty Quantity) {
2173 if (Quantity.isScalable())
2174 return DAG.getVScale(
2175 DL, VT, APInt(VT.getSizeInBits(), Quantity.getKnownMinValue()));
2176
2177 return DAG.getConstant(Quantity.getKnownMinValue(), DL, VT);
2178}
2179
2181 ElementCount EC) {
2182 return getFixedOrScalableQuantity(*this, DL, VT, EC);
2183}
2184
2186 return getFixedOrScalableQuantity(*this, DL, VT, TS);
2187}
2188
2190 ElementCount EC) {
2191 EVT IdxVT = TLI->getVectorIdxTy(getDataLayout());
2192 EVT MaskVT = TLI->getSetCCResultType(getDataLayout(), *getContext(), DataVT);
2193 return getNode(ISD::GET_ACTIVE_LANE_MASK, DL, MaskVT,
2194 getConstant(0, DL, IdxVT), getElementCount(DL, IdxVT, EC));
2195}
2196
2198 APInt One(ResVT.getScalarSizeInBits(), 1);
2199 return getStepVector(DL, ResVT, One);
2200}
2201
2203 const APInt &StepVal) {
2204 assert(ResVT.getScalarSizeInBits() == StepVal.getBitWidth());
2205 if (ResVT.isScalableVector())
2206 return getNode(
2207 ISD::STEP_VECTOR, DL, ResVT,
2208 getTargetConstant(StepVal, DL, ResVT.getVectorElementType()));
2209
2210 SmallVector<SDValue, 16> OpsStepConstants;
2211 for (uint64_t i = 0; i < ResVT.getVectorNumElements(); i++)
2212 OpsStepConstants.push_back(
2213 getConstant(StepVal * i, DL, ResVT.getVectorElementType()));
2214 return getBuildVector(ResVT, DL, OpsStepConstants);
2215}
2216
2217/// Swaps the values of N1 and N2. Swaps all indices in the shuffle mask M that
2218/// point at N1 to point at N2 and indices that point at N2 to point at N1.
2223
2225 SDValue N2, ArrayRef<int> Mask) {
2226 assert(VT.getVectorNumElements() == Mask.size() &&
2227 "Must have the same number of vector elements as mask elements!");
2228 assert(VT == N1.getValueType() && VT == N2.getValueType() &&
2229 "Invalid VECTOR_SHUFFLE");
2230
2231 // Canonicalize shuffle undef, undef -> undef
2232 if (N1.isUndef() && N2.isUndef())
2233 return getUNDEF(VT);
2234
2235 // Validate that all indices in Mask are within the range of the elements
2236 // input to the shuffle.
2237 int NElts = Mask.size();
2238 assert(llvm::all_of(Mask,
2239 [&](int M) { return M < (NElts * 2) && M >= -1; }) &&
2240 "Index out of range");
2241
2242 // Copy the mask so we can do any needed cleanup.
2243 SmallVector<int, 8> MaskVec(Mask);
2244
2245 // Canonicalize shuffle v, v -> v, undef
2246 if (N1 == N2) {
2247 N2 = getUNDEF(VT);
2248 for (int i = 0; i != NElts; ++i)
2249 if (MaskVec[i] >= NElts) MaskVec[i] -= NElts;
2250 }
2251
2252 // Canonicalize shuffle undef, v -> v, undef. Commute the shuffle mask.
2253 if (N1.isUndef())
2254 commuteShuffle(N1, N2, MaskVec);
2255
2256 if (TLI->hasVectorBlend()) {
2257 // If shuffling a splat, try to blend the splat instead. We do this here so
2258 // that even when this arises during lowering we don't have to re-handle it.
2259 auto BlendSplat = [&](BuildVectorSDNode *BV, int Offset) {
2260 BitVector UndefElements;
2261 SDValue Splat = BV->getSplatValue(&UndefElements);
2262 if (!Splat)
2263 return;
2264
2265 for (int i = 0; i < NElts; ++i) {
2266 if (MaskVec[i] < Offset || MaskVec[i] >= (Offset + NElts))
2267 continue;
2268
2269 // If this input comes from undef, mark it as such.
2270 if (UndefElements[MaskVec[i] - Offset]) {
2271 MaskVec[i] = -1;
2272 continue;
2273 }
2274
2275 // If we can blend a non-undef lane, use that instead.
2276 if (!UndefElements[i])
2277 MaskVec[i] = i + Offset;
2278 }
2279 };
2280 if (auto *N1BV = dyn_cast<BuildVectorSDNode>(N1))
2281 BlendSplat(N1BV, 0);
2282 if (auto *N2BV = dyn_cast<BuildVectorSDNode>(N2))
2283 BlendSplat(N2BV, NElts);
2284 }
2285
2286 // Canonicalize all index into lhs, -> shuffle lhs, undef
2287 // Canonicalize all index into rhs, -> shuffle rhs, undef
2288 bool AllLHS = true, AllRHS = true;
2289 bool N2Undef = N2.isUndef();
2290 for (int i = 0; i != NElts; ++i) {
2291 if (MaskVec[i] >= NElts) {
2292 if (N2Undef)
2293 MaskVec[i] = -1;
2294 else
2295 AllLHS = false;
2296 } else if (MaskVec[i] >= 0) {
2297 AllRHS = false;
2298 }
2299 }
2300 if (AllLHS && AllRHS)
2301 return getUNDEF(VT);
2302 if (AllLHS && !N2Undef)
2303 N2 = getUNDEF(VT);
2304 if (AllRHS) {
2305 N1 = getUNDEF(VT);
2306 commuteShuffle(N1, N2, MaskVec);
2307 }
2308 // Reset our undef status after accounting for the mask.
2309 N2Undef = N2.isUndef();
2310 // Re-check whether both sides ended up undef.
2311 if (N1.isUndef() && N2Undef)
2312 return getUNDEF(VT);
2313
2314 // If Identity shuffle return that node.
2315 bool Identity = true, AllSame = true;
2316 for (int i = 0; i != NElts; ++i) {
2317 if (MaskVec[i] >= 0 && MaskVec[i] != i) Identity = false;
2318 if (MaskVec[i] != MaskVec[0]) AllSame = false;
2319 }
2320 if (Identity && NElts)
2321 return N1;
2322
2323 // Shuffling a constant splat doesn't change the result.
2324 if (N2Undef) {
2325 SDValue V = N1;
2326
2327 // Look through any bitcasts. We check that these don't change the number
2328 // (and size) of elements and just changes their types.
2329 while (V.getOpcode() == ISD::BITCAST)
2330 V = V->getOperand(0);
2331
2332 // A splat should always show up as a build vector node.
2333 if (auto *BV = dyn_cast<BuildVectorSDNode>(V)) {
2334 BitVector UndefElements;
2335 SDValue Splat = BV->getSplatValue(&UndefElements);
2336 // If this is a splat of an undef, shuffling it is also undef.
2337 if (Splat && Splat.isUndef())
2338 return getUNDEF(VT);
2339
2340 bool SameNumElts =
2341 V.getValueType().getVectorNumElements() == VT.getVectorNumElements();
2342
2343 // We only have a splat which can skip shuffles if there is a splatted
2344 // value and no undef lanes rearranged by the shuffle.
2345 if (Splat && UndefElements.none()) {
2346 // Splat of <x, x, ..., x>, return <x, x, ..., x>, provided that the
2347 // number of elements match or the value splatted is a zero constant.
2348 if (SameNumElts || isNullConstant(Splat))
2349 return N1;
2350 }
2351
2352 // If the shuffle itself creates a splat, build the vector directly.
2353 if (AllSame && SameNumElts) {
2354 EVT BuildVT = BV->getValueType(0);
2355 const SDValue &Splatted = BV->getOperand(MaskVec[0]);
2356 SDValue NewBV = getSplatBuildVector(BuildVT, dl, Splatted);
2357
2358 // We may have jumped through bitcasts, so the type of the
2359 // BUILD_VECTOR may not match the type of the shuffle.
2360 if (BuildVT != VT)
2361 NewBV = getNode(ISD::BITCAST, dl, VT, NewBV);
2362 return NewBV;
2363 }
2364 }
2365 }
2366
2367 SDVTList VTs = getVTList(VT);
2369 SDValue Ops[2] = { N1, N2 };
2371 for (int i = 0; i != NElts; ++i)
2372 ID.AddInteger(MaskVec[i]);
2373
2374 void* IP = nullptr;
2375 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
2376 return SDValue(E, 0);
2377
2378 // Allocate the mask array for the node out of the BumpPtrAllocator, since
2379 // SDNode doesn't have access to it. This memory will be "leaked" when
2380 // the node is deallocated, but recovered when the NodeAllocator is released.
2381 int *MaskAlloc = OperandAllocator.Allocate<int>(NElts);
2382 llvm::copy(MaskVec, MaskAlloc);
2383
2384 auto *N = newSDNode<ShuffleVectorSDNode>(VTs, dl.getIROrder(),
2385 dl.getDebugLoc(), MaskAlloc);
2386 createOperands(N, Ops);
2387
2388 CSEMap.InsertNode(N, IP);
2389 InsertNode(N);
2390 SDValue V = SDValue(N, 0);
2391 NewSDValueDbgMsg(V, "Creating new node: ", this);
2392 return V;
2393}
2394
2396 EVT VT = SV.getValueType(0);
2397 SmallVector<int, 8> MaskVec(SV.getMask());
2399
2400 SDValue Op0 = SV.getOperand(0);
2401 SDValue Op1 = SV.getOperand(1);
2402 return getVectorShuffle(VT, SDLoc(&SV), Op1, Op0, MaskVec);
2403}
2404
2406 SDVTList VTs = getVTList(VT);
2408 AddNodeIDNode(ID, ISD::Register, VTs, {});
2409 ID.AddInteger(Reg.id());
2410 void *IP = nullptr;
2411 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2412 return SDValue(E, 0);
2413
2414 auto *N = newSDNode<RegisterSDNode>(Reg, VTs);
2415 N->SDNodeBits.IsDivergent = TLI->isSDNodeSourceOfDivergence(N, FLI, UA);
2416 CSEMap.InsertNode(N, IP);
2417 InsertNode(N);
2418 return SDValue(N, 0);
2419}
2420
2423 AddNodeIDNode(ID, ISD::RegisterMask, getVTList(MVT::Untyped), {});
2424 ID.AddPointer(RegMask);
2425 void *IP = nullptr;
2426 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2427 return SDValue(E, 0);
2428
2429 auto *N = newSDNode<RegisterMaskSDNode>(RegMask);
2430 CSEMap.InsertNode(N, IP);
2431 InsertNode(N);
2432 return SDValue(N, 0);
2433}
2434
2436 MCSymbol *Label) {
2437 return getLabelNode(ISD::EH_LABEL, dl, Root, Label);
2438}
2439
2440SDValue SelectionDAG::getLabelNode(unsigned Opcode, const SDLoc &dl,
2441 SDValue Root, MCSymbol *Label) {
2443 SDValue Ops[] = { Root };
2444 AddNodeIDNode(ID, Opcode, getVTList(MVT::Other), Ops);
2445 ID.AddPointer(Label);
2446 void *IP = nullptr;
2447 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2448 return SDValue(E, 0);
2449
2450 auto *N =
2451 newSDNode<LabelSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(), Label);
2452 createOperands(N, Ops);
2453
2454 CSEMap.InsertNode(N, IP);
2455 InsertNode(N);
2456 return SDValue(N, 0);
2457}
2458
2460 int64_t Offset, bool isTarget,
2461 unsigned TargetFlags) {
2462 unsigned Opc = isTarget ? ISD::TargetBlockAddress : ISD::BlockAddress;
2463 SDVTList VTs = getVTList(VT);
2464
2466 AddNodeIDNode(ID, Opc, VTs, {});
2467 ID.AddPointer(BA);
2468 ID.AddInteger(Offset);
2469 ID.AddInteger(TargetFlags);
2470 void *IP = nullptr;
2471 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2472 return SDValue(E, 0);
2473
2474 auto *N = newSDNode<BlockAddressSDNode>(Opc, VTs, BA, Offset, TargetFlags);
2475 CSEMap.InsertNode(N, IP);
2476 InsertNode(N);
2477 return SDValue(N, 0);
2478}
2479
2482 AddNodeIDNode(ID, ISD::SRCVALUE, getVTList(MVT::Other), {});
2483 ID.AddPointer(V);
2484
2485 void *IP = nullptr;
2486 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2487 return SDValue(E, 0);
2488
2489 auto *N = newSDNode<SrcValueSDNode>(V);
2490 CSEMap.InsertNode(N, IP);
2491 InsertNode(N);
2492 return SDValue(N, 0);
2493}
2494
2497 AddNodeIDNode(ID, ISD::MDNODE_SDNODE, getVTList(MVT::Other), {});
2498 ID.AddPointer(MD);
2499
2500 void *IP = nullptr;
2501 if (SDNode *E = FindNodeOrInsertPos(ID, IP))
2502 return SDValue(E, 0);
2503
2504 auto *N = newSDNode<MDNodeSDNode>(MD);
2505 CSEMap.InsertNode(N, IP);
2506 InsertNode(N);
2507 return SDValue(N, 0);
2508}
2509
2511 if (VT == V.getValueType())
2512 return V;
2513
2514 return getNode(ISD::BITCAST, SDLoc(V), VT, V);
2515}
2516
2518 unsigned SrcAS, unsigned DestAS) {
2519 SDVTList VTs = getVTList(VT);
2520 SDValue Ops[] = {Ptr};
2523 ID.AddInteger(SrcAS);
2524 ID.AddInteger(DestAS);
2525
2526 void *IP = nullptr;
2527 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
2528 return SDValue(E, 0);
2529
2530 auto *N = newSDNode<AddrSpaceCastSDNode>(dl.getIROrder(), dl.getDebugLoc(),
2531 VTs, SrcAS, DestAS);
2532 createOperands(N, Ops);
2533
2534 CSEMap.InsertNode(N, IP);
2535 InsertNode(N);
2536 return SDValue(N, 0);
2537}
2538
2540 return getNode(ISD::FREEZE, SDLoc(V), V.getValueType(), V);
2541}
2542
2544 bool PoisonOnly) {
2545 if (isGuaranteedNotToBeUndefOrPoison(V, DemandedElts, PoisonOnly))
2546 return V;
2547 return getFreeze(V);
2548}
2549
2550/// getShiftAmountOperand - Return the specified value casted to
2551/// the target's desired shift amount type.
2553 EVT OpTy = Op.getValueType();
2554 EVT ShTy = TLI->getShiftAmountTy(LHSTy, getDataLayout());
2555 if (OpTy == ShTy || OpTy.isVector()) return Op;
2556
2557 return getZExtOrTrunc(Op, SDLoc(Op), ShTy);
2558}
2559
2561 SDLoc dl(Node);
2563 const Value *V = cast<SrcValueSDNode>(Node->getOperand(2))->getValue();
2564 EVT VT = Node->getValueType(0);
2565 SDValue Tmp1 = Node->getOperand(0);
2566 SDValue Tmp2 = Node->getOperand(1);
2567 const MaybeAlign MA(Node->getConstantOperandVal(3));
2568
2569 SDValue VAListLoad = getLoad(TLI.getPointerTy(getDataLayout()), dl, Tmp1,
2570 Tmp2, MachinePointerInfo(V));
2571 SDValue VAList = VAListLoad;
2572
2573 if (MA && *MA > TLI.getMinStackArgumentAlignment()) {
2574 VAList = getNode(ISD::ADD, dl, VAList.getValueType(), VAList,
2575 getConstant(MA->value() - 1, dl, VAList.getValueType()));
2576
2577 VAList = getNode(
2578 ISD::AND, dl, VAList.getValueType(), VAList,
2579 getSignedConstant(-(int64_t)MA->value(), dl, VAList.getValueType()));
2580 }
2581
2582 // Increment the pointer, VAList, to the next vaarg
2583 Tmp1 = getNode(ISD::ADD, dl, VAList.getValueType(), VAList,
2584 getConstant(getDataLayout().getTypeAllocSize(
2585 VT.getTypeForEVT(*getContext())),
2586 dl, VAList.getValueType()));
2587 // Store the incremented VAList to the legalized pointer
2588 Tmp1 =
2589 getStore(VAListLoad.getValue(1), dl, Tmp1, Tmp2, MachinePointerInfo(V));
2590 // Load the actual argument out of the pointer VAList
2591 return getLoad(VT, dl, Tmp1, VAList, MachinePointerInfo());
2592}
2593
2595 SDLoc dl(Node);
2597 // This defaults to loading a pointer from the input and storing it to the
2598 // output, returning the chain.
2599 const Value *VD = cast<SrcValueSDNode>(Node->getOperand(3))->getValue();
2600 const Value *VS = cast<SrcValueSDNode>(Node->getOperand(4))->getValue();
2601 SDValue Tmp1 =
2602 getLoad(TLI.getPointerTy(getDataLayout()), dl, Node->getOperand(0),
2603 Node->getOperand(2), MachinePointerInfo(VS));
2604 return getStore(Tmp1.getValue(1), dl, Tmp1, Node->getOperand(1),
2605 MachinePointerInfo(VD));
2606}
2607
2609 const DataLayout &DL = getDataLayout();
2610 Type *Ty = VT.getTypeForEVT(*getContext());
2611 Align RedAlign = UseABI ? DL.getABITypeAlign(Ty) : DL.getPrefTypeAlign(Ty);
2612
2613 if (TLI->isTypeLegal(VT) || !VT.isVector())
2614 return RedAlign;
2615
2616 const TargetFrameLowering *TFI = MF->getSubtarget().getFrameLowering();
2617 const Align StackAlign = TFI->getStackAlign();
2618
2619 // See if we can choose a smaller ABI alignment in cases where it's an
2620 // illegal vector type that will get broken down.
2621 if (RedAlign > StackAlign) {
2622 EVT IntermediateVT;
2623 MVT RegisterVT;
2624 unsigned NumIntermediates;
2625 TLI->getVectorTypeBreakdown(*getContext(), VT, IntermediateVT,
2626 NumIntermediates, RegisterVT);
2627 Ty = IntermediateVT.getTypeForEVT(*getContext());
2628 Align RedAlign2 = UseABI ? DL.getABITypeAlign(Ty) : DL.getPrefTypeAlign(Ty);
2629 if (RedAlign2 < RedAlign)
2630 RedAlign = RedAlign2;
2631
2632 if (!getMachineFunction().getFrameInfo().isStackRealignable())
2633 // If the stack is not realignable, the alignment should be limited to the
2634 // StackAlignment
2635 RedAlign = std::min(RedAlign, StackAlign);
2636 }
2637
2638 return RedAlign;
2639}
2640
2642 MachineFrameInfo &MFI = MF->getFrameInfo();
2643 const TargetFrameLowering *TFI = MF->getSubtarget().getFrameLowering();
2644 int StackID = 0;
2645 if (Bytes.isScalable())
2646 StackID = TFI->getStackIDForScalableVectors();
2647 // The stack id gives an indication of whether the object is scalable or
2648 // not, so it's safe to pass in the minimum size here.
2649 int FrameIdx = MFI.CreateStackObject(Bytes.getKnownMinValue(), Alignment,
2650 false, nullptr, StackID);
2651 return getFrameIndex(FrameIdx, TLI->getFrameIndexTy(getDataLayout()));
2652}
2653
2655 Type *Ty = VT.getTypeForEVT(*getContext());
2656 Align StackAlign =
2657 std::max(getDataLayout().getPrefTypeAlign(Ty), Align(minAlign));
2658 return CreateStackTemporary(VT.getStoreSize(), StackAlign);
2659}
2660
2662 TypeSize VT1Size = VT1.getStoreSize();
2663 TypeSize VT2Size = VT2.getStoreSize();
2664 assert(VT1Size.isScalable() == VT2Size.isScalable() &&
2665 "Don't know how to choose the maximum size when creating a stack "
2666 "temporary");
2667 TypeSize Bytes = VT1Size.getKnownMinValue() > VT2Size.getKnownMinValue()
2668 ? VT1Size
2669 : VT2Size;
2670
2671 Type *Ty1 = VT1.getTypeForEVT(*getContext());
2672 Type *Ty2 = VT2.getTypeForEVT(*getContext());
2673 const DataLayout &DL = getDataLayout();
2674 Align Align = std::max(DL.getPrefTypeAlign(Ty1), DL.getPrefTypeAlign(Ty2));
2675 return CreateStackTemporary(Bytes, Align);
2676}
2677
2679 ISD::CondCode Cond, const SDLoc &dl,
2680 SDNodeFlags Flags) {
2681 EVT OpVT = N1.getValueType();
2682
2683 auto GetUndefBooleanConstant = [&]() {
2684 if (VT.getScalarType() == MVT::i1 ||
2685 TLI->getBooleanContents(OpVT) ==
2687 return getUNDEF(VT);
2688 // ZeroOrOne / ZeroOrNegative require specific values for the high bits,
2689 // so we cannot use getUNDEF(). Return zero instead.
2690 return getConstant(0, dl, VT);
2691 };
2692
2693 // These setcc operations always fold.
2694 switch (Cond) {
2695 default: break;
2696 case ISD::SETFALSE:
2697 case ISD::SETFALSE2: return getBoolConstant(false, dl, VT, OpVT);
2698 case ISD::SETTRUE:
2699 case ISD::SETTRUE2: return getBoolConstant(true, dl, VT, OpVT);
2700
2701 case ISD::SETOEQ:
2702 case ISD::SETOGT:
2703 case ISD::SETOGE:
2704 case ISD::SETOLT:
2705 case ISD::SETOLE:
2706 case ISD::SETONE:
2707 case ISD::SETO:
2708 case ISD::SETUO:
2709 case ISD::SETUEQ:
2710 case ISD::SETUNE:
2711 assert(!OpVT.isInteger() && "Illegal setcc for integer!");
2712 break;
2713 }
2714
2715 if (OpVT.isInteger()) {
2716 // For EQ and NE, we can always pick a value for the undef to make the
2717 // predicate pass or fail, so we can return undef.
2718 // Matches behavior in llvm::ConstantFoldCompareInstruction.
2719 // icmp eq/ne X, undef -> undef.
2720 if ((N1.isUndef() || N2.isUndef()) &&
2721 (Cond == ISD::SETEQ || Cond == ISD::SETNE))
2722 return GetUndefBooleanConstant();
2723
2724 // If both operands are undef, we can return undef for int comparison.
2725 // icmp undef, undef -> undef.
2726 if (N1.isUndef() && N2.isUndef())
2727 return GetUndefBooleanConstant();
2728
2729 // icmp X, X -> true/false
2730 // icmp X, undef -> true/false because undef could be X.
2731 if (N1.isUndef() || N2.isUndef() || N1 == N2)
2732 return getBoolConstant(ISD::isTrueWhenEqual(Cond), dl, VT, OpVT);
2733 }
2734
2736 const APInt &C2 = N2C->getAPIntValue();
2738 const APInt &C1 = N1C->getAPIntValue();
2739
2741 dl, VT, OpVT);
2742 }
2743 }
2744
2745 auto *N1CFP = dyn_cast<ConstantFPSDNode>(N1);
2746 auto *N2CFP = dyn_cast<ConstantFPSDNode>(N2);
2747
2748 if (N1CFP && N2CFP) {
2749 APFloat::cmpResult R = N1CFP->getValueAPF().compare(N2CFP->getValueAPF());
2750 switch (Cond) {
2751 default: break;
2752 case ISD::SETEQ: if (R==APFloat::cmpUnordered)
2753 return GetUndefBooleanConstant();
2754 [[fallthrough]];
2755 case ISD::SETOEQ: return getBoolConstant(R==APFloat::cmpEqual, dl, VT,
2756 OpVT);
2757 case ISD::SETNE: if (R==APFloat::cmpUnordered)
2758 return GetUndefBooleanConstant();
2759 [[fallthrough]];
2761 R==APFloat::cmpLessThan, dl, VT,
2762 OpVT);
2763 case ISD::SETLT: if (R==APFloat::cmpUnordered)
2764 return GetUndefBooleanConstant();
2765 [[fallthrough]];
2766 case ISD::SETOLT: return getBoolConstant(R==APFloat::cmpLessThan, dl, VT,
2767 OpVT);
2768 case ISD::SETGT: if (R==APFloat::cmpUnordered)
2769 return GetUndefBooleanConstant();
2770 [[fallthrough]];
2772 VT, OpVT);
2773 case ISD::SETLE: if (R==APFloat::cmpUnordered)
2774 return GetUndefBooleanConstant();
2775 [[fallthrough]];
2777 R==APFloat::cmpEqual, dl, VT,
2778 OpVT);
2779 case ISD::SETGE: if (R==APFloat::cmpUnordered)
2780 return GetUndefBooleanConstant();
2781 [[fallthrough]];
2783 R==APFloat::cmpEqual, dl, VT, OpVT);
2784 case ISD::SETO: return getBoolConstant(R!=APFloat::cmpUnordered, dl, VT,
2785 OpVT);
2786 case ISD::SETUO: return getBoolConstant(R==APFloat::cmpUnordered, dl, VT,
2787 OpVT);
2789 R==APFloat::cmpEqual, dl, VT,
2790 OpVT);
2791 case ISD::SETUNE: return getBoolConstant(R!=APFloat::cmpEqual, dl, VT,
2792 OpVT);
2794 R==APFloat::cmpLessThan, dl, VT,
2795 OpVT);
2797 R==APFloat::cmpUnordered, dl, VT,
2798 OpVT);
2800 VT, OpVT);
2801 case ISD::SETUGE: return getBoolConstant(R!=APFloat::cmpLessThan, dl, VT,
2802 OpVT);
2803 }
2804 } else if (N1CFP && OpVT.isSimple() && !N2.isUndef()) {
2805 // Ensure that the constant occurs on the RHS.
2807 if (!TLI->isCondCodeLegal(SwappedCond, OpVT.getSimpleVT()))
2808 return SDValue();
2809 return getSetCC(dl, VT, N2, N1, SwappedCond, /*Chian=*/{},
2810 /*IsSignaling=*/false, Flags);
2811 } else if ((N2CFP && N2CFP->getValueAPF().isNaN()) ||
2812 (OpVT.isFloatingPoint() && (N1.isUndef() || N2.isUndef()))) {
2813 // If an operand is known to be a nan (or undef that could be a nan), we can
2814 // fold it.
2815 // Choosing NaN for the undef will always make unordered comparison succeed
2816 // and ordered comparison fails.
2817 // Matches behavior in llvm::ConstantFoldCompareInstruction.
2818 switch (ISD::getUnorderedFlavor(Cond)) {
2819 default:
2820 llvm_unreachable("Unknown flavor!");
2821 case 0: // Known false.
2822 return getBoolConstant(false, dl, VT, OpVT);
2823 case 1: // Known true.
2824 return getBoolConstant(true, dl, VT, OpVT);
2825 case 2: // Undefined.
2826 return GetUndefBooleanConstant();
2827 }
2828 }
2829
2830 // Could not fold it.
2831 return SDValue();
2832}
2833
2834/// SignBitIsZero - Return true if the sign bit of Op is known to be zero. We
2835/// use this predicate to simplify operations downstream.
2837 unsigned BitWidth = Op.getScalarValueSizeInBits();
2839}
2840
2841// TODO: Should have argument to specify if sign bit of nan is ignorable.
2843 if (Depth >= MaxRecursionDepth)
2844 return false; // Limit search depth.
2845
2846 unsigned Opc = Op.getOpcode();
2847 switch (Opc) {
2848 case ISD::FABS:
2849 return true;
2850 case ISD::AssertNoFPClass: {
2851 FPClassTest NoFPClass =
2852 static_cast<FPClassTest>(Op.getConstantOperandVal(1));
2853
2854 const FPClassTest TestMask = fcNan | fcNegative;
2855 return (NoFPClass & TestMask) == TestMask;
2856 }
2857 case ISD::ARITH_FENCE:
2858 return SignBitIsZeroFP(Op.getOperand(0), Depth + 1);
2859 case ISD::FEXP:
2860 case ISD::FEXP2:
2861 case ISD::FEXP10:
2862 return Op->getFlags().hasNoNaNs();
2863 case ISD::FMINNUM:
2864 case ISD::FMINNUM_IEEE:
2865 case ISD::FMINIMUM:
2866 case ISD::FMINIMUMNUM:
2867 return SignBitIsZeroFP(Op.getOperand(1), Depth + 1) &&
2868 SignBitIsZeroFP(Op.getOperand(0), Depth + 1);
2869 case ISD::FMAXNUM:
2870 case ISD::FMAXNUM_IEEE:
2871 case ISD::FMAXIMUM:
2872 case ISD::FMAXIMUMNUM:
2873 // TODO: If we can ignore the sign bit of nans, only one side being known 0
2874 // is sufficient.
2875 return SignBitIsZeroFP(Op.getOperand(1), Depth + 1) &&
2876 SignBitIsZeroFP(Op.getOperand(0), Depth + 1);
2877 default:
2878 return false;
2879 }
2880
2881 llvm_unreachable("covered opcode switch");
2882}
2883
2884/// MaskedValueIsZero - Return true if 'V & Mask' is known to be zero. We use
2885/// this predicate to simplify operations downstream. Mask is known to be zero
2886/// for bits that V cannot have.
2888 unsigned Depth) const {
2889 return Mask.isSubsetOf(computeKnownBits(V, Depth).Zero);
2890}
2891
2892/// MaskedValueIsZero - Return true if 'V & Mask' is known to be zero in
2893/// DemandedElts. We use this predicate to simplify operations downstream.
2894/// Mask is known to be zero for bits that V cannot have.
2896 const APInt &DemandedElts,
2897 unsigned Depth) const {
2898 return Mask.isSubsetOf(computeKnownBits(V, DemandedElts, Depth).Zero);
2899}
2900
2901/// MaskedVectorIsZero - Return true if 'Op' is known to be zero in
2902/// DemandedElts. We use this predicate to simplify operations downstream.
2904 unsigned Depth /* = 0 */) const {
2905 return computeKnownBits(V, DemandedElts, Depth).isZero();
2906}
2907
2908/// MaskedValueIsAllOnes - Return true if '(Op & Mask) == Mask'.
2910 unsigned Depth) const {
2911 return Mask.isSubsetOf(computeKnownBits(V, Depth).One);
2912}
2913
2915 const APInt &DemandedElts,
2916 unsigned Depth) const {
2917 EVT VT = Op.getValueType();
2918 assert(VT.isVector() && !VT.isScalableVector() && "Only for fixed vectors!");
2919
2920 unsigned NumElts = VT.getVectorNumElements();
2921 assert(DemandedElts.getBitWidth() == NumElts && "Unexpected demanded mask.");
2922
2923 APInt KnownZeroElements = APInt::getZero(NumElts);
2924 for (unsigned EltIdx = 0; EltIdx != NumElts; ++EltIdx) {
2925 if (!DemandedElts[EltIdx])
2926 continue; // Don't query elements that are not demanded.
2927 APInt Mask = APInt::getOneBitSet(NumElts, EltIdx);
2928 if (MaskedVectorIsZero(Op, Mask, Depth))
2929 KnownZeroElements.setBit(EltIdx);
2930 }
2931 return KnownZeroElements;
2932}
2933
2934/// isSplatValue - Return true if the vector V has the same value
2935/// across all DemandedElts. For scalable vectors, we don't know the
2936/// number of lanes at compile time. Instead, we use a 1 bit APInt
2937/// to represent a conservative value for all lanes; that is, that
2938/// one bit value is implicitly splatted across all lanes.
2939bool SelectionDAG::isSplatValue(SDValue V, const APInt &DemandedElts,
2940 APInt &UndefElts, unsigned Depth) const {
2941 unsigned Opcode = V.getOpcode();
2942 EVT VT = V.getValueType();
2943 assert(VT.isVector() && "Vector type expected");
2944 assert((!VT.isScalableVector() || DemandedElts.getBitWidth() == 1) &&
2945 "scalable demanded bits are ignored");
2946
2947 if (!DemandedElts)
2948 return false; // No demanded elts, better to assume we don't know anything.
2949
2950 if (Depth >= MaxRecursionDepth)
2951 return false; // Limit search depth.
2952
2953 // Deal with some common cases here that work for both fixed and scalable
2954 // vector types.
2955 switch (Opcode) {
2956 case ISD::SPLAT_VECTOR:
2957 UndefElts = V.getOperand(0).isUndef()
2958 ? APInt::getAllOnes(DemandedElts.getBitWidth())
2959 : APInt(DemandedElts.getBitWidth(), 0);
2960 return true;
2961 case ISD::ADD:
2962 case ISD::SUB:
2963 case ISD::AND:
2964 case ISD::XOR:
2965 case ISD::OR: {
2966 APInt UndefLHS, UndefRHS;
2967 SDValue LHS = V.getOperand(0);
2968 SDValue RHS = V.getOperand(1);
2969 // Only recognize splats with the same demanded undef elements for both
2970 // operands, otherwise we might fail to handle binop-specific undef
2971 // handling.
2972 // e.g. (and undef, 0) -> 0 etc.
2973 if (isSplatValue(LHS, DemandedElts, UndefLHS, Depth + 1) &&
2974 isSplatValue(RHS, DemandedElts, UndefRHS, Depth + 1) &&
2975 (DemandedElts & UndefLHS) == (DemandedElts & UndefRHS)) {
2976 UndefElts = UndefLHS | UndefRHS;
2977 return true;
2978 }
2979 return false;
2980 }
2981 case ISD::ABS:
2982 case ISD::TRUNCATE:
2983 case ISD::SIGN_EXTEND:
2984 case ISD::ZERO_EXTEND:
2985 return isSplatValue(V.getOperand(0), DemandedElts, UndefElts, Depth + 1);
2986 default:
2987 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
2988 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID)
2989 return TLI->isSplatValueForTargetNode(V, DemandedElts, UndefElts, *this,
2990 Depth);
2991 break;
2992 }
2993
2994 // We don't support other cases than those above for scalable vectors at
2995 // the moment.
2996 if (VT.isScalableVector())
2997 return false;
2998
2999 unsigned NumElts = VT.getVectorNumElements();
3000 assert(NumElts == DemandedElts.getBitWidth() && "Vector size mismatch");
3001 UndefElts = APInt::getZero(NumElts);
3002
3003 switch (Opcode) {
3004 case ISD::BUILD_VECTOR: {
3005 SDValue Scl;
3006 for (unsigned i = 0; i != NumElts; ++i) {
3007 SDValue Op = V.getOperand(i);
3008 if (Op.isUndef()) {
3009 UndefElts.setBit(i);
3010 continue;
3011 }
3012 if (!DemandedElts[i])
3013 continue;
3014 if (Scl && Scl != Op)
3015 return false;
3016 Scl = Op;
3017 }
3018 return true;
3019 }
3020 case ISD::VECTOR_SHUFFLE: {
3021 // Check if this is a shuffle node doing a splat or a shuffle of a splat.
3022 APInt DemandedLHS = APInt::getZero(NumElts);
3023 APInt DemandedRHS = APInt::getZero(NumElts);
3024 ArrayRef<int> Mask = cast<ShuffleVectorSDNode>(V)->getMask();
3025 for (int i = 0; i != (int)NumElts; ++i) {
3026 int M = Mask[i];
3027 if (M < 0) {
3028 UndefElts.setBit(i);
3029 continue;
3030 }
3031 if (!DemandedElts[i])
3032 continue;
3033 if (M < (int)NumElts)
3034 DemandedLHS.setBit(M);
3035 else
3036 DemandedRHS.setBit(M - NumElts);
3037 }
3038
3039 // If we aren't demanding either op, assume there's no splat.
3040 // If we are demanding both ops, assume there's no splat.
3041 if ((DemandedLHS.isZero() && DemandedRHS.isZero()) ||
3042 (!DemandedLHS.isZero() && !DemandedRHS.isZero()))
3043 return false;
3044
3045 // See if the demanded elts of the source op is a splat or we only demand
3046 // one element, which should always be a splat.
3047 // TODO: Handle source ops splats with undefs.
3048 auto CheckSplatSrc = [&](SDValue Src, const APInt &SrcElts) {
3049 APInt SrcUndefs;
3050 return (SrcElts.popcount() == 1) ||
3051 (isSplatValue(Src, SrcElts, SrcUndefs, Depth + 1) &&
3052 (SrcElts & SrcUndefs).isZero());
3053 };
3054 if (!DemandedLHS.isZero())
3055 return CheckSplatSrc(V.getOperand(0), DemandedLHS);
3056 return CheckSplatSrc(V.getOperand(1), DemandedRHS);
3057 }
3059 // Offset the demanded elts by the subvector index.
3060 SDValue Src = V.getOperand(0);
3061 // We don't support scalable vectors at the moment.
3062 if (Src.getValueType().isScalableVector())
3063 return false;
3064 uint64_t Idx = V.getConstantOperandVal(1);
3065 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
3066 APInt UndefSrcElts;
3067 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
3068 if (isSplatValue(Src, DemandedSrcElts, UndefSrcElts, Depth + 1)) {
3069 UndefElts = UndefSrcElts.extractBits(NumElts, Idx);
3070 return true;
3071 }
3072 break;
3073 }
3077 // Widen the demanded elts by the src element count.
3078 SDValue Src = V.getOperand(0);
3079 // We don't support scalable vectors at the moment.
3080 if (Src.getValueType().isScalableVector())
3081 return false;
3082 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
3083 APInt UndefSrcElts;
3084 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts);
3085 if (isSplatValue(Src, DemandedSrcElts, UndefSrcElts, Depth + 1)) {
3086 UndefElts = UndefSrcElts.trunc(NumElts);
3087 return true;
3088 }
3089 break;
3090 }
3091 case ISD::BITCAST: {
3092 SDValue Src = V.getOperand(0);
3093 EVT SrcVT = Src.getValueType();
3094 unsigned SrcBitWidth = SrcVT.getScalarSizeInBits();
3095 unsigned BitWidth = VT.getScalarSizeInBits();
3096
3097 // Ignore bitcasts from unsupported types.
3098 // TODO: Add fp support?
3099 if (!SrcVT.isVector() || !SrcVT.isInteger() || !VT.isInteger())
3100 break;
3101
3102 // Bitcast 'small element' vector to 'large element' vector.
3103 if ((BitWidth % SrcBitWidth) == 0) {
3104 // See if each sub element is a splat.
3105 unsigned Scale = BitWidth / SrcBitWidth;
3106 unsigned NumSrcElts = SrcVT.getVectorNumElements();
3107 APInt ScaledDemandedElts =
3108 APIntOps::ScaleBitMask(DemandedElts, NumSrcElts);
3109 for (unsigned I = 0; I != Scale; ++I) {
3110 APInt SubUndefElts;
3111 APInt SubDemandedElt = APInt::getOneBitSet(Scale, I);
3112 APInt SubDemandedElts = APInt::getSplat(NumSrcElts, SubDemandedElt);
3113 SubDemandedElts &= ScaledDemandedElts;
3114 if (!isSplatValue(Src, SubDemandedElts, SubUndefElts, Depth + 1))
3115 return false;
3116 // TODO: Add support for merging sub undef elements.
3117 if (!SubUndefElts.isZero())
3118 return false;
3119 }
3120 return true;
3121 }
3122 break;
3123 }
3124 }
3125
3126 return false;
3127}
3128
3129/// Helper wrapper to main isSplatValue function.
3130bool SelectionDAG::isSplatValue(SDValue V, bool AllowUndefs) const {
3131 EVT VT = V.getValueType();
3132 assert(VT.isVector() && "Vector type expected");
3133
3134 APInt UndefElts;
3135 // Since the number of lanes in a scalable vector is unknown at compile time,
3136 // we track one bit which is implicitly broadcast to all lanes. This means
3137 // that all lanes in a scalable vector are considered demanded.
3138 APInt DemandedElts
3140 return isSplatValue(V, DemandedElts, UndefElts) &&
3141 (AllowUndefs || !UndefElts);
3142}
3143
3146
3147 EVT VT = V.getValueType();
3148 unsigned Opcode = V.getOpcode();
3149 switch (Opcode) {
3150 default: {
3151 APInt UndefElts;
3152 // Since the number of lanes in a scalable vector is unknown at compile time,
3153 // we track one bit which is implicitly broadcast to all lanes. This means
3154 // that all lanes in a scalable vector are considered demanded.
3155 APInt DemandedElts
3157
3158 if (isSplatValue(V, DemandedElts, UndefElts)) {
3159 if (VT.isScalableVector()) {
3160 // DemandedElts and UndefElts are ignored for scalable vectors, since
3161 // the only supported cases are SPLAT_VECTOR nodes.
3162 SplatIdx = 0;
3163 } else {
3164 // Handle case where all demanded elements are UNDEF.
3165 if (DemandedElts.isSubsetOf(UndefElts)) {
3166 SplatIdx = 0;
3167 return getUNDEF(VT);
3168 }
3169 SplatIdx = (UndefElts & DemandedElts).countr_one();
3170 }
3171 return V;
3172 }
3173 break;
3174 }
3175 case ISD::SPLAT_VECTOR:
3176 SplatIdx = 0;
3177 return V;
3178 case ISD::VECTOR_SHUFFLE: {
3179 assert(!VT.isScalableVector());
3180 // Check if this is a shuffle node doing a splat.
3181 // TODO - remove this and rely purely on SelectionDAG::isSplatValue,
3182 // getTargetVShiftNode currently struggles without the splat source.
3183 auto *SVN = cast<ShuffleVectorSDNode>(V);
3184 if (!SVN->isSplat())
3185 break;
3186 int Idx = SVN->getSplatIndex();
3187 int NumElts = V.getValueType().getVectorNumElements();
3188 SplatIdx = Idx % NumElts;
3189 return V.getOperand(Idx / NumElts);
3190 }
3191 }
3192
3193 return SDValue();
3194}
3195
3197 int SplatIdx;
3198 if (SDValue SrcVector = getSplatSourceVector(V, SplatIdx)) {
3199 EVT SVT = SrcVector.getValueType().getScalarType();
3200 EVT LegalSVT = SVT;
3201 if (LegalTypes && !TLI->isTypeLegal(SVT)) {
3202 if (!SVT.isInteger())
3203 return SDValue();
3204 LegalSVT = TLI->getTypeToTransformTo(*getContext(), LegalSVT);
3205 if (LegalSVT.bitsLT(SVT))
3206 return SDValue();
3207 }
3208 return getExtractVectorElt(SDLoc(V), LegalSVT, SrcVector, SplatIdx);
3209 }
3210 return SDValue();
3211}
3212
3213std::optional<ConstantRange>
3215 unsigned Depth) const {
3216 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3217 V.getOpcode() == ISD::SRA) &&
3218 "Unknown shift node");
3219 // Shifting more than the bitwidth is not valid.
3220 unsigned BitWidth = V.getScalarValueSizeInBits();
3221
3222 if (auto *Cst = dyn_cast<ConstantSDNode>(V.getOperand(1))) {
3223 const APInt &ShAmt = Cst->getAPIntValue();
3224 if (ShAmt.uge(BitWidth))
3225 return std::nullopt;
3226 return ConstantRange(ShAmt);
3227 }
3228
3229 if (auto *BV = dyn_cast<BuildVectorSDNode>(V.getOperand(1))) {
3230 const APInt *MinAmt = nullptr, *MaxAmt = nullptr;
3231 for (unsigned i = 0, e = BV->getNumOperands(); i != e; ++i) {
3232 if (!DemandedElts[i])
3233 continue;
3234 auto *SA = dyn_cast<ConstantSDNode>(BV->getOperand(i));
3235 if (!SA) {
3236 MinAmt = MaxAmt = nullptr;
3237 break;
3238 }
3239 const APInt &ShAmt = SA->getAPIntValue();
3240 if (ShAmt.uge(BitWidth))
3241 return std::nullopt;
3242 if (!MinAmt || MinAmt->ugt(ShAmt))
3243 MinAmt = &ShAmt;
3244 if (!MaxAmt || MaxAmt->ult(ShAmt))
3245 MaxAmt = &ShAmt;
3246 }
3247 assert(((!MinAmt && !MaxAmt) || (MinAmt && MaxAmt)) &&
3248 "Failed to find matching min/max shift amounts");
3249 if (MinAmt && MaxAmt)
3250 return ConstantRange(*MinAmt, *MaxAmt + 1);
3251 }
3252
3253 // Use computeKnownBits to find a hidden constant/knownbits (usually type
3254 // legalized). e.g. Hidden behind multiple bitcasts/build_vector/casts etc.
3255 KnownBits KnownAmt = computeKnownBits(V.getOperand(1), DemandedElts, Depth);
3256 if (KnownAmt.getMaxValue().ult(BitWidth))
3257 return ConstantRange::fromKnownBits(KnownAmt, /*IsSigned=*/false);
3258
3259 return std::nullopt;
3260}
3261
3262std::optional<unsigned>
3264 unsigned Depth) const {
3265 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3266 V.getOpcode() == ISD::SRA) &&
3267 "Unknown shift node");
3268 if (std::optional<ConstantRange> AmtRange =
3269 getValidShiftAmountRange(V, DemandedElts, Depth))
3270 if (const APInt *ShAmt = AmtRange->getSingleElement())
3271 return ShAmt->getZExtValue();
3272 return std::nullopt;
3273}
3274
3275std::optional<unsigned>
3277 EVT VT = V.getValueType();
3278 APInt DemandedElts = VT.isFixedLengthVector()
3280 : APInt(1, 1);
3281 return getValidShiftAmount(V, DemandedElts, Depth);
3282}
3283
3284std::optional<unsigned>
3286 unsigned Depth) const {
3287 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3288 V.getOpcode() == ISD::SRA) &&
3289 "Unknown shift node");
3290 if (std::optional<ConstantRange> AmtRange =
3291 getValidShiftAmountRange(V, DemandedElts, Depth))
3292 return AmtRange->getUnsignedMin().getZExtValue();
3293 return std::nullopt;
3294}
3295
3296std::optional<unsigned>
3298 EVT VT = V.getValueType();
3299 APInt DemandedElts = VT.isFixedLengthVector()
3301 : APInt(1, 1);
3302 return getValidMinimumShiftAmount(V, DemandedElts, Depth);
3303}
3304
3305std::optional<unsigned>
3307 unsigned Depth) const {
3308 assert((V.getOpcode() == ISD::SHL || V.getOpcode() == ISD::SRL ||
3309 V.getOpcode() == ISD::SRA) &&
3310 "Unknown shift node");
3311 if (std::optional<ConstantRange> AmtRange =
3312 getValidShiftAmountRange(V, DemandedElts, Depth))
3313 return AmtRange->getUnsignedMax().getZExtValue();
3314 return std::nullopt;
3315}
3316
3317std::optional<unsigned>
3319 EVT VT = V.getValueType();
3320 APInt DemandedElts = VT.isFixedLengthVector()
3322 : APInt(1, 1);
3323 return getValidMaximumShiftAmount(V, DemandedElts, Depth);
3324}
3325
3326/// Determine which bits of Op are known to be either zero or one and return
3327/// them in Known. For vectors, the known bits are those that are shared by
3328/// every vector element.
3330 EVT VT = Op.getValueType();
3331
3332 // Since the number of lanes in a scalable vector is unknown at compile time,
3333 // we track one bit which is implicitly broadcast to all lanes. This means
3334 // that all lanes in a scalable vector are considered demanded.
3335 APInt DemandedElts = VT.isFixedLengthVector()
3337 : APInt(1, 1);
3338 return computeKnownBits(Op, DemandedElts, Depth);
3339}
3340
3341/// Determine which bits of Op are known to be either zero or one and return
3342/// them in Known. The DemandedElts argument allows us to only collect the known
3343/// bits that are shared by the requested vector elements.
3345 unsigned Depth) const {
3346 unsigned BitWidth = Op.getScalarValueSizeInBits();
3347
3348 KnownBits Known(BitWidth); // Don't know anything.
3349
3350 if (auto OptAPInt = Op->bitcastToAPInt()) {
3351 // We know all of the bits for a constant!
3352 return KnownBits::makeConstant(*std::move(OptAPInt));
3353 }
3354
3355 if (Depth >= MaxRecursionDepth)
3356 return Known; // Limit search depth.
3357
3358 KnownBits Known2;
3359 unsigned NumElts = DemandedElts.getBitWidth();
3360 assert((!Op.getValueType().isScalableVector() || NumElts == 1) &&
3361 "DemandedElts for scalable vectors must be 1 to represent all lanes");
3362 assert((!Op.getValueType().isFixedLengthVector() ||
3363 NumElts == Op.getValueType().getVectorNumElements()) &&
3364 "Unexpected vector size");
3365
3366 if (!DemandedElts)
3367 return Known; // No demanded elts, better to assume we don't know anything.
3368
3369 unsigned Opcode = Op.getOpcode();
3370 switch (Opcode) {
3371 case ISD::MERGE_VALUES:
3372 return computeKnownBits(Op.getOperand(Op.getResNo()), DemandedElts,
3373 Depth + 1);
3374 case ISD::SPLAT_VECTOR: {
3375 SDValue SrcOp = Op.getOperand(0);
3376 assert(SrcOp.getValueSizeInBits() >= BitWidth &&
3377 "Expected SPLAT_VECTOR implicit truncation");
3378 // Implicitly truncate the bits to match the official semantics of
3379 // SPLAT_VECTOR.
3380 Known = computeKnownBits(SrcOp, Depth + 1).trunc(BitWidth);
3381 break;
3382 }
3384 unsigned ScalarSize = Op.getOperand(0).getScalarValueSizeInBits();
3385 assert(ScalarSize * Op.getNumOperands() == BitWidth &&
3386 "Expected SPLAT_VECTOR_PARTS scalars to cover element width");
3387 for (auto [I, SrcOp] : enumerate(Op->ops())) {
3388 Known.insertBits(computeKnownBits(SrcOp, Depth + 1), ScalarSize * I);
3389 }
3390 break;
3391 }
3392 case ISD::STEP_VECTOR: {
3393 const APInt &Step = Op.getConstantOperandAPInt(0);
3394
3395 if (Step.isPowerOf2())
3396 Known.Zero.setLowBits(Step.logBase2());
3397
3399
3400 if (!isUIntN(BitWidth, Op.getValueType().getVectorMinNumElements()))
3401 break;
3402 const APInt MinNumElts =
3403 APInt(BitWidth, Op.getValueType().getVectorMinNumElements());
3404
3405 bool Overflow;
3406 const APInt MaxNumElts = getVScaleRange(&F, BitWidth)
3408 .umul_ov(MinNumElts, Overflow);
3409 if (Overflow)
3410 break;
3411
3412 const APInt MaxValue = (MaxNumElts - 1).umul_ov(Step, Overflow);
3413 if (Overflow)
3414 break;
3415
3416 Known.Zero.setHighBits(MaxValue.countl_zero());
3417 break;
3418 }
3419 case ISD::BUILD_VECTOR:
3420 assert(!Op.getValueType().isScalableVector());
3421 // Collect the known bits that are shared by every demanded vector element.
3422 Known.setAllConflict();
3423 for (unsigned i = 0, e = Op.getNumOperands(); i != e; ++i) {
3424 if (!DemandedElts[i])
3425 continue;
3426
3427 SDValue SrcOp = Op.getOperand(i);
3428 Known2 = computeKnownBits(SrcOp, Depth + 1);
3429
3430 // BUILD_VECTOR can implicitly truncate sources, we must handle this.
3431 if (SrcOp.getValueSizeInBits() != BitWidth) {
3432 assert(SrcOp.getValueSizeInBits() > BitWidth &&
3433 "Expected BUILD_VECTOR implicit truncation");
3434 Known2 = Known2.trunc(BitWidth);
3435 }
3436
3437 // Known bits are the values that are shared by every demanded element.
3438 Known = Known.intersectWith(Known2);
3439
3440 // If we don't know any bits, early out.
3441 if (Known.isUnknown())
3442 break;
3443 }
3444 break;
3445 case ISD::VECTOR_COMPRESS: {
3446 SDValue Vec = Op.getOperand(0);
3447 SDValue PassThru = Op.getOperand(2);
3448 Known = computeKnownBits(PassThru, DemandedElts, Depth + 1);
3449 // If we don't know any bits, early out.
3450 if (Known.isUnknown())
3451 break;
3452 Known2 = computeKnownBits(Vec, Depth + 1);
3453 Known = Known.intersectWith(Known2);
3454 break;
3455 }
3456 case ISD::VECTOR_SHUFFLE: {
3457 assert(!Op.getValueType().isScalableVector());
3458 // Collect the known bits that are shared by every vector element referenced
3459 // by the shuffle.
3460 APInt DemandedLHS, DemandedRHS;
3462 assert(NumElts == SVN->getMask().size() && "Unexpected vector size");
3463 if (!getShuffleDemandedElts(NumElts, SVN->getMask(), DemandedElts,
3464 DemandedLHS, DemandedRHS))
3465 break;
3466
3467 // Known bits are the values that are shared by every demanded element.
3468 Known.setAllConflict();
3469 if (!!DemandedLHS) {
3470 SDValue LHS = Op.getOperand(0);
3471 Known2 = computeKnownBits(LHS, DemandedLHS, Depth + 1);
3472 Known = Known.intersectWith(Known2);
3473 }
3474 // If we don't know any bits, early out.
3475 if (Known.isUnknown())
3476 break;
3477 if (!!DemandedRHS) {
3478 SDValue RHS = Op.getOperand(1);
3479 Known2 = computeKnownBits(RHS, DemandedRHS, Depth + 1);
3480 Known = Known.intersectWith(Known2);
3481 }
3482 break;
3483 }
3484 case ISD::VSCALE: {
3486 const APInt &Multiplier = Op.getConstantOperandAPInt(0);
3487 Known = getVScaleRange(&F, BitWidth).multiply(Multiplier).toKnownBits();
3488 break;
3489 }
3490 case ISD::CONCAT_VECTORS: {
3491 if (Op.getValueType().isScalableVector())
3492 break;
3493 // Split DemandedElts and test each of the demanded subvectors.
3494 Known.setAllConflict();
3495 EVT SubVectorVT = Op.getOperand(0).getValueType();
3496 unsigned NumSubVectorElts = SubVectorVT.getVectorNumElements();
3497 unsigned NumSubVectors = Op.getNumOperands();
3498 for (unsigned i = 0; i != NumSubVectors; ++i) {
3499 APInt DemandedSub =
3500 DemandedElts.extractBits(NumSubVectorElts, i * NumSubVectorElts);
3501 if (!!DemandedSub) {
3502 SDValue Sub = Op.getOperand(i);
3503 Known2 = computeKnownBits(Sub, DemandedSub, Depth + 1);
3504 Known = Known.intersectWith(Known2);
3505 }
3506 // If we don't know any bits, early out.
3507 if (Known.isUnknown())
3508 break;
3509 }
3510 break;
3511 }
3512 case ISD::INSERT_SUBVECTOR: {
3513 if (Op.getValueType().isScalableVector())
3514 break;
3515 // Demand any elements from the subvector and the remainder from the src its
3516 // inserted into.
3517 SDValue Src = Op.getOperand(0);
3518 SDValue Sub = Op.getOperand(1);
3519 uint64_t Idx = Op.getConstantOperandVal(2);
3520 unsigned NumSubElts = Sub.getValueType().getVectorNumElements();
3521 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
3522 APInt DemandedSrcElts = DemandedElts;
3523 DemandedSrcElts.clearBits(Idx, Idx + NumSubElts);
3524
3525 Known.setAllConflict();
3526 if (!!DemandedSubElts) {
3527 Known = computeKnownBits(Sub, DemandedSubElts, Depth + 1);
3528 if (Known.isUnknown())
3529 break; // early-out.
3530 }
3531 if (!!DemandedSrcElts) {
3532 Known2 = computeKnownBits(Src, DemandedSrcElts, Depth + 1);
3533 Known = Known.intersectWith(Known2);
3534 }
3535 break;
3536 }
3538 // Offset the demanded elts by the subvector index.
3539 SDValue Src = Op.getOperand(0);
3540
3541 APInt DemandedSrcElts;
3542 if (Src.getValueType().isScalableVector())
3543 DemandedSrcElts = APInt(1, 1); // <=> 'demand all elements'
3544 else {
3545 uint64_t Idx = Op.getConstantOperandVal(1);
3546 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
3547 DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
3548 }
3549 Known = computeKnownBits(Src, DemandedSrcElts, Depth + 1);
3550 break;
3551 }
3552 case ISD::SCALAR_TO_VECTOR: {
3553 if (Op.getValueType().isScalableVector())
3554 break;
3555 // We know about scalar_to_vector as much as we know about it source,
3556 // which becomes the first element of otherwise unknown vector.
3557 if (DemandedElts != 1)
3558 break;
3559
3560 SDValue N0 = Op.getOperand(0);
3561 Known = computeKnownBits(N0, Depth + 1);
3562 if (N0.getValueSizeInBits() != BitWidth)
3563 Known = Known.trunc(BitWidth);
3564
3565 break;
3566 }
3567 case ISD::BITCAST: {
3568 if (Op.getValueType().isScalableVector())
3569 break;
3570
3571 SDValue N0 = Op.getOperand(0);
3572 EVT SubVT = N0.getValueType();
3573 unsigned SubBitWidth = SubVT.getScalarSizeInBits();
3574
3575 // Ignore bitcasts from unsupported types.
3576 if (!(SubVT.isInteger() || SubVT.isFloatingPoint()))
3577 break;
3578
3579 // Fast handling of 'identity' bitcasts.
3580 if (BitWidth == SubBitWidth) {
3581 Known = computeKnownBits(N0, DemandedElts, Depth + 1);
3582 break;
3583 }
3584
3585 bool IsLE = getDataLayout().isLittleEndian();
3586
3587 // Bitcast 'small element' vector to 'large element' scalar/vector.
3588 if ((BitWidth % SubBitWidth) == 0) {
3589 assert(N0.getValueType().isVector() && "Expected bitcast from vector");
3590
3591 // Collect known bits for the (larger) output by collecting the known
3592 // bits from each set of sub elements and shift these into place.
3593 // We need to separately call computeKnownBits for each set of
3594 // sub elements as the knownbits for each is likely to be different.
3595 unsigned SubScale = BitWidth / SubBitWidth;
3596 APInt SubDemandedElts(NumElts * SubScale, 0);
3597 for (unsigned i = 0; i != NumElts; ++i)
3598 if (DemandedElts[i])
3599 SubDemandedElts.setBit(i * SubScale);
3600
3601 for (unsigned i = 0; i != SubScale; ++i) {
3602 Known2 = computeKnownBits(N0, SubDemandedElts.shl(i),
3603 Depth + 1);
3604 unsigned Shifts = IsLE ? i : SubScale - 1 - i;
3605 Known.insertBits(Known2, SubBitWidth * Shifts);
3606 }
3607 }
3608
3609 // Bitcast 'large element' scalar/vector to 'small element' vector.
3610 if ((SubBitWidth % BitWidth) == 0) {
3611 assert(Op.getValueType().isVector() && "Expected bitcast to vector");
3612
3613 // Collect known bits for the (smaller) output by collecting the known
3614 // bits from the overlapping larger input elements and extracting the
3615 // sub sections we actually care about.
3616 unsigned SubScale = SubBitWidth / BitWidth;
3617 APInt SubDemandedElts =
3618 APIntOps::ScaleBitMask(DemandedElts, NumElts / SubScale);
3619 Known2 = computeKnownBits(N0, SubDemandedElts, Depth + 1);
3620
3621 Known.setAllConflict();
3622 for (unsigned i = 0; i != NumElts; ++i)
3623 if (DemandedElts[i]) {
3624 unsigned Shifts = IsLE ? i : NumElts - 1 - i;
3625 unsigned Offset = (Shifts % SubScale) * BitWidth;
3626 Known = Known.intersectWith(Known2.extractBits(BitWidth, Offset));
3627 // If we don't know any bits, early out.
3628 if (Known.isUnknown())
3629 break;
3630 }
3631 }
3632 break;
3633 }
3634 case ISD::AND:
3635 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3636 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3637
3638 Known &= Known2;
3639 break;
3640 case ISD::OR:
3641 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3642 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3643
3644 Known |= Known2;
3645 break;
3646 case ISD::XOR:
3647 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3648 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3649
3650 Known ^= Known2;
3651 break;
3652 case ISD::MUL: {
3653 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3654 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3655 bool SelfMultiply = Op.getOperand(0) == Op.getOperand(1);
3656 // TODO: SelfMultiply can be poison, but not undef.
3657 if (SelfMultiply)
3658 SelfMultiply &= isGuaranteedNotToBeUndefOrPoison(
3659 Op.getOperand(0), DemandedElts, false, Depth + 1);
3660 Known = KnownBits::mul(Known, Known2, SelfMultiply);
3661
3662 // If the multiplication is known not to overflow, the product of a number
3663 // with itself is non-negative. Only do this if we didn't already computed
3664 // the opposite value for the sign bit.
3665 if (Op->getFlags().hasNoSignedWrap() &&
3666 Op.getOperand(0) == Op.getOperand(1) &&
3667 !Known.isNegative())
3668 Known.makeNonNegative();
3669 break;
3670 }
3671 case ISD::MULHU: {
3672 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3673 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3674 Known = KnownBits::mulhu(Known, Known2);
3675 break;
3676 }
3677 case ISD::MULHS: {
3678 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3679 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3680 Known = KnownBits::mulhs(Known, Known2);
3681 break;
3682 }
3683 case ISD::ABDU: {
3684 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3685 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3686 Known = KnownBits::abdu(Known, Known2);
3687 break;
3688 }
3689 case ISD::ABDS: {
3690 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3691 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3692 Known = KnownBits::abds(Known, Known2);
3693 unsigned SignBits1 =
3694 ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
3695 if (SignBits1 == 1)
3696 break;
3697 unsigned SignBits0 =
3698 ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
3699 Known.Zero.setHighBits(std::min(SignBits0, SignBits1) - 1);
3700 break;
3701 }
3702 case ISD::UMUL_LOHI: {
3703 assert((Op.getResNo() == 0 || Op.getResNo() == 1) && "Unknown result");
3704 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3705 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3706 bool SelfMultiply = Op.getOperand(0) == Op.getOperand(1);
3707 if (Op.getResNo() == 0)
3708 Known = KnownBits::mul(Known, Known2, SelfMultiply);
3709 else
3710 Known = KnownBits::mulhu(Known, Known2);
3711 break;
3712 }
3713 case ISD::SMUL_LOHI: {
3714 assert((Op.getResNo() == 0 || Op.getResNo() == 1) && "Unknown result");
3715 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3716 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3717 bool SelfMultiply = Op.getOperand(0) == Op.getOperand(1);
3718 if (Op.getResNo() == 0)
3719 Known = KnownBits::mul(Known, Known2, SelfMultiply);
3720 else
3721 Known = KnownBits::mulhs(Known, Known2);
3722 break;
3723 }
3724 case ISD::AVGFLOORU: {
3725 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3726 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3727 Known = KnownBits::avgFloorU(Known, Known2);
3728 break;
3729 }
3730 case ISD::AVGCEILU: {
3731 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3732 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3733 Known = KnownBits::avgCeilU(Known, Known2);
3734 break;
3735 }
3736 case ISD::AVGFLOORS: {
3737 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3738 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3739 Known = KnownBits::avgFloorS(Known, Known2);
3740 break;
3741 }
3742 case ISD::AVGCEILS: {
3743 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3744 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3745 Known = KnownBits::avgCeilS(Known, Known2);
3746 break;
3747 }
3748 case ISD::SELECT:
3749 case ISD::VSELECT:
3750 Known = computeKnownBits(Op.getOperand(2), DemandedElts, Depth+1);
3751 // If we don't know any bits, early out.
3752 if (Known.isUnknown())
3753 break;
3754 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth+1);
3755
3756 // Only known if known in both the LHS and RHS.
3757 Known = Known.intersectWith(Known2);
3758 break;
3759 case ISD::SELECT_CC:
3760 Known = computeKnownBits(Op.getOperand(3), DemandedElts, Depth+1);
3761 // If we don't know any bits, early out.
3762 if (Known.isUnknown())
3763 break;
3764 Known2 = computeKnownBits(Op.getOperand(2), DemandedElts, Depth+1);
3765
3766 // Only known if known in both the LHS and RHS.
3767 Known = Known.intersectWith(Known2);
3768 break;
3769 case ISD::SMULO:
3770 case ISD::UMULO:
3771 if (Op.getResNo() != 1)
3772 break;
3773 // The boolean result conforms to getBooleanContents.
3774 // If we know the result of a setcc has the top bits zero, use this info.
3775 // We know that we have an integer-based boolean since these operations
3776 // are only available for integer.
3777 if (TLI->getBooleanContents(Op.getValueType().isVector(), false) ==
3779 BitWidth > 1)
3780 Known.Zero.setBitsFrom(1);
3781 break;
3782 case ISD::SETCC:
3783 case ISD::SETCCCARRY:
3784 case ISD::STRICT_FSETCC:
3785 case ISD::STRICT_FSETCCS: {
3786 unsigned OpNo = Op->isStrictFPOpcode() ? 1 : 0;
3787 // If we know the result of a setcc has the top bits zero, use this info.
3788 if (TLI->getBooleanContents(Op.getOperand(OpNo).getValueType()) ==
3790 BitWidth > 1)
3791 Known.Zero.setBitsFrom(1);
3792 break;
3793 }
3794 case ISD::SHL: {
3795 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3796 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3797
3798 bool NUW = Op->getFlags().hasNoUnsignedWrap();
3799 bool NSW = Op->getFlags().hasNoSignedWrap();
3800
3801 bool ShAmtNonZero = Known2.isNonZero();
3802
3803 Known = KnownBits::shl(Known, Known2, NUW, NSW, ShAmtNonZero);
3804
3805 // Minimum shift low bits are known zero.
3806 if (std::optional<unsigned> ShMinAmt =
3807 getValidMinimumShiftAmount(Op, DemandedElts, Depth + 1))
3808 Known.Zero.setLowBits(*ShMinAmt);
3809 break;
3810 }
3811 case ISD::SRL:
3812 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3813 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3814 Known = KnownBits::lshr(Known, Known2, /*ShAmtNonZero=*/false,
3815 Op->getFlags().hasExact());
3816
3817 // Minimum shift high bits are known zero.
3818 if (std::optional<unsigned> ShMinAmt =
3819 getValidMinimumShiftAmount(Op, DemandedElts, Depth + 1))
3820 Known.Zero.setHighBits(*ShMinAmt);
3821 break;
3822 case ISD::SRA:
3823 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3824 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3825 Known = KnownBits::ashr(Known, Known2, /*ShAmtNonZero=*/false,
3826 Op->getFlags().hasExact());
3827 break;
3828 case ISD::ROTL:
3829 case ISD::ROTR:
3830 if (ConstantSDNode *C =
3831 isConstOrConstSplat(Op.getOperand(1), DemandedElts)) {
3832 unsigned Amt = C->getAPIntValue().urem(BitWidth);
3833
3834 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3835
3836 // Canonicalize to ROTR.
3837 if (Opcode == ISD::ROTL && Amt != 0)
3838 Amt = BitWidth - Amt;
3839
3840 Known.Zero = Known.Zero.rotr(Amt);
3841 Known.One = Known.One.rotr(Amt);
3842 }
3843 break;
3844 case ISD::FSHL:
3845 case ISD::FSHR:
3846 if (ConstantSDNode *C = isConstOrConstSplat(Op.getOperand(2), DemandedElts)) {
3847 unsigned Amt = C->getAPIntValue().urem(BitWidth);
3848
3849 // For fshl, 0-shift returns the 1st arg.
3850 // For fshr, 0-shift returns the 2nd arg.
3851 if (Amt == 0) {
3852 Known = computeKnownBits(Op.getOperand(Opcode == ISD::FSHL ? 0 : 1),
3853 DemandedElts, Depth + 1);
3854 break;
3855 }
3856
3857 // fshl: (X << (Z % BW)) | (Y >> (BW - (Z % BW)))
3858 // fshr: (X << (BW - (Z % BW))) | (Y >> (Z % BW))
3859 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3860 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3861 if (Opcode == ISD::FSHL) {
3862 Known <<= Amt;
3863 Known2 >>= BitWidth - Amt;
3864 } else {
3865 Known <<= BitWidth - Amt;
3866 Known2 >>= Amt;
3867 }
3868 Known = Known.unionWith(Known2);
3869 }
3870 break;
3871 case ISD::SHL_PARTS:
3872 case ISD::SRA_PARTS:
3873 case ISD::SRL_PARTS: {
3874 assert((Op.getResNo() == 0 || Op.getResNo() == 1) && "Unknown result");
3875
3876 // Collect lo/hi source values and concatenate.
3877 unsigned LoBits = Op.getOperand(0).getScalarValueSizeInBits();
3878 unsigned HiBits = Op.getOperand(1).getScalarValueSizeInBits();
3879 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3880 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3881 Known = Known2.concat(Known);
3882
3883 // Collect shift amount.
3884 Known2 = computeKnownBits(Op.getOperand(2), DemandedElts, Depth + 1);
3885
3886 if (Opcode == ISD::SHL_PARTS)
3887 Known = KnownBits::shl(Known, Known2);
3888 else if (Opcode == ISD::SRA_PARTS)
3889 Known = KnownBits::ashr(Known, Known2);
3890 else // if (Opcode == ISD::SRL_PARTS)
3891 Known = KnownBits::lshr(Known, Known2);
3892
3893 // TODO: Minimum shift low/high bits are known zero.
3894
3895 if (Op.getResNo() == 0)
3896 Known = Known.extractBits(LoBits, 0);
3897 else
3898 Known = Known.extractBits(HiBits, LoBits);
3899 break;
3900 }
3902 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3903 EVT EVT = cast<VTSDNode>(Op.getOperand(1))->getVT();
3904 Known = Known.sextInReg(EVT.getScalarSizeInBits());
3905 break;
3906 }
3907 case ISD::CTTZ:
3908 case ISD::CTTZ_ZERO_UNDEF: {
3909 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3910 // If we have a known 1, its position is our upper bound.
3911 unsigned PossibleTZ = Known2.countMaxTrailingZeros();
3912 unsigned LowBits = llvm::bit_width(PossibleTZ);
3913 Known.Zero.setBitsFrom(LowBits);
3914 break;
3915 }
3916 case ISD::CTLZ:
3917 case ISD::CTLZ_ZERO_UNDEF: {
3918 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3919 // If we have a known 1, its position is our upper bound.
3920 unsigned PossibleLZ = Known2.countMaxLeadingZeros();
3921 unsigned LowBits = llvm::bit_width(PossibleLZ);
3922 Known.Zero.setBitsFrom(LowBits);
3923 break;
3924 }
3925 case ISD::CTLS: {
3926 unsigned MinRedundantSignBits =
3927 ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1) - 1;
3928 ConstantRange Range(APInt(BitWidth, MinRedundantSignBits),
3930 Known = Range.toKnownBits();
3931 break;
3932 }
3933 case ISD::CTPOP: {
3934 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3935 // If we know some of the bits are zero, they can't be one.
3936 unsigned PossibleOnes = Known2.countMaxPopulation();
3937 Known.Zero.setBitsFrom(llvm::bit_width(PossibleOnes));
3938 break;
3939 }
3940 case ISD::PARITY: {
3941 // Parity returns 0 everywhere but the LSB.
3942 Known.Zero.setBitsFrom(1);
3943 break;
3944 }
3945 case ISD::CLMUL: {
3946 Known = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
3947 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
3948 Known = KnownBits::clmul(Known, Known2);
3949 break;
3950 }
3951 case ISD::MGATHER:
3952 case ISD::MLOAD: {
3953 ISD::LoadExtType ETy =
3954 (Opcode == ISD::MGATHER)
3955 ? cast<MaskedGatherSDNode>(Op)->getExtensionType()
3956 : cast<MaskedLoadSDNode>(Op)->getExtensionType();
3957 if (ETy == ISD::ZEXTLOAD) {
3958 EVT MemVT = cast<MemSDNode>(Op)->getMemoryVT();
3959 KnownBits Known0(MemVT.getScalarSizeInBits());
3960 return Known0.zext(BitWidth);
3961 }
3962 break;
3963 }
3964 case ISD::LOAD: {
3966 const Constant *Cst = TLI->getTargetConstantFromLoad(LD);
3967 if (ISD::isNON_EXTLoad(LD) && Cst) {
3968 // Determine any common known bits from the loaded constant pool value.
3969 Type *CstTy = Cst->getType();
3970 if ((NumElts * BitWidth) == CstTy->getPrimitiveSizeInBits() &&
3971 !Op.getValueType().isScalableVector()) {
3972 // If its a vector splat, then we can (quickly) reuse the scalar path.
3973 // NOTE: We assume all elements match and none are UNDEF.
3974 if (CstTy->isVectorTy()) {
3975 if (const Constant *Splat = Cst->getSplatValue()) {
3976 Cst = Splat;
3977 CstTy = Cst->getType();
3978 }
3979 }
3980 // TODO - do we need to handle different bitwidths?
3981 if (CstTy->isVectorTy() && BitWidth == CstTy->getScalarSizeInBits()) {
3982 // Iterate across all vector elements finding common known bits.
3983 Known.setAllConflict();
3984 for (unsigned i = 0; i != NumElts; ++i) {
3985 if (!DemandedElts[i])
3986 continue;
3987 if (Constant *Elt = Cst->getAggregateElement(i)) {
3988 if (auto *CInt = dyn_cast<ConstantInt>(Elt)) {
3989 const APInt &Value = CInt->getValue();
3990 Known.One &= Value;
3991 Known.Zero &= ~Value;
3992 continue;
3993 }
3994 if (auto *CFP = dyn_cast<ConstantFP>(Elt)) {
3995 APInt Value = CFP->getValueAPF().bitcastToAPInt();
3996 Known.One &= Value;
3997 Known.Zero &= ~Value;
3998 continue;
3999 }
4000 }
4001 Known.One.clearAllBits();
4002 Known.Zero.clearAllBits();
4003 break;
4004 }
4005 } else if (BitWidth == CstTy->getPrimitiveSizeInBits()) {
4006 if (auto *CInt = dyn_cast<ConstantInt>(Cst)) {
4007 Known = KnownBits::makeConstant(CInt->getValue());
4008 } else if (auto *CFP = dyn_cast<ConstantFP>(Cst)) {
4009 Known =
4010 KnownBits::makeConstant(CFP->getValueAPF().bitcastToAPInt());
4011 }
4012 }
4013 }
4014 } else if (Op.getResNo() == 0) {
4015 unsigned ScalarMemorySize = LD->getMemoryVT().getScalarSizeInBits();
4016 KnownBits KnownScalarMemory(ScalarMemorySize);
4017 if (const MDNode *MD = LD->getRanges())
4018 computeKnownBitsFromRangeMetadata(*MD, KnownScalarMemory);
4019
4020 // Extend the Known bits from memory to the size of the scalar result.
4021 if (ISD::isZEXTLoad(Op.getNode()))
4022 Known = KnownScalarMemory.zext(BitWidth);
4023 else if (ISD::isSEXTLoad(Op.getNode()))
4024 Known = KnownScalarMemory.sext(BitWidth);
4025 else if (ISD::isEXTLoad(Op.getNode()))
4026 Known = KnownScalarMemory.anyext(BitWidth);
4027 else
4028 Known = KnownScalarMemory;
4029 assert(Known.getBitWidth() == BitWidth);
4030 return Known;
4031 }
4032 break;
4033 }
4035 if (Op.getValueType().isScalableVector())
4036 break;
4037 EVT InVT = Op.getOperand(0).getValueType();
4038 APInt InDemandedElts = DemandedElts.zext(InVT.getVectorNumElements());
4039 Known = computeKnownBits(Op.getOperand(0), InDemandedElts, Depth + 1);
4040 Known = Known.zext(BitWidth);
4041 break;
4042 }
4043 case ISD::ZERO_EXTEND: {
4044 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4045 Known = Known.zext(BitWidth);
4046 break;
4047 }
4049 if (Op.getValueType().isScalableVector())
4050 break;
4051 EVT InVT = Op.getOperand(0).getValueType();
4052 APInt InDemandedElts = DemandedElts.zext(InVT.getVectorNumElements());
4053 Known = computeKnownBits(Op.getOperand(0), InDemandedElts, Depth + 1);
4054 // If the sign bit is known to be zero or one, then sext will extend
4055 // it to the top bits, else it will just zext.
4056 Known = Known.sext(BitWidth);
4057 break;
4058 }
4059 case ISD::SIGN_EXTEND: {
4060 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4061 // If the sign bit is known to be zero or one, then sext will extend
4062 // it to the top bits, else it will just zext.
4063 Known = Known.sext(BitWidth);
4064 break;
4065 }
4067 if (Op.getValueType().isScalableVector())
4068 break;
4069 EVT InVT = Op.getOperand(0).getValueType();
4070 APInt InDemandedElts = DemandedElts.zext(InVT.getVectorNumElements());
4071 Known = computeKnownBits(Op.getOperand(0), InDemandedElts, Depth + 1);
4072 Known = Known.anyext(BitWidth);
4073 break;
4074 }
4075 case ISD::ANY_EXTEND: {
4076 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4077 Known = Known.anyext(BitWidth);
4078 break;
4079 }
4080 case ISD::TRUNCATE: {
4081 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4082 Known = Known.trunc(BitWidth);
4083 break;
4084 }
4085 case ISD::TRUNCATE_SSAT_S: {
4086 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4087 Known = Known.truncSSat(BitWidth);
4088 break;
4089 }
4090 case ISD::TRUNCATE_SSAT_U: {
4091 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4092 Known = Known.truncSSatU(BitWidth);
4093 break;
4094 }
4095 case ISD::TRUNCATE_USAT_U: {
4096 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4097 Known = Known.truncUSat(BitWidth);
4098 break;
4099 }
4100 case ISD::AssertZext: {
4101 EVT VT = cast<VTSDNode>(Op.getOperand(1))->getVT();
4103 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4104 Known.Zero |= (~InMask);
4105 Known.One &= (~Known.Zero);
4106 break;
4107 }
4108 case ISD::AssertAlign: {
4109 unsigned LogOfAlign = Log2(cast<AssertAlignSDNode>(Op)->getAlign());
4110 assert(LogOfAlign != 0);
4111
4112 // TODO: Should use maximum with source
4113 // If a node is guaranteed to be aligned, set low zero bits accordingly as
4114 // well as clearing one bits.
4115 Known.Zero.setLowBits(LogOfAlign);
4116 Known.One.clearLowBits(LogOfAlign);
4117 break;
4118 }
4119 case ISD::AssertNoFPClass: {
4120 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4121
4122 FPClassTest NoFPClass =
4123 static_cast<FPClassTest>(Op.getConstantOperandVal(1));
4124 const FPClassTest NegativeTestMask = fcNan | fcNegative;
4125 if ((NoFPClass & NegativeTestMask) == NegativeTestMask) {
4126 // Cannot be negative.
4127 Known.makeNonNegative();
4128 }
4129
4130 const FPClassTest PositiveTestMask = fcNan | fcPositive;
4131 if ((NoFPClass & PositiveTestMask) == PositiveTestMask) {
4132 // Cannot be positive.
4133 Known.makeNegative();
4134 }
4135
4136 break;
4137 }
4138 case ISD::FGETSIGN:
4139 // All bits are zero except the low bit.
4140 Known.Zero.setBitsFrom(1);
4141 break;
4142 case ISD::ADD:
4143 case ISD::SUB: {
4144 SDNodeFlags Flags = Op.getNode()->getFlags();
4145 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4146 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4148 Op.getOpcode() == ISD::ADD, Flags.hasNoSignedWrap(),
4149 Flags.hasNoUnsignedWrap(), Known, Known2);
4150 break;
4151 }
4152 case ISD::USUBO:
4153 case ISD::SSUBO:
4154 case ISD::USUBO_CARRY:
4155 case ISD::SSUBO_CARRY:
4156 if (Op.getResNo() == 1) {
4157 // If we know the result of a setcc has the top bits zero, use this info.
4158 if (TLI->getBooleanContents(Op.getOperand(0).getValueType()) ==
4160 BitWidth > 1)
4161 Known.Zero.setBitsFrom(1);
4162 break;
4163 }
4164 [[fallthrough]];
4165 case ISD::SUBC: {
4166 assert(Op.getResNo() == 0 &&
4167 "We only compute knownbits for the difference here.");
4168
4169 // With USUBO_CARRY and SSUBO_CARRY a borrow bit may be added in.
4170 KnownBits Borrow(1);
4171 if (Opcode == ISD::USUBO_CARRY || Opcode == ISD::SSUBO_CARRY) {
4172 Borrow = computeKnownBits(Op.getOperand(2), DemandedElts, Depth + 1);
4173 // Borrow has bit width 1
4174 Borrow = Borrow.trunc(1);
4175 } else {
4176 Borrow.setAllZero();
4177 }
4178
4179 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4180 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4181 Known = KnownBits::computeForSubBorrow(Known, Known2, Borrow);
4182 break;
4183 }
4184 case ISD::UADDO:
4185 case ISD::SADDO:
4186 case ISD::UADDO_CARRY:
4187 case ISD::SADDO_CARRY:
4188 if (Op.getResNo() == 1) {
4189 // If we know the result of a setcc has the top bits zero, use this info.
4190 if (TLI->getBooleanContents(Op.getOperand(0).getValueType()) ==
4192 BitWidth > 1)
4193 Known.Zero.setBitsFrom(1);
4194 break;
4195 }
4196 [[fallthrough]];
4197 case ISD::ADDC:
4198 case ISD::ADDE: {
4199 assert(Op.getResNo() == 0 && "We only compute knownbits for the sum here.");
4200
4201 // With ADDE and UADDO_CARRY, a carry bit may be added in.
4202 KnownBits Carry(1);
4203 if (Opcode == ISD::ADDE)
4204 // Can't track carry from glue, set carry to unknown.
4205 Carry.resetAll();
4206 else if (Opcode == ISD::UADDO_CARRY || Opcode == ISD::SADDO_CARRY) {
4207 Carry = computeKnownBits(Op.getOperand(2), DemandedElts, Depth + 1);
4208 // Carry has bit width 1
4209 Carry = Carry.trunc(1);
4210 } else {
4211 Carry.setAllZero();
4212 }
4213
4214 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4215 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4216 Known = KnownBits::computeForAddCarry(Known, Known2, Carry);
4217 break;
4218 }
4219 case ISD::UDIV: {
4220 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4221 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4222 Known = KnownBits::udiv(Known, Known2, Op->getFlags().hasExact());
4223 break;
4224 }
4225 case ISD::SDIV: {
4226 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4227 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4228 Known = KnownBits::sdiv(Known, Known2, Op->getFlags().hasExact());
4229 break;
4230 }
4231 case ISD::SREM: {
4232 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4233 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4234 Known = KnownBits::srem(Known, Known2);
4235 break;
4236 }
4237 case ISD::UREM: {
4238 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4239 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4240 Known = KnownBits::urem(Known, Known2);
4241 break;
4242 }
4243 case ISD::EXTRACT_ELEMENT: {
4244 Known = computeKnownBits(Op.getOperand(0), Depth+1);
4245 const unsigned Index = Op.getConstantOperandVal(1);
4246 const unsigned EltBitWidth = Op.getValueSizeInBits();
4247
4248 // Remove low part of known bits mask
4249 Known.Zero = Known.Zero.getHiBits(Known.getBitWidth() - Index * EltBitWidth);
4250 Known.One = Known.One.getHiBits(Known.getBitWidth() - Index * EltBitWidth);
4251
4252 // Remove high part of known bit mask
4253 Known = Known.trunc(EltBitWidth);
4254 break;
4255 }
4257 SDValue InVec = Op.getOperand(0);
4258 SDValue EltNo = Op.getOperand(1);
4259 EVT VecVT = InVec.getValueType();
4260 // computeKnownBits not yet implemented for scalable vectors.
4261 if (VecVT.isScalableVector())
4262 break;
4263 const unsigned EltBitWidth = VecVT.getScalarSizeInBits();
4264 const unsigned NumSrcElts = VecVT.getVectorNumElements();
4265
4266 // If BitWidth > EltBitWidth the value is anyext:ed. So we do not know
4267 // anything about the extended bits.
4268 if (BitWidth > EltBitWidth)
4269 Known = Known.trunc(EltBitWidth);
4270
4271 // If we know the element index, just demand that vector element, else for
4272 // an unknown element index, ignore DemandedElts and demand them all.
4273 APInt DemandedSrcElts = APInt::getAllOnes(NumSrcElts);
4274 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
4275 if (ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts))
4276 DemandedSrcElts =
4277 APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue());
4278
4279 Known = computeKnownBits(InVec, DemandedSrcElts, Depth + 1);
4280 if (BitWidth > EltBitWidth)
4281 Known = Known.anyext(BitWidth);
4282 break;
4283 }
4285 if (Op.getValueType().isScalableVector())
4286 break;
4287
4288 // If we know the element index, split the demand between the
4289 // source vector and the inserted element, otherwise assume we need
4290 // the original demanded vector elements and the value.
4291 SDValue InVec = Op.getOperand(0);
4292 SDValue InVal = Op.getOperand(1);
4293 SDValue EltNo = Op.getOperand(2);
4294 bool DemandedVal = true;
4295 APInt DemandedVecElts = DemandedElts;
4296 auto *CEltNo = dyn_cast<ConstantSDNode>(EltNo);
4297 if (CEltNo && CEltNo->getAPIntValue().ult(NumElts)) {
4298 unsigned EltIdx = CEltNo->getZExtValue();
4299 DemandedVal = !!DemandedElts[EltIdx];
4300 DemandedVecElts.clearBit(EltIdx);
4301 }
4302 Known.setAllConflict();
4303 if (DemandedVal) {
4304 Known2 = computeKnownBits(InVal, Depth + 1);
4305 Known = Known.intersectWith(Known2.zextOrTrunc(BitWidth));
4306 }
4307 if (!!DemandedVecElts) {
4308 Known2 = computeKnownBits(InVec, DemandedVecElts, Depth + 1);
4309 Known = Known.intersectWith(Known2);
4310 }
4311 break;
4312 }
4313 case ISD::BITREVERSE: {
4314 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4315 Known = Known2.reverseBits();
4316 break;
4317 }
4318 case ISD::BSWAP: {
4319 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4320 Known = Known2.byteSwap();
4321 break;
4322 }
4323 case ISD::ABS: {
4324 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4325 Known = Known2.abs();
4326 Known.Zero.setHighBits(
4327 ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1) - 1);
4328 break;
4329 }
4330 case ISD::USUBSAT: {
4331 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4332 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4333 Known = KnownBits::usub_sat(Known, Known2);
4334 break;
4335 }
4336 case ISD::UMIN: {
4337 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4338 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4339 Known = KnownBits::umin(Known, Known2);
4340 break;
4341 }
4342 case ISD::UMAX: {
4343 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4344 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4345 Known = KnownBits::umax(Known, Known2);
4346 break;
4347 }
4348 case ISD::SMIN:
4349 case ISD::SMAX: {
4350 // If we have a clamp pattern, we know that the number of sign bits will be
4351 // the minimum of the clamp min/max range.
4352 bool IsMax = (Opcode == ISD::SMAX);
4353 ConstantSDNode *CstLow = nullptr, *CstHigh = nullptr;
4354 if ((CstLow = isConstOrConstSplat(Op.getOperand(1), DemandedElts)))
4355 if (Op.getOperand(0).getOpcode() == (IsMax ? ISD::SMIN : ISD::SMAX))
4356 CstHigh =
4357 isConstOrConstSplat(Op.getOperand(0).getOperand(1), DemandedElts);
4358 if (CstLow && CstHigh) {
4359 if (!IsMax)
4360 std::swap(CstLow, CstHigh);
4361
4362 const APInt &ValueLow = CstLow->getAPIntValue();
4363 const APInt &ValueHigh = CstHigh->getAPIntValue();
4364 if (ValueLow.sle(ValueHigh)) {
4365 unsigned LowSignBits = ValueLow.getNumSignBits();
4366 unsigned HighSignBits = ValueHigh.getNumSignBits();
4367 unsigned MinSignBits = std::min(LowSignBits, HighSignBits);
4368 if (ValueLow.isNegative() && ValueHigh.isNegative()) {
4369 Known.One.setHighBits(MinSignBits);
4370 break;
4371 }
4372 if (ValueLow.isNonNegative() && ValueHigh.isNonNegative()) {
4373 Known.Zero.setHighBits(MinSignBits);
4374 break;
4375 }
4376 }
4377 }
4378
4379 Known = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4380 Known2 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
4381 if (IsMax)
4382 Known = KnownBits::smax(Known, Known2);
4383 else
4384 Known = KnownBits::smin(Known, Known2);
4385
4386 // For SMAX, if CstLow is non-negative we know the result will be
4387 // non-negative and thus all sign bits are 0.
4388 // TODO: There's an equivalent of this for smin with negative constant for
4389 // known ones.
4390 if (IsMax && CstLow) {
4391 const APInt &ValueLow = CstLow->getAPIntValue();
4392 if (ValueLow.isNonNegative()) {
4393 unsigned SignBits = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
4394 Known.Zero.setHighBits(std::min(SignBits, ValueLow.getNumSignBits()));
4395 }
4396 }
4397
4398 break;
4399 }
4400 case ISD::UINT_TO_FP: {
4401 Known.makeNonNegative();
4402 break;
4403 }
4404 case ISD::SINT_TO_FP: {
4405 Known2 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
4406 if (Known2.isNonNegative())
4407 Known.makeNonNegative();
4408 else if (Known2.isNegative())
4409 Known.makeNegative();
4410 break;
4411 }
4412 case ISD::FP_TO_UINT_SAT: {
4413 // FP_TO_UINT_SAT produces an unsigned value that fits in the saturating VT.
4414 EVT VT = cast<VTSDNode>(Op.getOperand(1))->getVT();
4416 break;
4417 }
4418 case ISD::ATOMIC_LOAD: {
4419 // If we are looking at the loaded value.
4420 if (Op.getResNo() == 0) {
4421 auto *AT = cast<AtomicSDNode>(Op);
4422 unsigned ScalarMemorySize = AT->getMemoryVT().getScalarSizeInBits();
4423 KnownBits KnownScalarMemory(ScalarMemorySize);
4424 if (const MDNode *MD = AT->getRanges())
4425 computeKnownBitsFromRangeMetadata(*MD, KnownScalarMemory);
4426
4427 switch (AT->getExtensionType()) {
4428 case ISD::ZEXTLOAD:
4429 Known = KnownScalarMemory.zext(BitWidth);
4430 break;
4431 case ISD::SEXTLOAD:
4432 Known = KnownScalarMemory.sext(BitWidth);
4433 break;
4434 case ISD::EXTLOAD:
4435 switch (TLI->getExtendForAtomicOps()) {
4436 case ISD::ZERO_EXTEND:
4437 Known = KnownScalarMemory.zext(BitWidth);
4438 break;
4439 case ISD::SIGN_EXTEND:
4440 Known = KnownScalarMemory.sext(BitWidth);
4441 break;
4442 default:
4443 Known = KnownScalarMemory.anyext(BitWidth);
4444 break;
4445 }
4446 break;
4447 case ISD::NON_EXTLOAD:
4448 Known = KnownScalarMemory;
4449 break;
4450 }
4451 assert(Known.getBitWidth() == BitWidth);
4452 }
4453 break;
4454 }
4456 if (Op.getResNo() == 1) {
4457 // The boolean result conforms to getBooleanContents.
4458 // If we know the result of a setcc has the top bits zero, use this info.
4459 // We know that we have an integer-based boolean since these operations
4460 // are only available for integer.
4461 if (TLI->getBooleanContents(Op.getValueType().isVector(), false) ==
4463 BitWidth > 1)
4464 Known.Zero.setBitsFrom(1);
4465 break;
4466 }
4467 [[fallthrough]];
4469 case ISD::ATOMIC_SWAP:
4480 case ISD::ATOMIC_LOAD_UMAX: {
4481 // If we are looking at the loaded value.
4482 if (Op.getResNo() == 0) {
4483 auto *AT = cast<AtomicSDNode>(Op);
4484 unsigned MemBits = AT->getMemoryVT().getScalarSizeInBits();
4485
4486 if (TLI->getExtendForAtomicOps() == ISD::ZERO_EXTEND)
4487 Known.Zero.setBitsFrom(MemBits);
4488 }
4489 break;
4490 }
4491 case ISD::FrameIndex:
4493 TLI->computeKnownBitsForFrameIndex(cast<FrameIndexSDNode>(Op)->getIndex(),
4494 Known, getMachineFunction());
4495 break;
4496
4497 default:
4498 if (Opcode < ISD::BUILTIN_OP_END)
4499 break;
4500 [[fallthrough]];
4504 // TODO: Probably okay to remove after audit; here to reduce change size
4505 // in initial enablement patch for scalable vectors
4506 if (Op.getValueType().isScalableVector())
4507 break;
4508
4509 // Allow the target to implement this method for its nodes.
4510 TLI->computeKnownBitsForTargetNode(Op, Known, DemandedElts, *this, Depth);
4511 break;
4512 }
4513
4514 return Known;
4515}
4516
4517/// Convert ConstantRange OverflowResult into SelectionDAG::OverflowKind.
4530
4533 // X + 0 never overflow
4534 if (isNullConstant(N1))
4535 return OFK_Never;
4536
4537 // If both operands each have at least two sign bits, the addition
4538 // cannot overflow.
4539 if (ComputeNumSignBits(N0) > 1 && ComputeNumSignBits(N1) > 1)
4540 return OFK_Never;
4541
4542 // TODO: Add ConstantRange::signedAddMayOverflow handling.
4543 return OFK_Sometime;
4544}
4545
4548 // X + 0 never overflow
4549 if (isNullConstant(N1))
4550 return OFK_Never;
4551
4552 // mulhi + 1 never overflow
4553 KnownBits N1Known = computeKnownBits(N1);
4554 if (N0.getOpcode() == ISD::UMUL_LOHI && N0.getResNo() == 1 &&
4555 N1Known.getMaxValue().ult(2))
4556 return OFK_Never;
4557
4558 KnownBits N0Known = computeKnownBits(N0);
4559 if (N1.getOpcode() == ISD::UMUL_LOHI && N1.getResNo() == 1 &&
4560 N0Known.getMaxValue().ult(2))
4561 return OFK_Never;
4562
4563 // Fallback to ConstantRange::unsignedAddMayOverflow handling.
4564 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, false);
4565 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, false);
4566 return mapOverflowResult(N0Range.unsignedAddMayOverflow(N1Range));
4567}
4568
4571 // X - 0 never overflow
4572 if (isNullConstant(N1))
4573 return OFK_Never;
4574
4575 // If both operands each have at least two sign bits, the subtraction
4576 // cannot overflow.
4577 if (ComputeNumSignBits(N0) > 1 && ComputeNumSignBits(N1) > 1)
4578 return OFK_Never;
4579
4580 KnownBits N0Known = computeKnownBits(N0);
4581 KnownBits N1Known = computeKnownBits(N1);
4582 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, true);
4583 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, true);
4584 return mapOverflowResult(N0Range.signedSubMayOverflow(N1Range));
4585}
4586
4589 // X - 0 never overflow
4590 if (isNullConstant(N1))
4591 return OFK_Never;
4592
4593 KnownBits N0Known = computeKnownBits(N0);
4594 KnownBits N1Known = computeKnownBits(N1);
4595 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, false);
4596 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, false);
4597 return mapOverflowResult(N0Range.unsignedSubMayOverflow(N1Range));
4598}
4599
4602 // X * 0 and X * 1 never overflow.
4603 if (isNullConstant(N1) || isOneConstant(N1))
4604 return OFK_Never;
4605
4606 KnownBits N0Known = computeKnownBits(N0);
4607 KnownBits N1Known = computeKnownBits(N1);
4608 ConstantRange N0Range = ConstantRange::fromKnownBits(N0Known, false);
4609 ConstantRange N1Range = ConstantRange::fromKnownBits(N1Known, false);
4610 return mapOverflowResult(N0Range.unsignedMulMayOverflow(N1Range));
4611}
4612
4615 // X * 0 and X * 1 never overflow.
4616 if (isNullConstant(N1) || isOneConstant(N1))
4617 return OFK_Never;
4618
4619 // Get the size of the result.
4620 unsigned BitWidth = N0.getScalarValueSizeInBits();
4621
4622 // Sum of the sign bits.
4623 unsigned SignBits = ComputeNumSignBits(N0) + ComputeNumSignBits(N1);
4624
4625 // If we have enough sign bits, then there's no overflow.
4626 if (SignBits > BitWidth + 1)
4627 return OFK_Never;
4628
4629 if (SignBits == BitWidth + 1) {
4630 // The overflow occurs when the true multiplication of the
4631 // the operands is the minimum negative number.
4632 KnownBits N0Known = computeKnownBits(N0);
4633 KnownBits N1Known = computeKnownBits(N1);
4634 // If one of the operands is non-negative, then there's no
4635 // overflow.
4636 if (N0Known.isNonNegative() || N1Known.isNonNegative())
4637 return OFK_Never;
4638 }
4639
4640 return OFK_Sometime;
4641}
4642
4644 unsigned Depth) const {
4645 EVT VT = Val.getValueType();
4646
4647 // Since the number of lanes in a scalable vector is unknown at compile time,
4648 // we track one bit which is implicitly broadcast to all lanes. This means
4649 // that all lanes in a scalable vector are considered demanded.
4650 APInt DemandedElts = VT.isFixedLengthVector()
4652 : APInt(1, 1);
4653
4654 return isKnownToBeAPowerOfTwo(Val, DemandedElts, OrZero, Depth);
4655}
4656
4658 const APInt &DemandedElts,
4659 bool OrZero, unsigned Depth) const {
4660 if (Depth >= MaxRecursionDepth)
4661 return false; // Limit search depth.
4662
4663 EVT OpVT = Val.getValueType();
4664 unsigned BitWidth = OpVT.getScalarSizeInBits();
4665 [[maybe_unused]] unsigned NumElts = DemandedElts.getBitWidth();
4666 assert((!OpVT.isScalableVector() || NumElts == 1) &&
4667 "DemandedElts for scalable vectors must be 1 to represent all lanes");
4668 assert(
4669 (!OpVT.isFixedLengthVector() || NumElts == OpVT.getVectorNumElements()) &&
4670 "Unexpected vector size");
4671
4672 auto IsPowerOfTwoOrZero = [BitWidth, OrZero](const ConstantSDNode *C) {
4673 APInt V = C->getAPIntValue().zextOrTrunc(BitWidth);
4674 return (OrZero && V.isZero()) || V.isPowerOf2();
4675 };
4676
4677 // Is the constant a known power of 2 or zero?
4678 if (ISD::matchUnaryPredicate(Val, IsPowerOfTwoOrZero))
4679 return true;
4680
4681 switch (Val.getOpcode()) {
4682 case ISD::BUILD_VECTOR:
4683 // Are all operands of a build vector constant powers of two or zero?
4684 if (all_of(enumerate(Val->ops()), [&](auto P) {
4685 auto *C = dyn_cast<ConstantSDNode>(P.value());
4686 return !DemandedElts[P.index()] || (C && IsPowerOfTwoOrZero(C));
4687 }))
4688 return true;
4689 break;
4690
4691 case ISD::SPLAT_VECTOR:
4692 // Is the operand of a splat vector a constant power of two?
4693 if (auto *C = dyn_cast<ConstantSDNode>(Val->getOperand(0)))
4694 if (IsPowerOfTwoOrZero(C))
4695 return true;
4696 break;
4697
4699 SDValue InVec = Val.getOperand(0);
4700 SDValue EltNo = Val.getOperand(1);
4701 EVT VecVT = InVec.getValueType();
4702
4703 // Skip scalable vectors or implicit extensions.
4704 if (VecVT.isScalableVector() ||
4705 OpVT.getScalarSizeInBits() != VecVT.getScalarSizeInBits())
4706 break;
4707
4708 // If we know the element index, just demand that vector element, else for
4709 // an unknown element index, ignore DemandedElts and demand them all.
4710 const unsigned NumSrcElts = VecVT.getVectorNumElements();
4711 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
4712 APInt DemandedSrcElts =
4713 ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts)
4714 ? APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue())
4715 : APInt::getAllOnes(NumSrcElts);
4716 return isKnownToBeAPowerOfTwo(InVec, DemandedSrcElts, OrZero, Depth + 1);
4717 }
4718
4719 case ISD::AND: {
4720 // Looking for `x & -x` pattern:
4721 // If x == 0:
4722 // x & -x -> 0
4723 // If x != 0:
4724 // x & -x -> non-zero pow2
4725 // so if we find the pattern return whether we know `x` is non-zero.
4726 SDValue X;
4727 if (sd_match(Val, m_And(m_Value(X), m_Neg(m_Deferred(X)))))
4728 return OrZero || isKnownNeverZero(X, DemandedElts, Depth);
4729 break;
4730 }
4731
4732 case ISD::SHL: {
4733 // A left-shift of a constant one will have exactly one bit set because
4734 // shifting the bit off the end is undefined.
4735 auto *C = isConstOrConstSplat(Val.getOperand(0), DemandedElts);
4736 if (C && C->getAPIntValue() == 1)
4737 return true;
4738 return (OrZero || isKnownNeverZero(Val, DemandedElts, Depth)) &&
4739 isKnownToBeAPowerOfTwo(Val.getOperand(0), DemandedElts, OrZero,
4740 Depth + 1);
4741 }
4742
4743 case ISD::SRL: {
4744 // A logical right-shift of a constant sign-bit will have exactly
4745 // one bit set.
4746 auto *C = isConstOrConstSplat(Val.getOperand(0));
4747 if (C && C->getAPIntValue().isSignMask())
4748 return true;
4749 return isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4750 Depth + 1) &&
4751 isKnownNeverZero(Val, Depth);
4752 }
4753
4754 case ISD::ROTL:
4755 case ISD::ROTR:
4756 return isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4757 Depth + 1);
4758
4759 case ISD::SMIN:
4760 case ISD::SMAX:
4761 case ISD::UMIN:
4762 case ISD::UMAX:
4763 return isKnownToBeAPowerOfTwo(Val.getOperand(1), /*OrZero=*/false,
4764 Depth + 1) &&
4765 isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4766 Depth + 1);
4767
4768 case ISD::SELECT:
4769 case ISD::VSELECT:
4770 return isKnownToBeAPowerOfTwo(Val.getOperand(2), DemandedElts, OrZero,
4771 Depth + 1) &&
4772 isKnownToBeAPowerOfTwo(Val.getOperand(1), DemandedElts, OrZero,
4773 Depth + 1);
4774
4775 case ISD::ZERO_EXTEND:
4776 return isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4777 Depth + 1);
4778
4779 case ISD::VSCALE:
4780 // vscale(power-of-two) is a power-of-two
4781 return isKnownToBeAPowerOfTwo(Val.getOperand(0), /*OrZero=*/false,
4782 Depth + 1);
4783 }
4784
4785 // More could be done here, though the above checks are enough
4786 // to handle some common cases.
4787 return false;
4788}
4789
4791 if (ConstantFPSDNode *C1 = isConstOrConstSplatFP(Val, true))
4792 return C1->getValueAPF().getExactLog2Abs() >= 0;
4793
4794 if (Val.getOpcode() == ISD::UINT_TO_FP || Val.getOpcode() == ISD::SINT_TO_FP)
4795 return isKnownToBeAPowerOfTwo(Val.getOperand(0), Depth + 1);
4796
4797 return false;
4798}
4799
4801 EVT VT = Op.getValueType();
4802
4803 // Since the number of lanes in a scalable vector is unknown at compile time,
4804 // we track one bit which is implicitly broadcast to all lanes. This means
4805 // that all lanes in a scalable vector are considered demanded.
4806 APInt DemandedElts = VT.isFixedLengthVector()
4808 : APInt(1, 1);
4809 return ComputeNumSignBits(Op, DemandedElts, Depth);
4810}
4811
4812unsigned SelectionDAG::ComputeNumSignBits(SDValue Op, const APInt &DemandedElts,
4813 unsigned Depth) const {
4814 EVT VT = Op.getValueType();
4815 assert((VT.isInteger() || VT.isFloatingPoint()) && "Invalid VT!");
4816 unsigned VTBits = VT.getScalarSizeInBits();
4817 unsigned NumElts = DemandedElts.getBitWidth();
4818 unsigned Tmp, Tmp2;
4819 unsigned FirstAnswer = 1;
4820
4821 assert((!VT.isScalableVector() || NumElts == 1) &&
4822 "DemandedElts for scalable vectors must be 1 to represent all lanes");
4823
4824 if (auto *C = dyn_cast<ConstantSDNode>(Op)) {
4825 const APInt &Val = C->getAPIntValue();
4826 return Val.getNumSignBits();
4827 }
4828
4829 if (Depth >= MaxRecursionDepth)
4830 return 1; // Limit search depth.
4831
4832 if (!DemandedElts)
4833 return 1; // No demanded elts, better to assume we don't know anything.
4834
4835 unsigned Opcode = Op.getOpcode();
4836 switch (Opcode) {
4837 default: break;
4838 case ISD::AssertSext:
4839 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getSizeInBits();
4840 return VTBits-Tmp+1;
4841 case ISD::AssertZext:
4842 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getSizeInBits();
4843 return VTBits-Tmp;
4844 case ISD::FREEZE:
4845 if (isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), DemandedElts,
4846 /*PoisonOnly=*/false))
4847 return ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
4848 break;
4849 case ISD::MERGE_VALUES:
4850 return ComputeNumSignBits(Op.getOperand(Op.getResNo()), DemandedElts,
4851 Depth + 1);
4852 case ISD::SPLAT_VECTOR: {
4853 // Check if the sign bits of source go down as far as the truncated value.
4854 unsigned NumSrcBits = Op.getOperand(0).getValueSizeInBits();
4855 unsigned NumSrcSignBits = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
4856 if (NumSrcSignBits > (NumSrcBits - VTBits))
4857 return NumSrcSignBits - (NumSrcBits - VTBits);
4858 break;
4859 }
4860 case ISD::BUILD_VECTOR:
4861 assert(!VT.isScalableVector());
4862 Tmp = VTBits;
4863 for (unsigned i = 0, e = Op.getNumOperands(); (i < e) && (Tmp > 1); ++i) {
4864 if (!DemandedElts[i])
4865 continue;
4866
4867 SDValue SrcOp = Op.getOperand(i);
4868 // BUILD_VECTOR can implicitly truncate sources, we handle this specially
4869 // for constant nodes to ensure we only look at the sign bits.
4871 APInt T = C->getAPIntValue().trunc(VTBits);
4872 Tmp2 = T.getNumSignBits();
4873 } else {
4874 Tmp2 = ComputeNumSignBits(SrcOp, Depth + 1);
4875
4876 if (SrcOp.getValueSizeInBits() != VTBits) {
4877 assert(SrcOp.getValueSizeInBits() > VTBits &&
4878 "Expected BUILD_VECTOR implicit truncation");
4879 unsigned ExtraBits = SrcOp.getValueSizeInBits() - VTBits;
4880 Tmp2 = (Tmp2 > ExtraBits ? Tmp2 - ExtraBits : 1);
4881 }
4882 }
4883 Tmp = std::min(Tmp, Tmp2);
4884 }
4885 return Tmp;
4886
4887 case ISD::VECTOR_COMPRESS: {
4888 SDValue Vec = Op.getOperand(0);
4889 SDValue PassThru = Op.getOperand(2);
4890 Tmp = ComputeNumSignBits(PassThru, DemandedElts, Depth + 1);
4891 if (Tmp == 1)
4892 return 1;
4893 Tmp2 = ComputeNumSignBits(Vec, Depth + 1);
4894 Tmp = std::min(Tmp, Tmp2);
4895 return Tmp;
4896 }
4897
4898 case ISD::VECTOR_SHUFFLE: {
4899 // Collect the minimum number of sign bits that are shared by every vector
4900 // element referenced by the shuffle.
4901 APInt DemandedLHS, DemandedRHS;
4903 assert(NumElts == SVN->getMask().size() && "Unexpected vector size");
4904 if (!getShuffleDemandedElts(NumElts, SVN->getMask(), DemandedElts,
4905 DemandedLHS, DemandedRHS))
4906 return 1;
4907
4908 Tmp = std::numeric_limits<unsigned>::max();
4909 if (!!DemandedLHS)
4910 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedLHS, Depth + 1);
4911 if (!!DemandedRHS) {
4912 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedRHS, Depth + 1);
4913 Tmp = std::min(Tmp, Tmp2);
4914 }
4915 // If we don't know anything, early out and try computeKnownBits fall-back.
4916 if (Tmp == 1)
4917 break;
4918 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
4919 return Tmp;
4920 }
4921
4922 case ISD::BITCAST: {
4923 if (VT.isScalableVector())
4924 break;
4925 SDValue N0 = Op.getOperand(0);
4926 EVT SrcVT = N0.getValueType();
4927 unsigned SrcBits = SrcVT.getScalarSizeInBits();
4928
4929 // Ignore bitcasts from unsupported types..
4930 if (!(SrcVT.isInteger() || SrcVT.isFloatingPoint()))
4931 break;
4932
4933 // Fast handling of 'identity' bitcasts.
4934 if (VTBits == SrcBits)
4935 return ComputeNumSignBits(N0, DemandedElts, Depth + 1);
4936
4937 bool IsLE = getDataLayout().isLittleEndian();
4938
4939 // Bitcast 'large element' scalar/vector to 'small element' vector.
4940 if ((SrcBits % VTBits) == 0) {
4941 assert(VT.isVector() && "Expected bitcast to vector");
4942
4943 unsigned Scale = SrcBits / VTBits;
4944 APInt SrcDemandedElts =
4945 APIntOps::ScaleBitMask(DemandedElts, NumElts / Scale);
4946
4947 // Fast case - sign splat can be simply split across the small elements.
4948 Tmp = ComputeNumSignBits(N0, SrcDemandedElts, Depth + 1);
4949 if (Tmp == SrcBits)
4950 return VTBits;
4951
4952 // Slow case - determine how far the sign extends into each sub-element.
4953 Tmp2 = VTBits;
4954 for (unsigned i = 0; i != NumElts; ++i)
4955 if (DemandedElts[i]) {
4956 unsigned SubOffset = i % Scale;
4957 SubOffset = (IsLE ? ((Scale - 1) - SubOffset) : SubOffset);
4958 SubOffset = SubOffset * VTBits;
4959 if (Tmp <= SubOffset)
4960 return 1;
4961 Tmp2 = std::min(Tmp2, Tmp - SubOffset);
4962 }
4963 return Tmp2;
4964 }
4965 break;
4966 }
4967
4969 // FP_TO_SINT_SAT produces a signed value that fits in the saturating VT.
4970 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getScalarSizeInBits();
4971 return VTBits - Tmp + 1;
4972 case ISD::SIGN_EXTEND:
4973 Tmp = VTBits - Op.getOperand(0).getScalarValueSizeInBits();
4974 return ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth+1) + Tmp;
4976 // Max of the input and what this extends.
4977 Tmp = cast<VTSDNode>(Op.getOperand(1))->getVT().getScalarSizeInBits();
4978 Tmp = VTBits-Tmp+1;
4979 Tmp2 = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth+1);
4980 return std::max(Tmp, Tmp2);
4982 if (VT.isScalableVector())
4983 break;
4984 SDValue Src = Op.getOperand(0);
4985 EVT SrcVT = Src.getValueType();
4986 APInt DemandedSrcElts = DemandedElts.zext(SrcVT.getVectorNumElements());
4987 Tmp = VTBits - SrcVT.getScalarSizeInBits();
4988 return ComputeNumSignBits(Src, DemandedSrcElts, Depth+1) + Tmp;
4989 }
4990 case ISD::SRA:
4991 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
4992 // SRA X, C -> adds C sign bits.
4993 if (std::optional<unsigned> ShAmt =
4994 getValidMinimumShiftAmount(Op, DemandedElts, Depth + 1))
4995 Tmp = std::min(Tmp + *ShAmt, VTBits);
4996 return Tmp;
4997 case ISD::SHL:
4998 if (std::optional<ConstantRange> ShAmtRange =
4999 getValidShiftAmountRange(Op, DemandedElts, Depth + 1)) {
5000 unsigned MaxShAmt = ShAmtRange->getUnsignedMax().getZExtValue();
5001 unsigned MinShAmt = ShAmtRange->getUnsignedMin().getZExtValue();
5002 // Try to look through ZERO/SIGN/ANY_EXTEND. If all extended bits are
5003 // shifted out, then we can compute the number of sign bits for the
5004 // operand being extended. A future improvement could be to pass along the
5005 // "shifted left by" information in the recursive calls to
5006 // ComputeKnownSignBits. Allowing us to handle this more generically.
5007 if (ISD::isExtOpcode(Op.getOperand(0).getOpcode())) {
5008 SDValue Ext = Op.getOperand(0);
5009 EVT ExtVT = Ext.getValueType();
5010 SDValue Extendee = Ext.getOperand(0);
5011 EVT ExtendeeVT = Extendee.getValueType();
5012 unsigned SizeDifference =
5013 ExtVT.getScalarSizeInBits() - ExtendeeVT.getScalarSizeInBits();
5014 if (SizeDifference <= MinShAmt) {
5015 Tmp = SizeDifference +
5016 ComputeNumSignBits(Extendee, DemandedElts, Depth + 1);
5017 if (MaxShAmt < Tmp)
5018 return Tmp - MaxShAmt;
5019 }
5020 }
5021 // shl destroys sign bits, ensure it doesn't shift out all sign bits.
5022 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5023 if (MaxShAmt < Tmp)
5024 return Tmp - MaxShAmt;
5025 }
5026 break;
5027 case ISD::AND:
5028 case ISD::OR:
5029 case ISD::XOR: // NOT is handled here.
5030 // Logical binary ops preserve the number of sign bits at the worst.
5031 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth+1);
5032 if (Tmp != 1) {
5033 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth+1);
5034 FirstAnswer = std::min(Tmp, Tmp2);
5035 // We computed what we know about the sign bits as our first
5036 // answer. Now proceed to the generic code that uses
5037 // computeKnownBits, and pick whichever answer is better.
5038 }
5039 break;
5040
5041 case ISD::SELECT:
5042 case ISD::VSELECT:
5043 Tmp = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth+1);
5044 if (Tmp == 1) return 1; // Early out.
5045 Tmp2 = ComputeNumSignBits(Op.getOperand(2), DemandedElts, Depth+1);
5046 return std::min(Tmp, Tmp2);
5047 case ISD::SELECT_CC:
5048 Tmp = ComputeNumSignBits(Op.getOperand(2), DemandedElts, Depth+1);
5049 if (Tmp == 1) return 1; // Early out.
5050 Tmp2 = ComputeNumSignBits(Op.getOperand(3), DemandedElts, Depth+1);
5051 return std::min(Tmp, Tmp2);
5052
5053 case ISD::SMIN:
5054 case ISD::SMAX: {
5055 // If we have a clamp pattern, we know that the number of sign bits will be
5056 // the minimum of the clamp min/max range.
5057 bool IsMax = (Opcode == ISD::SMAX);
5058 ConstantSDNode *CstLow = nullptr, *CstHigh = nullptr;
5059 if ((CstLow = isConstOrConstSplat(Op.getOperand(1), DemandedElts)))
5060 if (Op.getOperand(0).getOpcode() == (IsMax ? ISD::SMIN : ISD::SMAX))
5061 CstHigh =
5062 isConstOrConstSplat(Op.getOperand(0).getOperand(1), DemandedElts);
5063 if (CstLow && CstHigh) {
5064 if (!IsMax)
5065 std::swap(CstLow, CstHigh);
5066 if (CstLow->getAPIntValue().sle(CstHigh->getAPIntValue())) {
5067 Tmp = CstLow->getAPIntValue().getNumSignBits();
5068 Tmp2 = CstHigh->getAPIntValue().getNumSignBits();
5069 return std::min(Tmp, Tmp2);
5070 }
5071 }
5072
5073 // Fallback - just get the minimum number of sign bits of the operands.
5074 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5075 if (Tmp == 1)
5076 return 1; // Early out.
5077 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5078 return std::min(Tmp, Tmp2);
5079 }
5080 case ISD::UMIN:
5081 case ISD::UMAX:
5082 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5083 if (Tmp == 1)
5084 return 1; // Early out.
5085 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5086 return std::min(Tmp, Tmp2);
5087 case ISD::SSUBO_CARRY:
5088 case ISD::USUBO_CARRY:
5089 // sub_carry(x,x,c) -> 0/-1 (sext carry)
5090 if (Op.getResNo() == 0 && Op.getOperand(0) == Op.getOperand(1))
5091 return VTBits;
5092 [[fallthrough]];
5093 case ISD::SADDO:
5094 case ISD::UADDO:
5095 case ISD::SADDO_CARRY:
5096 case ISD::UADDO_CARRY:
5097 case ISD::SSUBO:
5098 case ISD::USUBO:
5099 case ISD::SMULO:
5100 case ISD::UMULO:
5101 if (Op.getResNo() != 1)
5102 break;
5103 // The boolean result conforms to getBooleanContents. Fall through.
5104 // If setcc returns 0/-1, all bits are sign bits.
5105 // We know that we have an integer-based boolean since these operations
5106 // are only available for integer.
5107 if (TLI->getBooleanContents(VT.isVector(), false) ==
5109 return VTBits;
5110 break;
5111 case ISD::SETCC:
5112 case ISD::SETCCCARRY:
5113 case ISD::STRICT_FSETCC:
5114 case ISD::STRICT_FSETCCS: {
5115 unsigned OpNo = Op->isStrictFPOpcode() ? 1 : 0;
5116 // If setcc returns 0/-1, all bits are sign bits.
5117 if (TLI->getBooleanContents(Op.getOperand(OpNo).getValueType()) ==
5119 return VTBits;
5120 break;
5121 }
5122 case ISD::ROTL:
5123 case ISD::ROTR:
5124 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5125
5126 // If we're rotating an 0/-1 value, then it stays an 0/-1 value.
5127 if (Tmp == VTBits)
5128 return VTBits;
5129
5130 if (ConstantSDNode *C =
5131 isConstOrConstSplat(Op.getOperand(1), DemandedElts)) {
5132 unsigned RotAmt = C->getAPIntValue().urem(VTBits);
5133
5134 // Handle rotate right by N like a rotate left by 32-N.
5135 if (Opcode == ISD::ROTR)
5136 RotAmt = (VTBits - RotAmt) % VTBits;
5137
5138 // If we aren't rotating out all of the known-in sign bits, return the
5139 // number that are left. This handles rotl(sext(x), 1) for example.
5140 if (Tmp > (RotAmt + 1)) return (Tmp - RotAmt);
5141 }
5142 break;
5143 case ISD::ADD:
5144 case ISD::ADDC:
5145 // TODO: Move Operand 1 check before Operand 0 check
5146 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5147 if (Tmp == 1) return 1; // Early out.
5148
5149 // Special case decrementing a value (ADD X, -1):
5150 if (ConstantSDNode *CRHS =
5151 isConstOrConstSplat(Op.getOperand(1), DemandedElts))
5152 if (CRHS->isAllOnes()) {
5153 KnownBits Known =
5154 computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
5155
5156 // If the input is known to be 0 or 1, the output is 0/-1, which is all
5157 // sign bits set.
5158 if ((Known.Zero | 1).isAllOnes())
5159 return VTBits;
5160
5161 // If we are subtracting one from a positive number, there is no carry
5162 // out of the result.
5163 if (Known.isNonNegative())
5164 return Tmp;
5165 }
5166
5167 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5168 if (Tmp2 == 1) return 1; // Early out.
5169
5170 // Add can have at most one carry bit. Thus we know that the output
5171 // is, at worst, one more bit than the inputs.
5172 return std::min(Tmp, Tmp2) - 1;
5173 case ISD::SUB:
5174 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5175 if (Tmp2 == 1) return 1; // Early out.
5176
5177 // Handle NEG.
5178 if (ConstantSDNode *CLHS =
5179 isConstOrConstSplat(Op.getOperand(0), DemandedElts))
5180 if (CLHS->isZero()) {
5181 KnownBits Known =
5182 computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
5183 // If the input is known to be 0 or 1, the output is 0/-1, which is all
5184 // sign bits set.
5185 if ((Known.Zero | 1).isAllOnes())
5186 return VTBits;
5187
5188 // If the input is known to be positive (the sign bit is known clear),
5189 // the output of the NEG has the same number of sign bits as the input.
5190 if (Known.isNonNegative())
5191 return Tmp2;
5192
5193 // Otherwise, we treat this like a SUB.
5194 }
5195
5196 // Sub can have at most one carry bit. Thus we know that the output
5197 // is, at worst, one more bit than the inputs.
5198 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5199 if (Tmp == 1) return 1; // Early out.
5200 return std::min(Tmp, Tmp2) - 1;
5201 case ISD::MUL: {
5202 // The output of the Mul can be at most twice the valid bits in the inputs.
5203 unsigned SignBitsOp0 = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
5204 if (SignBitsOp0 == 1)
5205 break;
5206 unsigned SignBitsOp1 = ComputeNumSignBits(Op.getOperand(1), Depth + 1);
5207 if (SignBitsOp1 == 1)
5208 break;
5209 unsigned OutValidBits =
5210 (VTBits - SignBitsOp0 + 1) + (VTBits - SignBitsOp1 + 1);
5211 return OutValidBits > VTBits ? 1 : VTBits - OutValidBits + 1;
5212 }
5213 case ISD::AVGCEILS:
5214 case ISD::AVGFLOORS:
5215 Tmp = ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5216 if (Tmp == 1)
5217 return 1; // Early out.
5218 Tmp2 = ComputeNumSignBits(Op.getOperand(1), DemandedElts, Depth + 1);
5219 return std::min(Tmp, Tmp2);
5220 case ISD::SREM:
5221 // The sign bit is the LHS's sign bit, except when the result of the
5222 // remainder is zero. The magnitude of the result should be less than or
5223 // equal to the magnitude of the LHS. Therefore, the result should have
5224 // at least as many sign bits as the left hand side.
5225 return ComputeNumSignBits(Op.getOperand(0), DemandedElts, Depth + 1);
5226 case ISD::TRUNCATE: {
5227 // Check if the sign bits of source go down as far as the truncated value.
5228 unsigned NumSrcBits = Op.getOperand(0).getScalarValueSizeInBits();
5229 unsigned NumSrcSignBits = ComputeNumSignBits(Op.getOperand(0), Depth + 1);
5230 if (NumSrcSignBits > (NumSrcBits - VTBits))
5231 return NumSrcSignBits - (NumSrcBits - VTBits);
5232 break;
5233 }
5234 case ISD::EXTRACT_ELEMENT: {
5235 if (VT.isScalableVector())
5236 break;
5237 const int KnownSign = ComputeNumSignBits(Op.getOperand(0), Depth+1);
5238 const int BitWidth = Op.getValueSizeInBits();
5239 const int Items = Op.getOperand(0).getValueSizeInBits() / BitWidth;
5240
5241 // Get reverse index (starting from 1), Op1 value indexes elements from
5242 // little end. Sign starts at big end.
5243 const int rIndex = Items - 1 - Op.getConstantOperandVal(1);
5244
5245 // If the sign portion ends in our element the subtraction gives correct
5246 // result. Otherwise it gives either negative or > bitwidth result
5247 return std::clamp(KnownSign - rIndex * BitWidth, 1, BitWidth);
5248 }
5250 if (VT.isScalableVector())
5251 break;
5252 // If we know the element index, split the demand between the
5253 // source vector and the inserted element, otherwise assume we need
5254 // the original demanded vector elements and the value.
5255 SDValue InVec = Op.getOperand(0);
5256 SDValue InVal = Op.getOperand(1);
5257 SDValue EltNo = Op.getOperand(2);
5258 bool DemandedVal = true;
5259 APInt DemandedVecElts = DemandedElts;
5260 auto *CEltNo = dyn_cast<ConstantSDNode>(EltNo);
5261 if (CEltNo && CEltNo->getAPIntValue().ult(NumElts)) {
5262 unsigned EltIdx = CEltNo->getZExtValue();
5263 DemandedVal = !!DemandedElts[EltIdx];
5264 DemandedVecElts.clearBit(EltIdx);
5265 }
5266 Tmp = std::numeric_limits<unsigned>::max();
5267 if (DemandedVal) {
5268 // TODO - handle implicit truncation of inserted elements.
5269 if (InVal.getScalarValueSizeInBits() != VTBits)
5270 break;
5271 Tmp2 = ComputeNumSignBits(InVal, Depth + 1);
5272 Tmp = std::min(Tmp, Tmp2);
5273 }
5274 if (!!DemandedVecElts) {
5275 Tmp2 = ComputeNumSignBits(InVec, DemandedVecElts, Depth + 1);
5276 Tmp = std::min(Tmp, Tmp2);
5277 }
5278 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
5279 return Tmp;
5280 }
5282 SDValue InVec = Op.getOperand(0);
5283 SDValue EltNo = Op.getOperand(1);
5284 EVT VecVT = InVec.getValueType();
5285 // ComputeNumSignBits not yet implemented for scalable vectors.
5286 if (VecVT.isScalableVector())
5287 break;
5288 const unsigned BitWidth = Op.getValueSizeInBits();
5289 const unsigned EltBitWidth = Op.getOperand(0).getScalarValueSizeInBits();
5290 const unsigned NumSrcElts = VecVT.getVectorNumElements();
5291
5292 // If BitWidth > EltBitWidth the value is anyext:ed, and we do not know
5293 // anything about sign bits. But if the sizes match we can derive knowledge
5294 // about sign bits from the vector operand.
5295 if (BitWidth != EltBitWidth)
5296 break;
5297
5298 // If we know the element index, just demand that vector element, else for
5299 // an unknown element index, ignore DemandedElts and demand them all.
5300 APInt DemandedSrcElts = APInt::getAllOnes(NumSrcElts);
5301 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
5302 if (ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts))
5303 DemandedSrcElts =
5304 APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue());
5305
5306 return ComputeNumSignBits(InVec, DemandedSrcElts, Depth + 1);
5307 }
5309 // Offset the demanded elts by the subvector index.
5310 SDValue Src = Op.getOperand(0);
5311
5312 APInt DemandedSrcElts;
5313 if (Src.getValueType().isScalableVector())
5314 DemandedSrcElts = APInt(1, 1);
5315 else {
5316 uint64_t Idx = Op.getConstantOperandVal(1);
5317 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
5318 DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
5319 }
5320 return ComputeNumSignBits(Src, DemandedSrcElts, Depth + 1);
5321 }
5322 case ISD::CONCAT_VECTORS: {
5323 if (VT.isScalableVector())
5324 break;
5325 // Determine the minimum number of sign bits across all demanded
5326 // elts of the input vectors. Early out if the result is already 1.
5327 Tmp = std::numeric_limits<unsigned>::max();
5328 EVT SubVectorVT = Op.getOperand(0).getValueType();
5329 unsigned NumSubVectorElts = SubVectorVT.getVectorNumElements();
5330 unsigned NumSubVectors = Op.getNumOperands();
5331 for (unsigned i = 0; (i < NumSubVectors) && (Tmp > 1); ++i) {
5332 APInt DemandedSub =
5333 DemandedElts.extractBits(NumSubVectorElts, i * NumSubVectorElts);
5334 if (!DemandedSub)
5335 continue;
5336 Tmp2 = ComputeNumSignBits(Op.getOperand(i), DemandedSub, Depth + 1);
5337 Tmp = std::min(Tmp, Tmp2);
5338 }
5339 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
5340 return Tmp;
5341 }
5342 case ISD::INSERT_SUBVECTOR: {
5343 if (VT.isScalableVector())
5344 break;
5345 // Demand any elements from the subvector and the remainder from the src its
5346 // inserted into.
5347 SDValue Src = Op.getOperand(0);
5348 SDValue Sub = Op.getOperand(1);
5349 uint64_t Idx = Op.getConstantOperandVal(2);
5350 unsigned NumSubElts = Sub.getValueType().getVectorNumElements();
5351 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
5352 APInt DemandedSrcElts = DemandedElts;
5353 DemandedSrcElts.clearBits(Idx, Idx + NumSubElts);
5354
5355 Tmp = std::numeric_limits<unsigned>::max();
5356 if (!!DemandedSubElts) {
5357 Tmp = ComputeNumSignBits(Sub, DemandedSubElts, Depth + 1);
5358 if (Tmp == 1)
5359 return 1; // early-out
5360 }
5361 if (!!DemandedSrcElts) {
5362 Tmp2 = ComputeNumSignBits(Src, DemandedSrcElts, Depth + 1);
5363 Tmp = std::min(Tmp, Tmp2);
5364 }
5365 assert(Tmp <= VTBits && "Failed to determine minimum sign bits");
5366 return Tmp;
5367 }
5368 case ISD::LOAD: {
5369 // If we are looking at the loaded value of the SDNode.
5370 if (Op.getResNo() != 0)
5371 break;
5372
5374 if (const MDNode *Ranges = LD->getRanges()) {
5375 if (DemandedElts != 1)
5376 break;
5377
5379 if (VTBits > CR.getBitWidth()) {
5380 switch (LD->getExtensionType()) {
5381 case ISD::SEXTLOAD:
5382 CR = CR.signExtend(VTBits);
5383 break;
5384 case ISD::ZEXTLOAD:
5385 CR = CR.zeroExtend(VTBits);
5386 break;
5387 default:
5388 break;
5389 }
5390 }
5391
5392 if (VTBits != CR.getBitWidth())
5393 break;
5394 return std::min(CR.getSignedMin().getNumSignBits(),
5396 }
5397
5398 unsigned ExtType = LD->getExtensionType();
5399 switch (ExtType) {
5400 default:
5401 break;
5402 case ISD::SEXTLOAD: // e.g. i16->i32 = '17' bits known.
5403 Tmp = LD->getMemoryVT().getScalarSizeInBits();
5404 return VTBits - Tmp + 1;
5405 case ISD::ZEXTLOAD: // e.g. i16->i32 = '16' bits known.
5406 Tmp = LD->getMemoryVT().getScalarSizeInBits();
5407 return VTBits - Tmp;
5408 case ISD::NON_EXTLOAD:
5409 if (const Constant *Cst = TLI->getTargetConstantFromLoad(LD)) {
5410 // We only need to handle vectors - computeKnownBits should handle
5411 // scalar cases.
5412 Type *CstTy = Cst->getType();
5413 if (CstTy->isVectorTy() && !VT.isScalableVector() &&
5414 (NumElts * VTBits) == CstTy->getPrimitiveSizeInBits() &&
5415 VTBits == CstTy->getScalarSizeInBits()) {
5416 Tmp = VTBits;
5417 for (unsigned i = 0; i != NumElts; ++i) {
5418 if (!DemandedElts[i])
5419 continue;
5420 if (Constant *Elt = Cst->getAggregateElement(i)) {
5421 if (auto *CInt = dyn_cast<ConstantInt>(Elt)) {
5422 const APInt &Value = CInt->getValue();
5423 Tmp = std::min(Tmp, Value.getNumSignBits());
5424 continue;
5425 }
5426 if (auto *CFP = dyn_cast<ConstantFP>(Elt)) {
5427 APInt Value = CFP->getValueAPF().bitcastToAPInt();
5428 Tmp = std::min(Tmp, Value.getNumSignBits());
5429 continue;
5430 }
5431 }
5432 // Unknown type. Conservatively assume no bits match sign bit.
5433 return 1;
5434 }
5435 return Tmp;
5436 }
5437 }
5438 break;
5439 }
5440
5441 break;
5442 }
5445 case ISD::ATOMIC_SWAP:
5457 case ISD::ATOMIC_LOAD: {
5458 auto *AT = cast<AtomicSDNode>(Op);
5459 // If we are looking at the loaded value.
5460 if (Op.getResNo() == 0) {
5461 Tmp = AT->getMemoryVT().getScalarSizeInBits();
5462 if (Tmp == VTBits)
5463 return 1; // early-out
5464
5465 // For atomic_load, prefer to use the extension type.
5466 if (Op->getOpcode() == ISD::ATOMIC_LOAD) {
5467 switch (AT->getExtensionType()) {
5468 default:
5469 break;
5470 case ISD::SEXTLOAD:
5471 return VTBits - Tmp + 1;
5472 case ISD::ZEXTLOAD:
5473 return VTBits - Tmp;
5474 }
5475 }
5476
5477 if (TLI->getExtendForAtomicOps() == ISD::SIGN_EXTEND)
5478 return VTBits - Tmp + 1;
5479 if (TLI->getExtendForAtomicOps() == ISD::ZERO_EXTEND)
5480 return VTBits - Tmp;
5481 }
5482 break;
5483 }
5484 }
5485
5486 // Allow the target to implement this method for its nodes.
5487 if (Opcode >= ISD::BUILTIN_OP_END ||
5488 Opcode == ISD::INTRINSIC_WO_CHAIN ||
5489 Opcode == ISD::INTRINSIC_W_CHAIN ||
5490 Opcode == ISD::INTRINSIC_VOID) {
5491 // TODO: This can probably be removed once target code is audited. This
5492 // is here purely to reduce patch size and review complexity.
5493 if (!VT.isScalableVector()) {
5494 unsigned NumBits =
5495 TLI->ComputeNumSignBitsForTargetNode(Op, DemandedElts, *this, Depth);
5496 if (NumBits > 1)
5497 FirstAnswer = std::max(FirstAnswer, NumBits);
5498 }
5499 }
5500
5501 // Finally, if we can prove that the top bits of the result are 0's or 1's,
5502 // use this information.
5503 KnownBits Known = computeKnownBits(Op, DemandedElts, Depth);
5504 return std::max(FirstAnswer, Known.countMinSignBits());
5505}
5506
5508 unsigned Depth) const {
5509 unsigned SignBits = ComputeNumSignBits(Op, Depth);
5510 return Op.getScalarValueSizeInBits() - SignBits + 1;
5511}
5512
5514 const APInt &DemandedElts,
5515 unsigned Depth) const {
5516 unsigned SignBits = ComputeNumSignBits(Op, DemandedElts, Depth);
5517 return Op.getScalarValueSizeInBits() - SignBits + 1;
5518}
5519
5521 unsigned Depth) const {
5522 // Early out for FREEZE.
5523 if (Op.getOpcode() == ISD::FREEZE)
5524 return true;
5525
5526 EVT VT = Op.getValueType();
5527 APInt DemandedElts = VT.isFixedLengthVector()
5529 : APInt(1, 1);
5530 return isGuaranteedNotToBeUndefOrPoison(Op, DemandedElts, PoisonOnly, Depth);
5531}
5532
5534 const APInt &DemandedElts,
5535 bool PoisonOnly,
5536 unsigned Depth) const {
5537 unsigned Opcode = Op.getOpcode();
5538
5539 // Early out for FREEZE.
5540 if (Opcode == ISD::FREEZE)
5541 return true;
5542
5543 if (Depth >= MaxRecursionDepth)
5544 return false; // Limit search depth.
5545
5546 if (isIntOrFPConstant(Op))
5547 return true;
5548
5549 switch (Opcode) {
5550 case ISD::CONDCODE:
5551 case ISD::VALUETYPE:
5552 case ISD::FrameIndex:
5554 case ISD::CopyFromReg:
5555 return true;
5556
5557 case ISD::POISON:
5558 return false;
5559
5560 case ISD::UNDEF:
5561 return PoisonOnly;
5562
5563 case ISD::BUILD_VECTOR:
5564 // NOTE: BUILD_VECTOR has implicit truncation of wider scalar elements -
5565 // this shouldn't affect the result.
5566 for (unsigned i = 0, e = Op.getNumOperands(); i < e; ++i) {
5567 if (!DemandedElts[i])
5568 continue;
5570 Depth + 1))
5571 return false;
5572 }
5573 return true;
5574
5576 SDValue Src = Op.getOperand(0);
5577 if (Src.getValueType().isScalableVector())
5578 break;
5579 uint64_t Idx = Op.getConstantOperandVal(1);
5580 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
5581 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
5582 return isGuaranteedNotToBeUndefOrPoison(Src, DemandedSrcElts, PoisonOnly,
5583 Depth + 1);
5584 }
5585
5586 case ISD::INSERT_SUBVECTOR: {
5587 if (Op.getValueType().isScalableVector())
5588 break;
5589 SDValue Src = Op.getOperand(0);
5590 SDValue Sub = Op.getOperand(1);
5591 uint64_t Idx = Op.getConstantOperandVal(2);
5592 unsigned NumSubElts = Sub.getValueType().getVectorNumElements();
5593 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
5594 APInt DemandedSrcElts = DemandedElts;
5595 DemandedSrcElts.clearBits(Idx, Idx + NumSubElts);
5596
5597 if (!!DemandedSubElts && !isGuaranteedNotToBeUndefOrPoison(
5598 Sub, DemandedSubElts, PoisonOnly, Depth + 1))
5599 return false;
5600 if (!!DemandedSrcElts && !isGuaranteedNotToBeUndefOrPoison(
5601 Src, DemandedSrcElts, PoisonOnly, Depth + 1))
5602 return false;
5603 return true;
5604 }
5605
5607 SDValue Src = Op.getOperand(0);
5608 auto *IndexC = dyn_cast<ConstantSDNode>(Op.getOperand(1));
5609 EVT SrcVT = Src.getValueType();
5610 if (SrcVT.isFixedLengthVector() && IndexC &&
5611 IndexC->getAPIntValue().ult(SrcVT.getVectorNumElements())) {
5612 APInt DemandedSrcElts = APInt::getOneBitSet(SrcVT.getVectorNumElements(),
5613 IndexC->getZExtValue());
5614 return isGuaranteedNotToBeUndefOrPoison(Src, DemandedSrcElts, PoisonOnly,
5615 Depth + 1);
5616 }
5617 break;
5618 }
5619
5621 SDValue InVec = Op.getOperand(0);
5622 SDValue InVal = Op.getOperand(1);
5623 SDValue EltNo = Op.getOperand(2);
5624 EVT VT = InVec.getValueType();
5625 auto *IndexC = dyn_cast<ConstantSDNode>(EltNo);
5626 if (IndexC && VT.isFixedLengthVector() &&
5627 IndexC->getAPIntValue().ult(VT.getVectorNumElements())) {
5628 if (DemandedElts[IndexC->getZExtValue()] &&
5630 return false;
5631 APInt InVecDemandedElts = DemandedElts;
5632 InVecDemandedElts.clearBit(IndexC->getZExtValue());
5633 if (!!InVecDemandedElts &&
5635 peekThroughInsertVectorElt(InVec, InVecDemandedElts),
5636 InVecDemandedElts, PoisonOnly, Depth + 1))
5637 return false;
5638 return true;
5639 }
5640 break;
5641 }
5642
5644 // Check upper (known undef) elements.
5645 if (DemandedElts.ugt(1) && !PoisonOnly)
5646 return false;
5647 // Check element zero.
5648 if (DemandedElts[0] && !isGuaranteedNotToBeUndefOrPoison(
5649 Op.getOperand(0), PoisonOnly, Depth + 1))
5650 return false;
5651 return true;
5652
5653 case ISD::SPLAT_VECTOR:
5654 return isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), PoisonOnly,
5655 Depth + 1);
5656
5657 case ISD::VECTOR_SHUFFLE: {
5658 APInt DemandedLHS, DemandedRHS;
5659 auto *SVN = cast<ShuffleVectorSDNode>(Op);
5660 if (!getShuffleDemandedElts(DemandedElts.getBitWidth(), SVN->getMask(),
5661 DemandedElts, DemandedLHS, DemandedRHS,
5662 /*AllowUndefElts=*/false))
5663 return false;
5664 if (!DemandedLHS.isZero() &&
5665 !isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), DemandedLHS,
5666 PoisonOnly, Depth + 1))
5667 return false;
5668 if (!DemandedRHS.isZero() &&
5669 !isGuaranteedNotToBeUndefOrPoison(Op.getOperand(1), DemandedRHS,
5670 PoisonOnly, Depth + 1))
5671 return false;
5672 return true;
5673 }
5674
5675 case ISD::SHL:
5676 case ISD::SRL:
5677 case ISD::SRA:
5678 // Shift amount operand is checked by canCreateUndefOrPoison. So it is
5679 // enough to check operand 0 if Op can't create undef/poison.
5680 return !canCreateUndefOrPoison(Op, DemandedElts, PoisonOnly,
5681 /*ConsiderFlags*/ true, Depth) &&
5682 isGuaranteedNotToBeUndefOrPoison(Op.getOperand(0), DemandedElts,
5683 PoisonOnly, Depth + 1);
5684
5685 case ISD::BSWAP:
5686 case ISD::CTPOP:
5687 case ISD::BITREVERSE:
5688 case ISD::AND:
5689 case ISD::OR:
5690 case ISD::XOR:
5691 case ISD::ADD:
5692 case ISD::SUB:
5693 case ISD::MUL:
5694 case ISD::SADDSAT:
5695 case ISD::UADDSAT:
5696 case ISD::SSUBSAT:
5697 case ISD::USUBSAT:
5698 case ISD::SSHLSAT:
5699 case ISD::USHLSAT:
5700 case ISD::SMIN:
5701 case ISD::SMAX:
5702 case ISD::UMIN:
5703 case ISD::UMAX:
5704 case ISD::ZERO_EXTEND:
5705 case ISD::SIGN_EXTEND:
5706 case ISD::ANY_EXTEND:
5707 case ISD::TRUNCATE:
5708 case ISD::VSELECT: {
5709 // If Op can't create undef/poison and none of its operands are undef/poison
5710 // then Op is never undef/poison. A difference from the more common check
5711 // below, outside the switch, is that we handle elementwise operations for
5712 // which the DemandedElts mask is valid for all operands here.
5713 return !canCreateUndefOrPoison(Op, DemandedElts, PoisonOnly,
5714 /*ConsiderFlags*/ true, Depth) &&
5715 all_of(Op->ops(), [&](SDValue V) {
5716 return isGuaranteedNotToBeUndefOrPoison(V, DemandedElts,
5717 PoisonOnly, Depth + 1);
5718 });
5719 }
5720
5721 // TODO: Search for noundef attributes from library functions.
5722
5723 // TODO: Pointers dereferenced by ISD::LOAD/STORE ops are noundef.
5724
5725 default:
5726 // Allow the target to implement this method for its nodes.
5727 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
5728 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID)
5729 return TLI->isGuaranteedNotToBeUndefOrPoisonForTargetNode(
5730 Op, DemandedElts, *this, PoisonOnly, Depth);
5731 break;
5732 }
5733
5734 // If Op can't create undef/poison and none of its operands are undef/poison
5735 // then Op is never undef/poison.
5736 // NOTE: TargetNodes can handle this in themselves in
5737 // isGuaranteedNotToBeUndefOrPoisonForTargetNode or let
5738 // TargetLowering::isGuaranteedNotToBeUndefOrPoisonForTargetNode handle it.
5739 return !canCreateUndefOrPoison(Op, PoisonOnly, /*ConsiderFlags*/ true,
5740 Depth) &&
5741 all_of(Op->ops(), [&](SDValue V) {
5742 return isGuaranteedNotToBeUndefOrPoison(V, PoisonOnly, Depth + 1);
5743 });
5744}
5745
5747 bool ConsiderFlags,
5748 unsigned Depth) const {
5749 EVT VT = Op.getValueType();
5750 APInt DemandedElts = VT.isFixedLengthVector()
5752 : APInt(1, 1);
5753 return canCreateUndefOrPoison(Op, DemandedElts, PoisonOnly, ConsiderFlags,
5754 Depth);
5755}
5756
5758 bool PoisonOnly, bool ConsiderFlags,
5759 unsigned Depth) const {
5760 if (ConsiderFlags && Op->hasPoisonGeneratingFlags())
5761 return true;
5762
5763 unsigned Opcode = Op.getOpcode();
5764 switch (Opcode) {
5765 case ISD::AssertSext:
5766 case ISD::AssertZext:
5767 case ISD::AssertAlign:
5769 // Assertion nodes can create poison if the assertion fails.
5770 return true;
5771
5772 case ISD::FREEZE:
5776 case ISD::SADDSAT:
5777 case ISD::UADDSAT:
5778 case ISD::SSUBSAT:
5779 case ISD::USUBSAT:
5780 case ISD::MULHU:
5781 case ISD::MULHS:
5782 case ISD::AVGFLOORS:
5783 case ISD::AVGFLOORU:
5784 case ISD::AVGCEILS:
5785 case ISD::AVGCEILU:
5786 case ISD::ABDU:
5787 case ISD::ABDS:
5788 case ISD::SMIN:
5789 case ISD::SMAX:
5790 case ISD::SCMP:
5791 case ISD::UMIN:
5792 case ISD::UMAX:
5793 case ISD::UCMP:
5794 case ISD::AND:
5795 case ISD::XOR:
5796 case ISD::ROTL:
5797 case ISD::ROTR:
5798 case ISD::FSHL:
5799 case ISD::FSHR:
5800 case ISD::BSWAP:
5801 case ISD::CTTZ:
5802 case ISD::CTLZ:
5803 case ISD::CTLS:
5804 case ISD::CTPOP:
5805 case ISD::BITREVERSE:
5806 case ISD::PARITY:
5807 case ISD::SIGN_EXTEND:
5808 case ISD::TRUNCATE:
5812 case ISD::BITCAST:
5813 case ISD::BUILD_VECTOR:
5814 case ISD::BUILD_PAIR:
5815 case ISD::SPLAT_VECTOR:
5816 case ISD::FABS:
5817 return false;
5818
5819 case ISD::ABS:
5820 // ISD::ABS defines abs(INT_MIN) -> INT_MIN and never generates poison.
5821 // Different to Intrinsic::abs.
5822 return false;
5823
5824 case ISD::ADDC:
5825 case ISD::SUBC:
5826 case ISD::ADDE:
5827 case ISD::SUBE:
5828 case ISD::SADDO:
5829 case ISD::SSUBO:
5830 case ISD::SMULO:
5831 case ISD::SADDO_CARRY:
5832 case ISD::SSUBO_CARRY:
5833 case ISD::UADDO:
5834 case ISD::USUBO:
5835 case ISD::UMULO:
5836 case ISD::UADDO_CARRY:
5837 case ISD::USUBO_CARRY:
5838 // No poison on result or overflow flags.
5839 return false;
5840
5841 case ISD::SELECT_CC:
5842 case ISD::SETCC: {
5843 // Integer setcc cannot create undef or poison.
5844 if (Op.getOperand(0).getValueType().isInteger())
5845 return false;
5846
5847 // FP compares are more complicated. They can create poison for nan/infinity
5848 // based on options and flags. The options and flags also cause special
5849 // nonan condition codes to be used. Those condition codes may be preserved
5850 // even if the nonan flag is dropped somewhere.
5851 unsigned CCOp = Opcode == ISD::SETCC ? 2 : 4;
5852 ISD::CondCode CCCode = cast<CondCodeSDNode>(Op.getOperand(CCOp))->get();
5853 return (unsigned)CCCode & 0x10U;
5854 }
5855
5856 case ISD::OR:
5857 case ISD::ZERO_EXTEND:
5858 case ISD::SELECT:
5859 case ISD::VSELECT:
5860 case ISD::ADD:
5861 case ISD::SUB:
5862 case ISD::MUL:
5863 case ISD::FNEG:
5864 case ISD::FADD:
5865 case ISD::FSUB:
5866 case ISD::FMUL:
5867 case ISD::FDIV:
5868 case ISD::FREM:
5869 case ISD::FCOPYSIGN:
5870 case ISD::FMA:
5871 case ISD::FMAD:
5872 case ISD::FMULADD:
5873 case ISD::FP_EXTEND:
5879 // No poison except from flags (which is handled above)
5880 return false;
5881
5882 case ISD::SHL:
5883 case ISD::SRL:
5884 case ISD::SRA:
5885 // If the max shift amount isn't in range, then the shift can
5886 // create poison.
5887 return !getValidMaximumShiftAmount(Op, DemandedElts, Depth + 1);
5888
5891 // If the amount is zero then the result will be poison.
5892 // TODO: Add isKnownNeverZero DemandedElts handling.
5893 return !isKnownNeverZero(Op.getOperand(0), Depth + 1);
5894
5896 // Check if we demand any upper (undef) elements.
5897 return !PoisonOnly && DemandedElts.ugt(1);
5898
5901 // Ensure that the element index is in bounds.
5902 EVT VecVT = Op.getOperand(0).getValueType();
5903 SDValue Idx = Op.getOperand(Opcode == ISD::INSERT_VECTOR_ELT ? 2 : 1);
5904 KnownBits KnownIdx = computeKnownBits(Idx, Depth + 1);
5905 return KnownIdx.getMaxValue().uge(VecVT.getVectorMinNumElements());
5906 }
5907
5908 case ISD::VECTOR_SHUFFLE: {
5909 // Check for any demanded shuffle element that is undef.
5910 auto *SVN = cast<ShuffleVectorSDNode>(Op);
5911 for (auto [Idx, Elt] : enumerate(SVN->getMask()))
5912 if (Elt < 0 && DemandedElts[Idx])
5913 return true;
5914 return false;
5915 }
5916
5918 return false;
5919
5920 default:
5921 // Allow the target to implement this method for its nodes.
5922 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
5923 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID)
5924 return TLI->canCreateUndefOrPoisonForTargetNode(
5925 Op, DemandedElts, *this, PoisonOnly, ConsiderFlags, Depth);
5926 break;
5927 }
5928
5929 // Be conservative and return true.
5930 return true;
5931}
5932
5933bool SelectionDAG::isADDLike(SDValue Op, bool NoWrap) const {
5934 unsigned Opcode = Op.getOpcode();
5935 if (Opcode == ISD::OR)
5936 return Op->getFlags().hasDisjoint() ||
5937 haveNoCommonBitsSet(Op.getOperand(0), Op.getOperand(1));
5938 if (Opcode == ISD::XOR)
5939 return !NoWrap && isMinSignedConstant(Op.getOperand(1));
5940 return false;
5941}
5942
5944 return Op.getNumOperands() == 2 && isa<ConstantSDNode>(Op.getOperand(1)) &&
5945 (Op.isAnyAdd() || isADDLike(Op));
5946}
5947
5949 unsigned Depth) const {
5950 EVT VT = Op.getValueType();
5951
5952 // Since the number of lanes in a scalable vector is unknown at compile time,
5953 // we track one bit which is implicitly broadcast to all lanes. This means
5954 // that all lanes in a scalable vector are considered demanded.
5955 APInt DemandedElts = VT.isFixedLengthVector()
5957 : APInt(1, 1);
5958
5959 return isKnownNeverNaN(Op, DemandedElts, SNaN, Depth);
5960}
5961
5963 bool SNaN, unsigned Depth) const {
5964 assert(!DemandedElts.isZero() && "No demanded elements");
5965
5966 // If we're told that NaNs won't happen, assume they won't.
5967 if (getTarget().Options.NoNaNsFPMath || Op->getFlags().hasNoNaNs())
5968 return true;
5969
5970 if (Depth >= MaxRecursionDepth)
5971 return false; // Limit search depth.
5972
5973 // If the value is a constant, we can obviously see if it is a NaN or not.
5975 return !C->getValueAPF().isNaN() ||
5976 (SNaN && !C->getValueAPF().isSignaling());
5977 }
5978
5979 unsigned Opcode = Op.getOpcode();
5980 switch (Opcode) {
5981 case ISD::FADD:
5982 case ISD::FSUB:
5983 case ISD::FMUL:
5984 case ISD::FDIV:
5985 case ISD::FREM:
5986 case ISD::FSIN:
5987 case ISD::FCOS:
5988 case ISD::FTAN:
5989 case ISD::FASIN:
5990 case ISD::FACOS:
5991 case ISD::FATAN:
5992 case ISD::FATAN2:
5993 case ISD::FSINH:
5994 case ISD::FCOSH:
5995 case ISD::FTANH:
5996 case ISD::FMA:
5997 case ISD::FMULADD:
5998 case ISD::FMAD: {
5999 if (SNaN)
6000 return true;
6001 // TODO: Need isKnownNeverInfinity
6002 return false;
6003 }
6004 case ISD::FCANONICALIZE:
6005 case ISD::FEXP:
6006 case ISD::FEXP2:
6007 case ISD::FEXP10:
6008 case ISD::FTRUNC:
6009 case ISD::FFLOOR:
6010 case ISD::FCEIL:
6011 case ISD::FROUND:
6012 case ISD::FROUNDEVEN:
6013 case ISD::LROUND:
6014 case ISD::LLROUND:
6015 case ISD::FRINT:
6016 case ISD::LRINT:
6017 case ISD::LLRINT:
6018 case ISD::FNEARBYINT:
6019 case ISD::FLDEXP: {
6020 if (SNaN)
6021 return true;
6022 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6023 }
6024 case ISD::FABS:
6025 case ISD::FNEG:
6026 case ISD::FCOPYSIGN: {
6027 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6028 }
6029 case ISD::SELECT:
6030 return isKnownNeverNaN(Op.getOperand(1), DemandedElts, SNaN, Depth + 1) &&
6031 isKnownNeverNaN(Op.getOperand(2), DemandedElts, SNaN, Depth + 1);
6032 case ISD::FP_EXTEND:
6033 case ISD::FP_ROUND: {
6034 if (SNaN)
6035 return true;
6036 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6037 }
6038 case ISD::SINT_TO_FP:
6039 case ISD::UINT_TO_FP:
6040 return true;
6041 case ISD::FSQRT: // Need is known positive
6042 case ISD::FLOG:
6043 case ISD::FLOG2:
6044 case ISD::FLOG10:
6045 case ISD::FPOWI:
6046 case ISD::FPOW: {
6047 if (SNaN)
6048 return true;
6049 // TODO: Refine on operand
6050 return false;
6051 }
6052 case ISD::FMINNUM:
6053 case ISD::FMAXNUM:
6054 case ISD::FMINIMUMNUM:
6055 case ISD::FMAXIMUMNUM: {
6056 // Only one needs to be known not-nan, since it will be returned if the
6057 // other ends up being one.
6058 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1) ||
6059 isKnownNeverNaN(Op.getOperand(1), DemandedElts, SNaN, Depth + 1);
6060 }
6061 case ISD::FMINNUM_IEEE:
6062 case ISD::FMAXNUM_IEEE: {
6063 if (SNaN)
6064 return true;
6065 // This can return a NaN if either operand is an sNaN, or if both operands
6066 // are NaN.
6067 return (isKnownNeverNaN(Op.getOperand(0), DemandedElts, false, Depth + 1) &&
6068 isKnownNeverSNaN(Op.getOperand(1), DemandedElts, Depth + 1)) ||
6069 (isKnownNeverNaN(Op.getOperand(1), DemandedElts, false, Depth + 1) &&
6070 isKnownNeverSNaN(Op.getOperand(0), DemandedElts, Depth + 1));
6071 }
6072 case ISD::FMINIMUM:
6073 case ISD::FMAXIMUM: {
6074 // TODO: Does this quiet or return the origina NaN as-is?
6075 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1) &&
6076 isKnownNeverNaN(Op.getOperand(1), DemandedElts, SNaN, Depth + 1);
6077 }
6079 SDValue Src = Op.getOperand(0);
6080 auto *Idx = dyn_cast<ConstantSDNode>(Op.getOperand(1));
6081 EVT SrcVT = Src.getValueType();
6082 if (SrcVT.isFixedLengthVector() && Idx &&
6083 Idx->getAPIntValue().ult(SrcVT.getVectorNumElements())) {
6084 APInt DemandedSrcElts = APInt::getOneBitSet(SrcVT.getVectorNumElements(),
6085 Idx->getZExtValue());
6086 return isKnownNeverNaN(Src, DemandedSrcElts, SNaN, Depth + 1);
6087 }
6088 return isKnownNeverNaN(Src, SNaN, Depth + 1);
6089 }
6091 SDValue Src = Op.getOperand(0);
6092 if (Src.getValueType().isFixedLengthVector()) {
6093 unsigned Idx = Op.getConstantOperandVal(1);
6094 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
6095 APInt DemandedSrcElts = DemandedElts.zext(NumSrcElts).shl(Idx);
6096 return isKnownNeverNaN(Src, DemandedSrcElts, SNaN, Depth + 1);
6097 }
6098 return isKnownNeverNaN(Src, SNaN, Depth + 1);
6099 }
6100 case ISD::INSERT_SUBVECTOR: {
6101 SDValue BaseVector = Op.getOperand(0);
6102 SDValue SubVector = Op.getOperand(1);
6103 EVT BaseVectorVT = BaseVector.getValueType();
6104 if (BaseVectorVT.isFixedLengthVector()) {
6105 unsigned Idx = Op.getConstantOperandVal(2);
6106 unsigned NumBaseElts = BaseVectorVT.getVectorNumElements();
6107 unsigned NumSubElts = SubVector.getValueType().getVectorNumElements();
6108
6109 // Clear/Extract the bits at the position where the subvector will be
6110 // inserted.
6111 APInt DemandedMask =
6112 APInt::getBitsSet(NumBaseElts, Idx, Idx + NumSubElts);
6113 APInt DemandedSrcElts = DemandedElts & ~DemandedMask;
6114 APInt DemandedSubElts = DemandedElts.extractBits(NumSubElts, Idx);
6115
6116 bool NeverNaN = true;
6117 if (!DemandedSrcElts.isZero())
6118 NeverNaN &=
6119 isKnownNeverNaN(BaseVector, DemandedSrcElts, SNaN, Depth + 1);
6120 if (NeverNaN && !DemandedSubElts.isZero())
6121 NeverNaN &=
6122 isKnownNeverNaN(SubVector, DemandedSubElts, SNaN, Depth + 1);
6123 return NeverNaN;
6124 }
6125 return isKnownNeverNaN(BaseVector, SNaN, Depth + 1) &&
6126 isKnownNeverNaN(SubVector, SNaN, Depth + 1);
6127 }
6128 case ISD::BUILD_VECTOR: {
6129 unsigned NumElts = Op.getNumOperands();
6130 for (unsigned I = 0; I != NumElts; ++I)
6131 if (DemandedElts[I] &&
6132 !isKnownNeverNaN(Op.getOperand(I), SNaN, Depth + 1))
6133 return false;
6134 return true;
6135 }
6136 case ISD::SPLAT_VECTOR:
6137 return isKnownNeverNaN(Op.getOperand(0), SNaN, Depth + 1);
6138 case ISD::AssertNoFPClass: {
6139 FPClassTest NoFPClass =
6140 static_cast<FPClassTest>(Op.getConstantOperandVal(1));
6141 if ((NoFPClass & fcNan) == fcNan)
6142 return true;
6143 if (SNaN && (NoFPClass & fcSNan) == fcSNan)
6144 return true;
6145 return isKnownNeverNaN(Op.getOperand(0), DemandedElts, SNaN, Depth + 1);
6146 }
6147 default:
6148 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::INTRINSIC_WO_CHAIN ||
6149 Opcode == ISD::INTRINSIC_W_CHAIN || Opcode == ISD::INTRINSIC_VOID) {
6150 return TLI->isKnownNeverNaNForTargetNode(Op, DemandedElts, *this, SNaN,
6151 Depth);
6152 }
6153
6154 return false;
6155 }
6156}
6157
6159 assert(Op.getValueType().isFloatingPoint() &&
6160 "Floating point type expected");
6161
6162 // If the value is a constant, we can obviously see if it is a zero or not.
6164 Op, [](ConstantFPSDNode *C) { return !C->isZero(); });
6165}
6166
6168 EVT VT = Op.getValueType();
6169
6170 // Since the number of lanes in a scalable vector is unknown at compile time,
6171 // we track one bit which is implicitly broadcast to all lanes. This means
6172 // that all lanes in a scalable vector are considered demanded.
6173 APInt DemandedElts = VT.isFixedLengthVector()
6175 : APInt(1, 1);
6176
6177 return isKnownNeverZero(Op, DemandedElts, Depth);
6178}
6179
6181 unsigned Depth) const {
6182 if (Depth >= MaxRecursionDepth)
6183 return false; // Limit search depth.
6184
6185 EVT OpVT = Op.getValueType();
6186 unsigned BitWidth = OpVT.getScalarSizeInBits();
6187
6188 assert(!Op.getValueType().isFloatingPoint() &&
6189 "Floating point types unsupported - use isKnownNeverZeroFloat");
6190
6191 // If the value is a constant, we can obviously see if it is a zero or not.
6192 auto IsNeverZero = [BitWidth](const ConstantSDNode *C) {
6193 APInt V = C->getAPIntValue().zextOrTrunc(BitWidth);
6194 return !V.isZero();
6195 };
6196
6197 if (ISD::matchUnaryPredicate(Op, IsNeverZero))
6198 return true;
6199
6200 // TODO: Recognize more cases here. Most of the cases are also incomplete to
6201 // some degree.
6202 switch (Op.getOpcode()) {
6203 default:
6204 break;
6205
6206 case ISD::BUILD_VECTOR:
6207 // Are all operands of a build vector constant non-zero?
6208 if (all_of(enumerate(Op->ops()), [&](auto P) {
6209 auto *C = dyn_cast<ConstantSDNode>(P.value());
6210 return !DemandedElts[P.index()] || (C && IsNeverZero(C));
6211 }))
6212 return true;
6213 break;
6214
6215 case ISD::SPLAT_VECTOR:
6216 // Is the operand of a splat vector a constant non-zero?
6217 if (auto *C = dyn_cast<ConstantSDNode>(Op->getOperand(0)))
6218 if (IsNeverZero(C))
6219 return true;
6220 break;
6221
6223 SDValue InVec = Op.getOperand(0);
6224 SDValue EltNo = Op.getOperand(1);
6225 EVT VecVT = InVec.getValueType();
6226
6227 // Skip scalable vectors or implicit extensions.
6228 if (VecVT.isScalableVector() ||
6229 OpVT.getScalarSizeInBits() != VecVT.getScalarSizeInBits())
6230 break;
6231
6232 // If we know the element index, just demand that vector element, else for
6233 // an unknown element index, ignore DemandedElts and demand them all.
6234 const unsigned NumSrcElts = VecVT.getVectorNumElements();
6235 APInt DemandedSrcElts = APInt::getAllOnes(NumSrcElts);
6236 auto *ConstEltNo = dyn_cast<ConstantSDNode>(EltNo);
6237 if (ConstEltNo && ConstEltNo->getAPIntValue().ult(NumSrcElts))
6238 DemandedSrcElts =
6239 APInt::getOneBitSet(NumSrcElts, ConstEltNo->getZExtValue());
6240
6241 return isKnownNeverZero(InVec, DemandedSrcElts, Depth + 1);
6242 }
6243
6244 case ISD::OR:
6245 return isKnownNeverZero(Op.getOperand(1), Depth + 1) ||
6246 isKnownNeverZero(Op.getOperand(0), Depth + 1);
6247
6248 case ISD::VSELECT:
6249 case ISD::SELECT:
6250 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) &&
6251 isKnownNeverZero(Op.getOperand(2), DemandedElts, Depth + 1);
6252
6253 case ISD::SHL: {
6254 if (Op->getFlags().hasNoSignedWrap() || Op->getFlags().hasNoUnsignedWrap())
6255 return isKnownNeverZero(Op.getOperand(0), Depth + 1);
6256 KnownBits ValKnown = computeKnownBits(Op.getOperand(0), Depth + 1);
6257 // 1 << X is never zero.
6258 if (ValKnown.One[0])
6259 return true;
6260 // If max shift cnt of known ones is non-zero, result is non-zero.
6261 APInt MaxCnt = computeKnownBits(Op.getOperand(1), Depth + 1).getMaxValue();
6262 if (MaxCnt.ult(ValKnown.getBitWidth()) &&
6263 !ValKnown.One.shl(MaxCnt).isZero())
6264 return true;
6265 break;
6266 }
6267 case ISD::UADDSAT:
6268 case ISD::UMAX:
6269 return isKnownNeverZero(Op.getOperand(1), Depth + 1) ||
6270 isKnownNeverZero(Op.getOperand(0), Depth + 1);
6271
6272 // For smin/smax: If either operand is known negative/positive
6273 // respectively we don't need the other to be known at all.
6274 case ISD::SMAX: {
6275 KnownBits Op1 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
6276 if (Op1.isStrictlyPositive())
6277 return true;
6278
6279 KnownBits Op0 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
6280 if (Op0.isStrictlyPositive())
6281 return true;
6282
6283 if (Op1.isNonZero() && Op0.isNonZero())
6284 return true;
6285
6286 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) &&
6287 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6288 }
6289 case ISD::SMIN: {
6290 KnownBits Op1 = computeKnownBits(Op.getOperand(1), DemandedElts, Depth + 1);
6291 if (Op1.isNegative())
6292 return true;
6293
6294 KnownBits Op0 = computeKnownBits(Op.getOperand(0), DemandedElts, Depth + 1);
6295 if (Op0.isNegative())
6296 return true;
6297
6298 if (Op1.isNonZero() && Op0.isNonZero())
6299 return true;
6300
6301 return isKnownNeverZero(Op.getOperand(1), DemandedElts, Depth + 1) &&
6302 isKnownNeverZero(Op.getOperand(0), DemandedElts, Depth + 1);
6303 }
6304 case ISD::UMIN:
6305 return isKnownNeverZero(Op.getOperand(1), Depth + 1) &&
6306 isKnownNeverZero(Op.getOperand(0), Depth + 1);
6307
6308 case ISD::ROTL:
6309 case ISD::ROTR:
6310 case ISD::BITREVERSE:
6311 case ISD::BSWAP:
6312 case ISD::CTPOP:
6313 case ISD::ABS:
6314 return isKnownNeverZero(Op.getOperand(0), Depth + 1);
6315
6316 case ISD::SRA:
6317 case ISD::SRL: {
6318 if (Op->getFlags().hasExact())
6319 return isKnownNeverZero(Op.getOperand(0), Depth + 1);
6320 KnownBits ValKnown = computeKnownBits(Op.getOperand(0), Depth + 1);
6321 if (ValKnown.isNegative())
6322 return true;
6323 // If max shift cnt of known ones is non-zero, result is non-zero.
6324 APInt MaxCnt = computeKnownBits(Op.getOperand(1), Depth + 1).getMaxValue();
6325 if (MaxCnt.ult(ValKnown.getBitWidth()) &&
6326 !ValKnown.One.lshr(MaxCnt).isZero())
6327 return true;
6328 break;
6329 }
6330 case ISD::UDIV:
6331 case ISD::SDIV:
6332 // div exact can only produce a zero if the dividend is zero.
6333 // TODO: For udiv this is also true if Op1 u<= Op0
6334 if (Op->getFlags().hasExact())
6335 return isKnownNeverZero(Op.getOperand(0), Depth + 1);
6336 break;
6337
6338 case ISD::ADD:
6339 if (Op->getFlags().hasNoUnsignedWrap())
6340 if (isKnownNeverZero(Op.getOperand(1), Depth + 1) ||
6341 isKnownNeverZero(Op.getOperand(0), Depth + 1))
6342 return true;
6343 // TODO: There are a lot more cases we can prove for add.
6344 break;
6345
6346 case ISD::SUB: {
6347 if (isNullConstant(Op.getOperand(0)))
6348 return isKnownNeverZero(Op.getOperand(1), Depth + 1);
6349
6350 std::optional<bool> ne =
6351 KnownBits::ne(computeKnownBits(Op.getOperand(0), Depth + 1),
6352 computeKnownBits(Op.getOperand(1), Depth + 1));
6353 return ne && *ne;
6354 }
6355
6356 case ISD::MUL:
6357 if (Op->getFlags().hasNoSignedWrap() || Op->getFlags().hasNoUnsignedWrap())
6358 if (isKnownNeverZero(Op.getOperand(1), Depth + 1) &&
6359 isKnownNeverZero(Op.getOperand(0), Depth + 1))
6360 return true;
6361 break;
6362
6363 case ISD::ZERO_EXTEND:
6364 case ISD::SIGN_EXTEND:
6365 return isKnownNeverZero(Op.getOperand(0), Depth + 1);
6366 case ISD::VSCALE: {
6368 const APInt &Multiplier = Op.getConstantOperandAPInt(0);
6369 ConstantRange CR =
6370 getVScaleRange(&F, Op.getScalarValueSizeInBits()).multiply(Multiplier);
6371 if (!CR.contains(APInt(CR.getBitWidth(), 0)))
6372 return true;
6373 break;
6374 }
6375 }
6376
6378}
6379
6381 if (ConstantFPSDNode *C1 = isConstOrConstSplatFP(Op, true))
6382 return !C1->isNegative();
6383
6384 switch (Op.getOpcode()) {
6385 case ISD::FABS:
6386 case ISD::FEXP:
6387 case ISD::FEXP2:
6388 case ISD::FEXP10:
6389 return true;
6390 default:
6391 return false;
6392 }
6393
6394 llvm_unreachable("covered opcode switch");
6395}
6396
6398 assert(Use.getValueType().isFloatingPoint());
6399 const SDNode *User = Use.getUser();
6400 if (User->getFlags().hasNoSignedZeros())
6401 return true;
6402
6403 unsigned OperandNo = Use.getOperandNo();
6404 // Check if this use is insensitive to the sign of zero
6405 switch (User->getOpcode()) {
6406 case ISD::SETCC:
6407 // Comparisons: IEEE-754 specifies +0.0 == -0.0.
6408 case ISD::FABS:
6409 // fabs always produces +0.0.
6410 return true;
6411 case ISD::FCOPYSIGN:
6412 // copysign overwrites the sign bit of the first operand.
6413 return OperandNo == 0;
6414 case ISD::FADD:
6415 case ISD::FSUB: {
6416 // Arithmetic with non-zero constants fixes the uncertainty around the
6417 // sign bit.
6418 SDValue Other = User->getOperand(1 - OperandNo);
6420 }
6421 case ISD::FP_TO_SINT:
6422 case ISD::FP_TO_UINT:
6423 // fp-to-int conversions normalize signed zeros.
6424 return true;
6425 default:
6426 return false;
6427 }
6428}
6429
6431 if (Op->getFlags().hasNoSignedZeros())
6432 return true;
6433 // FIXME: Limit the amount of checked uses to not introduce a compile-time
6434 // regression. Ideally, this should be implemented as a demanded-bits
6435 // optimization that stems from the users.
6436 if (Op->use_size() > 2)
6437 return false;
6438 return all_of(Op->uses(),
6439 [&](const SDUse &Use) { return canIgnoreSignBitOfZero(Use); });
6440}
6441
6443 // Check the obvious case.
6444 if (A == B) return true;
6445
6446 // For negative and positive zero.
6449 if (CA->isZero() && CB->isZero()) return true;
6450
6451 // Otherwise they may not be equal.
6452 return false;
6453}
6454
6455// Only bits set in Mask must be negated, other bits may be arbitrary.
6457 if (isBitwiseNot(V, AllowUndefs))
6458 return V.getOperand(0);
6459
6460 // Handle any_extend (not (truncate X)) pattern, where Mask only sets
6461 // bits in the non-extended part.
6462 ConstantSDNode *MaskC = isConstOrConstSplat(Mask);
6463 if (!MaskC || V.getOpcode() != ISD::ANY_EXTEND)
6464 return SDValue();
6465 SDValue ExtArg = V.getOperand(0);
6466 if (ExtArg.getScalarValueSizeInBits() >=
6467 MaskC->getAPIntValue().getActiveBits() &&
6468 isBitwiseNot(ExtArg, AllowUndefs) &&
6469 ExtArg.getOperand(0).getOpcode() == ISD::TRUNCATE &&
6470 ExtArg.getOperand(0).getOperand(0).getValueType() == V.getValueType())
6471 return ExtArg.getOperand(0).getOperand(0);
6472 return SDValue();
6473}
6474
6476 // Match masked merge pattern (X & ~M) op (Y & M)
6477 // Including degenerate case (X & ~M) op M
6478 auto MatchNoCommonBitsPattern = [&](SDValue Not, SDValue Mask,
6479 SDValue Other) {
6480 if (SDValue NotOperand =
6481 getBitwiseNotOperand(Not, Mask, /* AllowUndefs */ true)) {
6482 if (NotOperand->getOpcode() == ISD::ZERO_EXTEND ||
6483 NotOperand->getOpcode() == ISD::TRUNCATE)
6484 NotOperand = NotOperand->getOperand(0);
6485
6486 if (Other == NotOperand)
6487 return true;
6488 if (Other->getOpcode() == ISD::AND)
6489 return NotOperand == Other->getOperand(0) ||
6490 NotOperand == Other->getOperand(1);
6491 }
6492 return false;
6493 };
6494
6495 if (A->getOpcode() == ISD::ZERO_EXTEND || A->getOpcode() == ISD::TRUNCATE)
6496 A = A->getOperand(0);
6497
6498 if (B->getOpcode() == ISD::ZERO_EXTEND || B->getOpcode() == ISD::TRUNCATE)
6499 B = B->getOperand(0);
6500
6501 if (A->getOpcode() == ISD::AND)
6502 return MatchNoCommonBitsPattern(A->getOperand(0), A->getOperand(1), B) ||
6503 MatchNoCommonBitsPattern(A->getOperand(1), A->getOperand(0), B);
6504 return false;
6505}
6506
6507// FIXME: unify with llvm::haveNoCommonBitsSet.
6509 assert(A.getValueType() == B.getValueType() &&
6510 "Values must have the same type");
6513 return true;
6516}
6517
6518static SDValue FoldSTEP_VECTOR(const SDLoc &DL, EVT VT, SDValue Step,
6519 SelectionDAG &DAG) {
6520 if (cast<ConstantSDNode>(Step)->isZero())
6521 return DAG.getConstant(0, DL, VT);
6522
6523 return SDValue();
6524}
6525
6528 SelectionDAG &DAG) {
6529 int NumOps = Ops.size();
6530 assert(NumOps != 0 && "Can't build an empty vector!");
6531 assert(!VT.isScalableVector() &&
6532 "BUILD_VECTOR cannot be used with scalable types");
6533 assert(VT.getVectorNumElements() == (unsigned)NumOps &&
6534 "Incorrect element count in BUILD_VECTOR!");
6535
6536 // BUILD_VECTOR of UNDEFs is UNDEF.
6537 if (llvm::all_of(Ops, [](SDValue Op) { return Op.isUndef(); }))
6538 return DAG.getUNDEF(VT);
6539
6540 // BUILD_VECTOR of seq extract/insert from the same vector + type is Identity.
6541 SDValue IdentitySrc;
6542 bool IsIdentity = true;
6543 for (int i = 0; i != NumOps; ++i) {
6545 Ops[i].getOperand(0).getValueType() != VT ||
6546 (IdentitySrc && Ops[i].getOperand(0) != IdentitySrc) ||
6547 !isa<ConstantSDNode>(Ops[i].getOperand(1)) ||
6548 Ops[i].getConstantOperandAPInt(1) != i) {
6549 IsIdentity = false;
6550 break;
6551 }
6552 IdentitySrc = Ops[i].getOperand(0);
6553 }
6554 if (IsIdentity)
6555 return IdentitySrc;
6556
6557 return SDValue();
6558}
6559
6560/// Try to simplify vector concatenation to an input value, undef, or build
6561/// vector.
6564 SelectionDAG &DAG) {
6565 assert(!Ops.empty() && "Can't concatenate an empty list of vectors!");
6567 [Ops](SDValue Op) {
6568 return Ops[0].getValueType() == Op.getValueType();
6569 }) &&
6570 "Concatenation of vectors with inconsistent value types!");
6571 assert((Ops[0].getValueType().getVectorElementCount() * Ops.size()) ==
6572 VT.getVectorElementCount() &&
6573 "Incorrect element count in vector concatenation!");
6574
6575 if (Ops.size() == 1)
6576 return Ops[0];
6577
6578 // Concat of UNDEFs is UNDEF.
6579 if (llvm::all_of(Ops, [](SDValue Op) { return Op.isUndef(); }))
6580 return DAG.getUNDEF(VT);
6581
6582 // Scan the operands and look for extract operations from a single source
6583 // that correspond to insertion at the same location via this concatenation:
6584 // concat (extract X, 0*subvec_elts), (extract X, 1*subvec_elts), ...
6585 SDValue IdentitySrc;
6586 bool IsIdentity = true;
6587 for (unsigned i = 0, e = Ops.size(); i != e; ++i) {
6588 SDValue Op = Ops[i];
6589 unsigned IdentityIndex = i * Op.getValueType().getVectorMinNumElements();
6590 if (Op.getOpcode() != ISD::EXTRACT_SUBVECTOR ||
6591 Op.getOperand(0).getValueType() != VT ||
6592 (IdentitySrc && Op.getOperand(0) != IdentitySrc) ||
6593 Op.getConstantOperandVal(1) != IdentityIndex) {
6594 IsIdentity = false;
6595 break;
6596 }
6597 assert((!IdentitySrc || IdentitySrc == Op.getOperand(0)) &&
6598 "Unexpected identity source vector for concat of extracts");
6599 IdentitySrc = Op.getOperand(0);
6600 }
6601 if (IsIdentity) {
6602 assert(IdentitySrc && "Failed to set source vector of extracts");
6603 return IdentitySrc;
6604 }
6605
6606 // The code below this point is only designed to work for fixed width
6607 // vectors, so we bail out for now.
6608 if (VT.isScalableVector())
6609 return SDValue();
6610
6611 // A CONCAT_VECTOR of scalar sources, such as UNDEF, BUILD_VECTOR and
6612 // single-element INSERT_VECTOR_ELT operands can be simplified to one big
6613 // BUILD_VECTOR.
6614 // FIXME: Add support for SCALAR_TO_VECTOR as well.
6615 EVT SVT = VT.getScalarType();
6617 for (SDValue Op : Ops) {
6618 EVT OpVT = Op.getValueType();
6619 if (Op.isUndef())
6620 Elts.append(OpVT.getVectorNumElements(), DAG.getUNDEF(SVT));
6621 else if (Op.getOpcode() == ISD::BUILD_VECTOR)
6622 Elts.append(Op->op_begin(), Op->op_end());
6623 else if (Op.getOpcode() == ISD::INSERT_VECTOR_ELT &&
6624 OpVT.getVectorNumElements() == 1 &&
6625 isNullConstant(Op.getOperand(2)))
6626 Elts.push_back(Op.getOperand(1));
6627 else
6628 return SDValue();
6629 }
6630
6631 // BUILD_VECTOR requires all inputs to be of the same type, find the
6632 // maximum type and extend them all.
6633 for (SDValue Op : Elts)
6634 SVT = (SVT.bitsLT(Op.getValueType()) ? Op.getValueType() : SVT);
6635
6636 if (SVT.bitsGT(VT.getScalarType())) {
6637 for (SDValue &Op : Elts) {
6638 if (Op.isUndef())
6639 Op = DAG.getUNDEF(SVT);
6640 else
6641 Op = DAG.getTargetLoweringInfo().isZExtFree(Op.getValueType(), SVT)
6642 ? DAG.getZExtOrTrunc(Op, DL, SVT)
6643 : DAG.getSExtOrTrunc(Op, DL, SVT);
6644 }
6645 }
6646
6647 SDValue V = DAG.getBuildVector(VT, DL, Elts);
6648 NewSDValueDbgMsg(V, "New node fold concat vectors: ", &DAG);
6649 return V;
6650}
6651
6652/// Gets or creates the specified node.
6653SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT) {
6654 SDVTList VTs = getVTList(VT);
6656 AddNodeIDNode(ID, Opcode, VTs, {});
6657 void *IP = nullptr;
6658 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP))
6659 return SDValue(E, 0);
6660
6661 auto *N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
6662 CSEMap.InsertNode(N, IP);
6663
6664 InsertNode(N);
6665 SDValue V = SDValue(N, 0);
6666 NewSDValueDbgMsg(V, "Creating new node: ", this);
6667 return V;
6668}
6669
6670SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
6671 SDValue N1) {
6672 SDNodeFlags Flags;
6673 if (Inserter)
6674 Flags = Inserter->getFlags();
6675 return getNode(Opcode, DL, VT, N1, Flags);
6676}
6677
6678SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
6679 SDValue N1, const SDNodeFlags Flags) {
6680 assert(N1.getOpcode() != ISD::DELETED_NODE && "Operand is DELETED_NODE!");
6681
6682 // Constant fold unary operations with a vector integer or float operand.
6683 switch (Opcode) {
6684 default:
6685 // FIXME: Entirely reasonable to perform folding of other unary
6686 // operations here as the need arises.
6687 break;
6688 case ISD::FNEG:
6689 case ISD::FABS:
6690 case ISD::FCEIL:
6691 case ISD::FTRUNC:
6692 case ISD::FFLOOR:
6693 case ISD::FP_EXTEND:
6694 case ISD::FP_TO_SINT:
6695 case ISD::FP_TO_UINT:
6696 case ISD::FP_TO_FP16:
6697 case ISD::FP_TO_BF16:
6698 case ISD::TRUNCATE:
6699 case ISD::ANY_EXTEND:
6700 case ISD::ZERO_EXTEND:
6701 case ISD::SIGN_EXTEND:
6702 case ISD::UINT_TO_FP:
6703 case ISD::SINT_TO_FP:
6704 case ISD::FP16_TO_FP:
6705 case ISD::BF16_TO_FP:
6706 case ISD::BITCAST:
6707 case ISD::ABS:
6708 case ISD::BITREVERSE:
6709 case ISD::BSWAP:
6710 case ISD::CTLZ:
6712 case ISD::CTTZ:
6714 case ISD::CTPOP:
6715 case ISD::CTLS:
6716 case ISD::STEP_VECTOR: {
6717 SDValue Ops = {N1};
6718 if (SDValue Fold = FoldConstantArithmetic(Opcode, DL, VT, Ops))
6719 return Fold;
6720 }
6721 }
6722
6723 unsigned OpOpcode = N1.getNode()->getOpcode();
6724 switch (Opcode) {
6725 case ISD::STEP_VECTOR:
6726 assert(VT.isScalableVector() &&
6727 "STEP_VECTOR can only be used with scalable types");
6728 assert(OpOpcode == ISD::TargetConstant &&
6729 VT.getVectorElementType() == N1.getValueType() &&
6730 "Unexpected step operand");
6731 break;
6732 case ISD::FREEZE:
6733 assert(VT == N1.getValueType() && "Unexpected VT!");
6734 if (isGuaranteedNotToBeUndefOrPoison(N1, /*PoisonOnly=*/false))
6735 return N1;
6736 break;
6737 case ISD::TokenFactor:
6738 case ISD::MERGE_VALUES:
6740 return N1; // Factor, merge or concat of one node? No need.
6741 case ISD::BUILD_VECTOR: {
6742 // Attempt to simplify BUILD_VECTOR.
6743 SDValue Ops[] = {N1};
6744 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
6745 return V;
6746 break;
6747 }
6748 case ISD::FP_ROUND: llvm_unreachable("Invalid method to make FP_ROUND node");
6749 case ISD::FP_EXTEND:
6751 "Invalid FP cast!");
6752 if (N1.getValueType() == VT) return N1; // noop conversion.
6753 assert((!VT.isVector() || VT.getVectorElementCount() ==
6755 "Vector element count mismatch!");
6756 assert(N1.getValueType().bitsLT(VT) && "Invalid fpext node, dst < src!");
6757 if (N1.isUndef())
6758 return getUNDEF(VT);
6759 break;
6760 case ISD::FP_TO_SINT:
6761 case ISD::FP_TO_UINT:
6762 if (N1.isUndef())
6763 return getUNDEF(VT);
6764 break;
6765 case ISD::SINT_TO_FP:
6766 case ISD::UINT_TO_FP:
6767 // [us]itofp(undef) = 0, because the result value is bounded.
6768 if (N1.isUndef())
6769 return getConstantFP(0.0, DL, VT);
6770 break;
6771 case ISD::SIGN_EXTEND:
6772 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6773 "Invalid SIGN_EXTEND!");
6774 assert(VT.isVector() == N1.getValueType().isVector() &&
6775 "SIGN_EXTEND result type type should be vector iff the operand "
6776 "type is vector!");
6777 if (N1.getValueType() == VT) return N1; // noop extension
6778 assert((!VT.isVector() || VT.getVectorElementCount() ==
6780 "Vector element count mismatch!");
6781 assert(N1.getValueType().bitsLT(VT) && "Invalid sext node, dst < src!");
6782 if (OpOpcode == ISD::SIGN_EXTEND || OpOpcode == ISD::ZERO_EXTEND) {
6783 SDNodeFlags Flags;
6784 if (OpOpcode == ISD::ZERO_EXTEND)
6785 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6786 SDValue NewVal = getNode(OpOpcode, DL, VT, N1.getOperand(0), Flags);
6787 transferDbgValues(N1, NewVal);
6788 return NewVal;
6789 }
6790
6791 if (OpOpcode == ISD::POISON)
6792 return getPOISON(VT);
6793
6794 if (N1.isUndef())
6795 // sext(undef) = 0, because the top bits will all be the same.
6796 return getConstant(0, DL, VT);
6797
6798 // Skip unnecessary sext_inreg pattern:
6799 // (sext (trunc x)) -> x iff the upper bits are all signbits.
6800 if (OpOpcode == ISD::TRUNCATE) {
6801 SDValue OpOp = N1.getOperand(0);
6802 if (OpOp.getValueType() == VT) {
6803 unsigned NumSignExtBits =
6805 if (ComputeNumSignBits(OpOp) > NumSignExtBits) {
6806 transferDbgValues(N1, OpOp);
6807 return OpOp;
6808 }
6809 }
6810 }
6811 break;
6812 case ISD::ZERO_EXTEND:
6813 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6814 "Invalid ZERO_EXTEND!");
6815 assert(VT.isVector() == N1.getValueType().isVector() &&
6816 "ZERO_EXTEND result type type should be vector iff the operand "
6817 "type is vector!");
6818 if (N1.getValueType() == VT) return N1; // noop extension
6819 assert((!VT.isVector() || VT.getVectorElementCount() ==
6821 "Vector element count mismatch!");
6822 assert(N1.getValueType().bitsLT(VT) && "Invalid zext node, dst < src!");
6823 if (OpOpcode == ISD::ZERO_EXTEND) { // (zext (zext x)) -> (zext x)
6824 SDNodeFlags Flags;
6825 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6826 SDValue NewVal =
6827 getNode(ISD::ZERO_EXTEND, DL, VT, N1.getOperand(0), Flags);
6828 transferDbgValues(N1, NewVal);
6829 return NewVal;
6830 }
6831
6832 if (OpOpcode == ISD::POISON)
6833 return getPOISON(VT);
6834
6835 if (N1.isUndef())
6836 // zext(undef) = 0, because the top bits will be zero.
6837 return getConstant(0, DL, VT);
6838
6839 // Skip unnecessary zext_inreg pattern:
6840 // (zext (trunc x)) -> x iff the upper bits are known zero.
6841 // TODO: Remove (zext (trunc (and x, c))) exception which some targets
6842 // use to recognise zext_inreg patterns.
6843 if (OpOpcode == ISD::TRUNCATE) {
6844 SDValue OpOp = N1.getOperand(0);
6845 if (OpOp.getValueType() == VT) {
6846 if (OpOp.getOpcode() != ISD::AND) {
6849 if (MaskedValueIsZero(OpOp, HiBits)) {
6850 transferDbgValues(N1, OpOp);
6851 return OpOp;
6852 }
6853 }
6854 }
6855 }
6856 break;
6857 case ISD::ANY_EXTEND:
6858 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6859 "Invalid ANY_EXTEND!");
6860 assert(VT.isVector() == N1.getValueType().isVector() &&
6861 "ANY_EXTEND result type type should be vector iff the operand "
6862 "type is vector!");
6863 if (N1.getValueType() == VT) return N1; // noop extension
6864 assert((!VT.isVector() || VT.getVectorElementCount() ==
6866 "Vector element count mismatch!");
6867 assert(N1.getValueType().bitsLT(VT) && "Invalid anyext node, dst < src!");
6868
6869 if (OpOpcode == ISD::ZERO_EXTEND || OpOpcode == ISD::SIGN_EXTEND ||
6870 OpOpcode == ISD::ANY_EXTEND) {
6871 SDNodeFlags Flags;
6872 if (OpOpcode == ISD::ZERO_EXTEND)
6873 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6874 // (ext (zext x)) -> (zext x) and (ext (sext x)) -> (sext x)
6875 return getNode(OpOpcode, DL, VT, N1.getOperand(0), Flags);
6876 }
6877 if (N1.isUndef())
6878 return getUNDEF(VT);
6879
6880 // (ext (trunc x)) -> x
6881 if (OpOpcode == ISD::TRUNCATE) {
6882 SDValue OpOp = N1.getOperand(0);
6883 if (OpOp.getValueType() == VT) {
6884 transferDbgValues(N1, OpOp);
6885 return OpOp;
6886 }
6887 }
6888 break;
6889 case ISD::TRUNCATE:
6890 assert(VT.isInteger() && N1.getValueType().isInteger() &&
6891 "Invalid TRUNCATE!");
6892 assert(VT.isVector() == N1.getValueType().isVector() &&
6893 "TRUNCATE result type type should be vector iff the operand "
6894 "type is vector!");
6895 if (N1.getValueType() == VT) return N1; // noop truncate
6896 assert((!VT.isVector() || VT.getVectorElementCount() ==
6898 "Vector element count mismatch!");
6899 assert(N1.getValueType().bitsGT(VT) && "Invalid truncate node, src < dst!");
6900 if (OpOpcode == ISD::TRUNCATE)
6901 return getNode(ISD::TRUNCATE, DL, VT, N1.getOperand(0));
6902 if (OpOpcode == ISD::ZERO_EXTEND || OpOpcode == ISD::SIGN_EXTEND ||
6903 OpOpcode == ISD::ANY_EXTEND) {
6904 // If the source is smaller than the dest, we still need an extend.
6906 VT.getScalarType())) {
6907 SDNodeFlags Flags;
6908 if (OpOpcode == ISD::ZERO_EXTEND)
6909 Flags.setNonNeg(N1->getFlags().hasNonNeg());
6910 return getNode(OpOpcode, DL, VT, N1.getOperand(0), Flags);
6911 }
6912 if (N1.getOperand(0).getValueType().bitsGT(VT))
6913 return getNode(ISD::TRUNCATE, DL, VT, N1.getOperand(0));
6914 return N1.getOperand(0);
6915 }
6916 if (N1.isUndef())
6917 return getUNDEF(VT);
6918 if (OpOpcode == ISD::VSCALE && !NewNodesMustHaveLegalTypes)
6919 return getVScale(DL, VT,
6921 break;
6925 assert(VT.isVector() && "This DAG node is restricted to vector types.");
6926 assert(N1.getValueType().bitsLE(VT) &&
6927 "The input must be the same size or smaller than the result.");
6930 "The destination vector type must have fewer lanes than the input.");
6931 break;
6932 case ISD::ABS:
6933 assert(VT.isInteger() && VT == N1.getValueType() && "Invalid ABS!");
6934 if (N1.isUndef())
6935 return getConstant(0, DL, VT);
6936 break;
6937 case ISD::BSWAP:
6938 assert(VT.isInteger() && VT == N1.getValueType() && "Invalid BSWAP!");
6939 assert((VT.getScalarSizeInBits() % 16 == 0) &&
6940 "BSWAP types must be a multiple of 16 bits!");
6941 if (N1.isUndef())
6942 return getUNDEF(VT);
6943 // bswap(bswap(X)) -> X.
6944 if (OpOpcode == ISD::BSWAP)
6945 return N1.getOperand(0);
6946 break;
6947 case ISD::BITREVERSE:
6948 assert(VT.isInteger() && VT == N1.getValueType() && "Invalid BITREVERSE!");
6949 if (N1.isUndef())
6950 return getUNDEF(VT);
6951 break;
6952 case ISD::BITCAST:
6954 "Cannot BITCAST between types of different sizes!");
6955 if (VT == N1.getValueType()) return N1; // noop conversion.
6956 if (OpOpcode == ISD::BITCAST) // bitconv(bitconv(x)) -> bitconv(x)
6957 return getNode(ISD::BITCAST, DL, VT, N1.getOperand(0));
6958 if (N1.isUndef())
6959 return getUNDEF(VT);
6960 break;
6962 assert(VT.isVector() && !N1.getValueType().isVector() &&
6963 (VT.getVectorElementType() == N1.getValueType() ||
6965 N1.getValueType().isInteger() &&
6967 "Illegal SCALAR_TO_VECTOR node!");
6968 if (N1.isUndef())
6969 return getUNDEF(VT);
6970 // scalar_to_vector(extract_vector_elt V, 0) -> V, top bits are undefined.
6971 if (OpOpcode == ISD::EXTRACT_VECTOR_ELT &&
6973 N1.getConstantOperandVal(1) == 0 &&
6974 N1.getOperand(0).getValueType() == VT)
6975 return N1.getOperand(0);
6976 break;
6977 case ISD::FNEG:
6978 // Negation of an unknown bag of bits is still completely undefined.
6979 if (N1.isUndef())
6980 return getUNDEF(VT);
6981
6982 if (OpOpcode == ISD::FNEG) // --X -> X
6983 return N1.getOperand(0);
6984 break;
6985 case ISD::FABS:
6986 if (OpOpcode == ISD::FNEG) // abs(-X) -> abs(X)
6987 return getNode(ISD::FABS, DL, VT, N1.getOperand(0));
6988 break;
6989 case ISD::VSCALE:
6990 assert(VT == N1.getValueType() && "Unexpected VT!");
6991 break;
6992 case ISD::CTPOP:
6993 if (N1.getValueType().getScalarType() == MVT::i1)
6994 return N1;
6995 break;
6996 case ISD::CTLZ:
6997 case ISD::CTTZ:
6998 if (N1.getValueType().getScalarType() == MVT::i1)
6999 return getNOT(DL, N1, N1.getValueType());
7000 break;
7001 case ISD::CTLS:
7002 if (N1.getValueType().getScalarType() == MVT::i1)
7003 return getConstant(0, DL, VT);
7004 break;
7005 case ISD::VECREDUCE_ADD:
7006 if (N1.getValueType().getScalarType() == MVT::i1)
7007 return getNode(ISD::VECREDUCE_XOR, DL, VT, N1);
7008 break;
7011 if (N1.getValueType().getScalarType() == MVT::i1)
7012 return getNode(ISD::VECREDUCE_OR, DL, VT, N1);
7013 break;
7016 if (N1.getValueType().getScalarType() == MVT::i1)
7017 return getNode(ISD::VECREDUCE_AND, DL, VT, N1);
7018 break;
7019 case ISD::SPLAT_VECTOR:
7020 assert(VT.isVector() && "Wrong return type!");
7021 // FIXME: Hexagon uses i32 scalar for a floating point zero vector so allow
7022 // that for now.
7024 (VT.isFloatingPoint() && N1.getValueType() == MVT::i32) ||
7026 N1.getValueType().isInteger() &&
7028 "Wrong operand type!");
7029 break;
7030 }
7031
7032 SDNode *N;
7033 SDVTList VTs = getVTList(VT);
7034 SDValue Ops[] = {N1};
7035 if (VT != MVT::Glue) { // Don't CSE glue producing nodes
7037 AddNodeIDNode(ID, Opcode, VTs, Ops);
7038 void *IP = nullptr;
7039 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
7040 E->intersectFlagsWith(Flags);
7041 return SDValue(E, 0);
7042 }
7043
7044 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
7045 N->setFlags(Flags);
7046 createOperands(N, Ops);
7047 CSEMap.InsertNode(N, IP);
7048 } else {
7049 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
7050 createOperands(N, Ops);
7051 }
7052
7053 InsertNode(N);
7054 SDValue V = SDValue(N, 0);
7055 NewSDValueDbgMsg(V, "Creating new node: ", this);
7056 return V;
7057}
7058
7059static std::optional<APInt> FoldValue(unsigned Opcode, const APInt &C1,
7060 const APInt &C2) {
7061 switch (Opcode) {
7062 case ISD::ADD: return C1 + C2;
7063 case ISD::SUB: return C1 - C2;
7064 case ISD::MUL: return C1 * C2;
7065 case ISD::AND: return C1 & C2;
7066 case ISD::OR: return C1 | C2;
7067 case ISD::XOR: return C1 ^ C2;
7068 case ISD::SHL: return C1 << C2;
7069 case ISD::SRL: return C1.lshr(C2);
7070 case ISD::SRA: return C1.ashr(C2);
7071 case ISD::ROTL: return C1.rotl(C2);
7072 case ISD::ROTR: return C1.rotr(C2);
7073 case ISD::SMIN: return C1.sle(C2) ? C1 : C2;
7074 case ISD::SMAX: return C1.sge(C2) ? C1 : C2;
7075 case ISD::UMIN: return C1.ule(C2) ? C1 : C2;
7076 case ISD::UMAX: return C1.uge(C2) ? C1 : C2;
7077 case ISD::SADDSAT: return C1.sadd_sat(C2);
7078 case ISD::UADDSAT: return C1.uadd_sat(C2);
7079 case ISD::SSUBSAT: return C1.ssub_sat(C2);
7080 case ISD::USUBSAT: return C1.usub_sat(C2);
7081 case ISD::SSHLSAT: return C1.sshl_sat(C2);
7082 case ISD::USHLSAT: return C1.ushl_sat(C2);
7083 case ISD::UDIV:
7084 if (!C2.getBoolValue())
7085 break;
7086 return C1.udiv(C2);
7087 case ISD::UREM:
7088 if (!C2.getBoolValue())
7089 break;
7090 return C1.urem(C2);
7091 case ISD::SDIV:
7092 if (!C2.getBoolValue())
7093 break;
7094 return C1.sdiv(C2);
7095 case ISD::SREM:
7096 if (!C2.getBoolValue())
7097 break;
7098 return C1.srem(C2);
7099 case ISD::AVGFLOORS:
7100 return APIntOps::avgFloorS(C1, C2);
7101 case ISD::AVGFLOORU:
7102 return APIntOps::avgFloorU(C1, C2);
7103 case ISD::AVGCEILS:
7104 return APIntOps::avgCeilS(C1, C2);
7105 case ISD::AVGCEILU:
7106 return APIntOps::avgCeilU(C1, C2);
7107 case ISD::ABDS:
7108 return APIntOps::abds(C1, C2);
7109 case ISD::ABDU:
7110 return APIntOps::abdu(C1, C2);
7111 case ISD::MULHS:
7112 return APIntOps::mulhs(C1, C2);
7113 case ISD::MULHU:
7114 return APIntOps::mulhu(C1, C2);
7115 case ISD::CLMUL:
7116 return APIntOps::clmul(C1, C2);
7117 case ISD::CLMULR:
7118 return APIntOps::clmulr(C1, C2);
7119 case ISD::CLMULH:
7120 return APIntOps::clmulh(C1, C2);
7121 }
7122 return std::nullopt;
7123}
7124// Handle constant folding with UNDEF.
7125// TODO: Handle more cases.
7126static std::optional<APInt> FoldValueWithUndef(unsigned Opcode, const APInt &C1,
7127 bool IsUndef1, const APInt &C2,
7128 bool IsUndef2) {
7129 if (!(IsUndef1 || IsUndef2))
7130 return FoldValue(Opcode, C1, C2);
7131
7132 // Fold and(x, undef) -> 0
7133 // Fold mul(x, undef) -> 0
7134 if (Opcode == ISD::AND || Opcode == ISD::MUL)
7135 return APInt::getZero(C1.getBitWidth());
7136
7137 return std::nullopt;
7138}
7139
7141 const GlobalAddressSDNode *GA,
7142 const SDNode *N2) {
7143 if (GA->getOpcode() != ISD::GlobalAddress)
7144 return SDValue();
7145 if (!TLI->isOffsetFoldingLegal(GA))
7146 return SDValue();
7147 auto *C2 = dyn_cast<ConstantSDNode>(N2);
7148 if (!C2)
7149 return SDValue();
7150 int64_t Offset = C2->getSExtValue();
7151 switch (Opcode) {
7152 case ISD::ADD:
7153 case ISD::PTRADD:
7154 break;
7155 case ISD::SUB: Offset = -uint64_t(Offset); break;
7156 default: return SDValue();
7157 }
7158 return getGlobalAddress(GA->getGlobal(), SDLoc(C2), VT,
7159 GA->getOffset() + uint64_t(Offset));
7160}
7161
7163 switch (Opcode) {
7164 case ISD::SDIV:
7165 case ISD::UDIV:
7166 case ISD::SREM:
7167 case ISD::UREM: {
7168 // If a divisor is zero/undef or any element of a divisor vector is
7169 // zero/undef, the whole op is undef.
7170 assert(Ops.size() == 2 && "Div/rem should have 2 operands");
7171 SDValue Divisor = Ops[1];
7172 if (Divisor.isUndef() || isNullConstant(Divisor))
7173 return true;
7174
7175 return ISD::isBuildVectorOfConstantSDNodes(Divisor.getNode()) &&
7176 llvm::any_of(Divisor->op_values(),
7177 [](SDValue V) { return V.isUndef() ||
7178 isNullConstant(V); });
7179 // TODO: Handle signed overflow.
7180 }
7181 // TODO: Handle oversized shifts.
7182 default:
7183 return false;
7184 }
7185}
7186
7189 SDNodeFlags Flags) {
7190 // If the opcode is a target-specific ISD node, there's nothing we can
7191 // do here and the operand rules may not line up with the below, so
7192 // bail early.
7193 // We can't create a scalar CONCAT_VECTORS so skip it. It will break
7194 // for concats involving SPLAT_VECTOR. Concats of BUILD_VECTORS are handled by
7195 // foldCONCAT_VECTORS in getNode before this is called.
7196 if (Opcode >= ISD::BUILTIN_OP_END || Opcode == ISD::CONCAT_VECTORS)
7197 return SDValue();
7198
7199 unsigned NumOps = Ops.size();
7200 if (NumOps == 0)
7201 return SDValue();
7202
7203 if (isUndef(Opcode, Ops))
7204 return getUNDEF(VT);
7205
7206 // Handle unary special cases.
7207 if (NumOps == 1) {
7208 SDValue N1 = Ops[0];
7209
7210 // Constant fold unary operations with an integer constant operand. Even
7211 // opaque constant will be folded, because the folding of unary operations
7212 // doesn't create new constants with different values. Nevertheless, the
7213 // opaque flag is preserved during folding to prevent future folding with
7214 // other constants.
7215 if (auto *C = dyn_cast<ConstantSDNode>(N1)) {
7216 const APInt &Val = C->getAPIntValue();
7217 switch (Opcode) {
7218 case ISD::SIGN_EXTEND:
7219 return getConstant(Val.sextOrTrunc(VT.getSizeInBits()), DL, VT,
7220 C->isTargetOpcode(), C->isOpaque());
7221 case ISD::TRUNCATE:
7222 if (C->isOpaque())
7223 break;
7224 [[fallthrough]];
7225 case ISD::ZERO_EXTEND:
7226 return getConstant(Val.zextOrTrunc(VT.getSizeInBits()), DL, VT,
7227 C->isTargetOpcode(), C->isOpaque());
7228 case ISD::ANY_EXTEND:
7229 // Some targets like RISCV prefer to sign extend some types.
7230 if (TLI->isSExtCheaperThanZExt(N1.getValueType(), VT))
7231 return getConstant(Val.sextOrTrunc(VT.getSizeInBits()), DL, VT,
7232 C->isTargetOpcode(), C->isOpaque());
7233 return getConstant(Val.zextOrTrunc(VT.getSizeInBits()), DL, VT,
7234 C->isTargetOpcode(), C->isOpaque());
7235 case ISD::ABS:
7236 return getConstant(Val.abs(), DL, VT, C->isTargetOpcode(),
7237 C->isOpaque());
7238 case ISD::BITREVERSE:
7239 return getConstant(Val.reverseBits(), DL, VT, C->isTargetOpcode(),
7240 C->isOpaque());
7241 case ISD::BSWAP:
7242 return getConstant(Val.byteSwap(), DL, VT, C->isTargetOpcode(),
7243 C->isOpaque());
7244 case ISD::CTPOP:
7245 return getConstant(Val.popcount(), DL, VT, C->isTargetOpcode(),
7246 C->isOpaque());
7247 case ISD::CTLZ:
7249 return getConstant(Val.countl_zero(), DL, VT, C->isTargetOpcode(),
7250 C->isOpaque());
7251 case ISD::CTTZ:
7253 return getConstant(Val.countr_zero(), DL, VT, C->isTargetOpcode(),
7254 C->isOpaque());
7255 case ISD::CTLS:
7256 // CTLS returns the number of extra sign bits so subtract one.
7257 return getConstant(Val.getNumSignBits() - 1, DL, VT,
7258 C->isTargetOpcode(), C->isOpaque());
7259 case ISD::UINT_TO_FP:
7260 case ISD::SINT_TO_FP: {
7262 (void)FPV.convertFromAPInt(Val, Opcode == ISD::SINT_TO_FP,
7264 return getConstantFP(FPV, DL, VT);
7265 }
7266 case ISD::FP16_TO_FP:
7267 case ISD::BF16_TO_FP: {
7268 bool Ignored;
7269 APFloat FPV(Opcode == ISD::FP16_TO_FP ? APFloat::IEEEhalf()
7270 : APFloat::BFloat(),
7271 (Val.getBitWidth() == 16) ? Val : Val.trunc(16));
7272
7273 // This can return overflow, underflow, or inexact; we don't care.
7274 // FIXME need to be more flexible about rounding mode.
7276 &Ignored);
7277 return getConstantFP(FPV, DL, VT);
7278 }
7279 case ISD::STEP_VECTOR:
7280 if (SDValue V = FoldSTEP_VECTOR(DL, VT, N1, *this))
7281 return V;
7282 break;
7283 case ISD::BITCAST:
7284 if (VT == MVT::f16 && C->getValueType(0) == MVT::i16)
7285 return getConstantFP(APFloat(APFloat::IEEEhalf(), Val), DL, VT);
7286 if (VT == MVT::f32 && C->getValueType(0) == MVT::i32)
7287 return getConstantFP(APFloat(APFloat::IEEEsingle(), Val), DL, VT);
7288 if (VT == MVT::f64 && C->getValueType(0) == MVT::i64)
7289 return getConstantFP(APFloat(APFloat::IEEEdouble(), Val), DL, VT);
7290 if (VT == MVT::f128 && C->getValueType(0) == MVT::i128)
7291 return getConstantFP(APFloat(APFloat::IEEEquad(), Val), DL, VT);
7292 break;
7293 }
7294 }
7295
7296 // Constant fold unary operations with a floating point constant operand.
7297 if (auto *C = dyn_cast<ConstantFPSDNode>(N1)) {
7298 APFloat V = C->getValueAPF(); // make copy
7299 switch (Opcode) {
7300 case ISD::FNEG:
7301 V.changeSign();
7302 return getConstantFP(V, DL, VT);
7303 case ISD::FABS:
7304 V.clearSign();
7305 return getConstantFP(V, DL, VT);
7306 case ISD::FCEIL: {
7307 APFloat::opStatus fs = V.roundToIntegral(APFloat::rmTowardPositive);
7309 return getConstantFP(V, DL, VT);
7310 return SDValue();
7311 }
7312 case ISD::FTRUNC: {
7313 APFloat::opStatus fs = V.roundToIntegral(APFloat::rmTowardZero);
7315 return getConstantFP(V, DL, VT);
7316 return SDValue();
7317 }
7318 case ISD::FFLOOR: {
7319 APFloat::opStatus fs = V.roundToIntegral(APFloat::rmTowardNegative);
7321 return getConstantFP(V, DL, VT);
7322 return SDValue();
7323 }
7324 case ISD::FP_EXTEND: {
7325 bool ignored;
7326 // This can return overflow, underflow, or inexact; we don't care.
7327 // FIXME need to be more flexible about rounding mode.
7328 (void)V.convert(VT.getFltSemantics(), APFloat::rmNearestTiesToEven,
7329 &ignored);
7330 return getConstantFP(V, DL, VT);
7331 }
7332 case ISD::FP_TO_SINT:
7333 case ISD::FP_TO_UINT: {
7334 bool ignored;
7335 APSInt IntVal(VT.getSizeInBits(), Opcode == ISD::FP_TO_UINT);
7336 // FIXME need to be more flexible about rounding mode.
7338 V.convertToInteger(IntVal, APFloat::rmTowardZero, &ignored);
7339 if (s == APFloat::opInvalidOp) // inexact is OK, in fact usual
7340 break;
7341 return getConstant(IntVal, DL, VT);
7342 }
7343 case ISD::FP_TO_FP16:
7344 case ISD::FP_TO_BF16: {
7345 bool Ignored;
7346 // This can return overflow, underflow, or inexact; we don't care.
7347 // FIXME need to be more flexible about rounding mode.
7348 (void)V.convert(Opcode == ISD::FP_TO_FP16 ? APFloat::IEEEhalf()
7349 : APFloat::BFloat(),
7351 return getConstant(V.bitcastToAPInt().getZExtValue(), DL, VT);
7352 }
7353 case ISD::BITCAST:
7354 if (VT == MVT::i16 && C->getValueType(0) == MVT::f16)
7355 return getConstant((uint16_t)V.bitcastToAPInt().getZExtValue(), DL,
7356 VT);
7357 if (VT == MVT::i16 && C->getValueType(0) == MVT::bf16)
7358 return getConstant((uint16_t)V.bitcastToAPInt().getZExtValue(), DL,
7359 VT);
7360 if (VT == MVT::i32 && C->getValueType(0) == MVT::f32)
7361 return getConstant((uint32_t)V.bitcastToAPInt().getZExtValue(), DL,
7362 VT);
7363 if (VT == MVT::i64 && C->getValueType(0) == MVT::f64)
7364 return getConstant(V.bitcastToAPInt().getZExtValue(), DL, VT);
7365 break;
7366 }
7367 }
7368
7369 // Early-out if we failed to constant fold a bitcast.
7370 if (Opcode == ISD::BITCAST)
7371 return SDValue();
7372 }
7373
7374 // Handle binops special cases.
7375 if (NumOps == 2) {
7376 if (SDValue CFP = foldConstantFPMath(Opcode, DL, VT, Ops))
7377 return CFP;
7378
7379 if (auto *C1 = dyn_cast<ConstantSDNode>(Ops[0])) {
7380 if (auto *C2 = dyn_cast<ConstantSDNode>(Ops[1])) {
7381 if (C1->isOpaque() || C2->isOpaque())
7382 return SDValue();
7383
7384 std::optional<APInt> FoldAttempt =
7385 FoldValue(Opcode, C1->getAPIntValue(), C2->getAPIntValue());
7386 if (!FoldAttempt)
7387 return SDValue();
7388
7389 SDValue Folded = getConstant(*FoldAttempt, DL, VT);
7390 assert((!Folded || !VT.isVector()) &&
7391 "Can't fold vectors ops with scalar operands");
7392 return Folded;
7393 }
7394 }
7395
7396 // fold (add Sym, c) -> Sym+c
7398 return FoldSymbolOffset(Opcode, VT, GA, Ops[1].getNode());
7399 if (TLI->isCommutativeBinOp(Opcode))
7401 return FoldSymbolOffset(Opcode, VT, GA, Ops[0].getNode());
7402
7403 // fold (sext_in_reg c1) -> c2
7404 if (Opcode == ISD::SIGN_EXTEND_INREG) {
7405 EVT EVT = cast<VTSDNode>(Ops[1])->getVT();
7406
7407 auto SignExtendInReg = [&](APInt Val, llvm::EVT ConstantVT) {
7408 unsigned FromBits = EVT.getScalarSizeInBits();
7409 Val <<= Val.getBitWidth() - FromBits;
7410 Val.ashrInPlace(Val.getBitWidth() - FromBits);
7411 return getConstant(Val, DL, ConstantVT);
7412 };
7413
7414 if (auto *C1 = dyn_cast<ConstantSDNode>(Ops[0])) {
7415 const APInt &Val = C1->getAPIntValue();
7416 return SignExtendInReg(Val, VT);
7417 }
7418
7420 SmallVector<SDValue, 8> ScalarOps;
7421 llvm::EVT OpVT = Ops[0].getOperand(0).getValueType();
7422 for (int I = 0, E = VT.getVectorNumElements(); I != E; ++I) {
7423 SDValue Op = Ops[0].getOperand(I);
7424 if (Op.isUndef()) {
7425 ScalarOps.push_back(getUNDEF(OpVT));
7426 continue;
7427 }
7428 const APInt &Val = cast<ConstantSDNode>(Op)->getAPIntValue();
7429 ScalarOps.push_back(SignExtendInReg(Val, OpVT));
7430 }
7431 return getBuildVector(VT, DL, ScalarOps);
7432 }
7433
7434 if (Ops[0].getOpcode() == ISD::SPLAT_VECTOR &&
7435 isa<ConstantSDNode>(Ops[0].getOperand(0)))
7436 return getNode(ISD::SPLAT_VECTOR, DL, VT,
7437 SignExtendInReg(Ops[0].getConstantOperandAPInt(0),
7438 Ops[0].getOperand(0).getValueType()));
7439 }
7440 }
7441
7442 // Handle fshl/fshr special cases.
7443 if (Opcode == ISD::FSHL || Opcode == ISD::FSHR) {
7444 auto *C1 = dyn_cast<ConstantSDNode>(Ops[0]);
7445 auto *C2 = dyn_cast<ConstantSDNode>(Ops[1]);
7446 auto *C3 = dyn_cast<ConstantSDNode>(Ops[2]);
7447
7448 if (C1 && C2 && C3) {
7449 if (C1->isOpaque() || C2->isOpaque() || C3->isOpaque())
7450 return SDValue();
7451 const APInt &V1 = C1->getAPIntValue(), &V2 = C2->getAPIntValue(),
7452 &V3 = C3->getAPIntValue();
7453
7454 APInt FoldedVal = Opcode == ISD::FSHL ? APIntOps::fshl(V1, V2, V3)
7455 : APIntOps::fshr(V1, V2, V3);
7456 return getConstant(FoldedVal, DL, VT);
7457 }
7458 }
7459
7460 // Handle fma/fmad special cases.
7461 if (Opcode == ISD::FMA || Opcode == ISD::FMAD || Opcode == ISD::FMULADD) {
7462 assert(VT.isFloatingPoint() && "This operator only applies to FP types!");
7463 assert(Ops[0].getValueType() == VT && Ops[1].getValueType() == VT &&
7464 Ops[2].getValueType() == VT && "FMA types must match!");
7468 if (C1 && C2 && C3) {
7469 APFloat V1 = C1->getValueAPF();
7470 const APFloat &V2 = C2->getValueAPF();
7471 const APFloat &V3 = C3->getValueAPF();
7472 if (Opcode == ISD::FMAD || Opcode == ISD::FMULADD) {
7475 } else
7477 return getConstantFP(V1, DL, VT);
7478 }
7479 }
7480
7481 // This is for vector folding only from here on.
7482 if (!VT.isVector())
7483 return SDValue();
7484
7485 ElementCount NumElts = VT.getVectorElementCount();
7486
7487 // See if we can fold through any bitcasted integer ops.
7488 if (NumOps == 2 && VT.isFixedLengthVector() && VT.isInteger() &&
7489 Ops[0].getValueType() == VT && Ops[1].getValueType() == VT &&
7490 (Ops[0].getOpcode() == ISD::BITCAST ||
7491 Ops[1].getOpcode() == ISD::BITCAST)) {
7494 auto *BV1 = dyn_cast<BuildVectorSDNode>(N1);
7495 auto *BV2 = dyn_cast<BuildVectorSDNode>(N2);
7496 if (BV1 && BV2 && N1.getValueType().isInteger() &&
7497 N2.getValueType().isInteger()) {
7498 bool IsLE = getDataLayout().isLittleEndian();
7499 unsigned EltBits = VT.getScalarSizeInBits();
7500 SmallVector<APInt> RawBits1, RawBits2;
7501 BitVector UndefElts1, UndefElts2;
7502 if (BV1->getConstantRawBits(IsLE, EltBits, RawBits1, UndefElts1) &&
7503 BV2->getConstantRawBits(IsLE, EltBits, RawBits2, UndefElts2)) {
7504 SmallVector<APInt> RawBits;
7505 for (unsigned I = 0, E = NumElts.getFixedValue(); I != E; ++I) {
7506 std::optional<APInt> Fold = FoldValueWithUndef(
7507 Opcode, RawBits1[I], UndefElts1[I], RawBits2[I], UndefElts2[I]);
7508 if (!Fold)
7509 break;
7510 RawBits.push_back(*Fold);
7511 }
7512 if (RawBits.size() == NumElts.getFixedValue()) {
7513 // We have constant folded, but we might need to cast this again back
7514 // to the original (possibly legalized) type.
7515 EVT BVVT, BVEltVT;
7516 if (N1.getValueType() == VT) {
7517 BVVT = N1.getValueType();
7518 BVEltVT = BV1->getOperand(0).getValueType();
7519 } else {
7520 BVVT = N2.getValueType();
7521 BVEltVT = BV2->getOperand(0).getValueType();
7522 }
7523 unsigned BVEltBits = BVEltVT.getSizeInBits();
7524 SmallVector<APInt> DstBits;
7525 BitVector DstUndefs;
7527 DstBits, RawBits, DstUndefs,
7528 BitVector(RawBits.size(), false));
7529 SmallVector<SDValue> Ops(DstBits.size(), getUNDEF(BVEltVT));
7530 for (unsigned I = 0, E = DstBits.size(); I != E; ++I) {
7531 if (DstUndefs[I])
7532 continue;
7533 Ops[I] = getConstant(DstBits[I].sext(BVEltBits), DL, BVEltVT);
7534 }
7535 return getBitcast(VT, getBuildVector(BVVT, DL, Ops));
7536 }
7537 }
7538 }
7539 }
7540
7541 // Fold (mul step_vector(C0), C1) to (step_vector(C0 * C1)).
7542 // (shl step_vector(C0), C1) -> (step_vector(C0 << C1))
7543 if ((Opcode == ISD::MUL || Opcode == ISD::SHL) &&
7544 Ops[0].getOpcode() == ISD::STEP_VECTOR) {
7545 APInt RHSVal;
7546 if (ISD::isConstantSplatVector(Ops[1].getNode(), RHSVal)) {
7547 APInt NewStep = Opcode == ISD::MUL
7548 ? Ops[0].getConstantOperandAPInt(0) * RHSVal
7549 : Ops[0].getConstantOperandAPInt(0) << RHSVal;
7550 return getStepVector(DL, VT, NewStep);
7551 }
7552 }
7553
7554 auto IsScalarOrSameVectorSize = [NumElts](const SDValue &Op) {
7555 return !Op.getValueType().isVector() ||
7556 Op.getValueType().getVectorElementCount() == NumElts;
7557 };
7558
7559 auto IsBuildVectorSplatVectorOrUndef = [](const SDValue &Op) {
7560 return Op.isUndef() || Op.getOpcode() == ISD::CONDCODE ||
7561 Op.getOpcode() == ISD::BUILD_VECTOR ||
7562 Op.getOpcode() == ISD::SPLAT_VECTOR;
7563 };
7564
7565 // All operands must be vector types with the same number of elements as
7566 // the result type and must be either UNDEF or a build/splat vector
7567 // or UNDEF scalars.
7568 if (!llvm::all_of(Ops, IsBuildVectorSplatVectorOrUndef) ||
7569 !llvm::all_of(Ops, IsScalarOrSameVectorSize))
7570 return SDValue();
7571
7572 // If we are comparing vectors, then the result needs to be a i1 boolean that
7573 // is then extended back to the legal result type depending on how booleans
7574 // are represented.
7575 EVT SVT = (Opcode == ISD::SETCC ? MVT::i1 : VT.getScalarType());
7576 ISD::NodeType ExtendCode =
7577 (Opcode == ISD::SETCC && SVT != VT.getScalarType())
7578 ? TargetLowering::getExtendForContent(TLI->getBooleanContents(VT))
7580
7581 // Find legal integer scalar type for constant promotion and
7582 // ensure that its scalar size is at least as large as source.
7583 EVT LegalSVT = VT.getScalarType();
7584 if (NewNodesMustHaveLegalTypes && LegalSVT.isInteger()) {
7585 LegalSVT = TLI->getTypeToTransformTo(*getContext(), LegalSVT);
7586 if (LegalSVT.bitsLT(VT.getScalarType()))
7587 return SDValue();
7588 }
7589
7590 // For scalable vector types we know we're dealing with SPLAT_VECTORs. We
7591 // only have one operand to check. For fixed-length vector types we may have
7592 // a combination of BUILD_VECTOR and SPLAT_VECTOR.
7593 unsigned NumVectorElts = NumElts.isScalable() ? 1 : NumElts.getFixedValue();
7594
7595 // Constant fold each scalar lane separately.
7596 SmallVector<SDValue, 4> ScalarResults;
7597 for (unsigned I = 0; I != NumVectorElts; I++) {
7598 SmallVector<SDValue, 4> ScalarOps;
7599 for (SDValue Op : Ops) {
7600 EVT InSVT = Op.getValueType().getScalarType();
7601 if (Op.getOpcode() != ISD::BUILD_VECTOR &&
7602 Op.getOpcode() != ISD::SPLAT_VECTOR) {
7603 if (Op.isUndef())
7604 ScalarOps.push_back(getUNDEF(InSVT));
7605 else
7606 ScalarOps.push_back(Op);
7607 continue;
7608 }
7609
7610 SDValue ScalarOp =
7611 Op.getOperand(Op.getOpcode() == ISD::SPLAT_VECTOR ? 0 : I);
7612 EVT ScalarVT = ScalarOp.getValueType();
7613
7614 // Build vector (integer) scalar operands may need implicit
7615 // truncation - do this before constant folding.
7616 if (ScalarVT.isInteger() && ScalarVT.bitsGT(InSVT)) {
7617 // Don't create illegally-typed nodes unless they're constants or undef
7618 // - if we fail to constant fold we can't guarantee the (dead) nodes
7619 // we're creating will be cleaned up before being visited for
7620 // legalization.
7621 if (NewNodesMustHaveLegalTypes && !ScalarOp.isUndef() &&
7622 !isa<ConstantSDNode>(ScalarOp) &&
7623 TLI->getTypeAction(*getContext(), InSVT) !=
7625 return SDValue();
7626 ScalarOp = getNode(ISD::TRUNCATE, DL, InSVT, ScalarOp);
7627 }
7628
7629 ScalarOps.push_back(ScalarOp);
7630 }
7631
7632 // Constant fold the scalar operands.
7633 SDValue ScalarResult = getNode(Opcode, DL, SVT, ScalarOps, Flags);
7634
7635 // Scalar folding only succeeded if the result is a constant or UNDEF.
7636 if (!ScalarResult.isUndef() && ScalarResult.getOpcode() != ISD::Constant &&
7637 ScalarResult.getOpcode() != ISD::ConstantFP)
7638 return SDValue();
7639
7640 // Legalize the (integer) scalar constant if necessary. We only do
7641 // this once we know the folding succeeded, since otherwise we would
7642 // get a node with illegal type which has a user.
7643 if (LegalSVT != SVT)
7644 ScalarResult = getNode(ExtendCode, DL, LegalSVT, ScalarResult);
7645
7646 ScalarResults.push_back(ScalarResult);
7647 }
7648
7649 SDValue V = NumElts.isScalable() ? getSplatVector(VT, DL, ScalarResults[0])
7650 : getBuildVector(VT, DL, ScalarResults);
7651 NewSDValueDbgMsg(V, "New node fold constant vector: ", this);
7652 return V;
7653}
7654
7657 // TODO: Add support for unary/ternary fp opcodes.
7658 if (Ops.size() != 2)
7659 return SDValue();
7660
7661 // TODO: We don't do any constant folding for strict FP opcodes here, but we
7662 // should. That will require dealing with a potentially non-default
7663 // rounding mode, checking the "opStatus" return value from the APFloat
7664 // math calculations, and possibly other variations.
7665 SDValue N1 = Ops[0];
7666 SDValue N2 = Ops[1];
7667 ConstantFPSDNode *N1CFP = isConstOrConstSplatFP(N1, /*AllowUndefs*/ false);
7668 ConstantFPSDNode *N2CFP = isConstOrConstSplatFP(N2, /*AllowUndefs*/ false);
7669 if (N1CFP && N2CFP) {
7670 APFloat C1 = N1CFP->getValueAPF(); // make copy
7671 const APFloat &C2 = N2CFP->getValueAPF();
7672 switch (Opcode) {
7673 case ISD::FADD:
7675 return getConstantFP(C1, DL, VT);
7676 case ISD::FSUB:
7678 return getConstantFP(C1, DL, VT);
7679 case ISD::FMUL:
7681 return getConstantFP(C1, DL, VT);
7682 case ISD::FDIV:
7684 return getConstantFP(C1, DL, VT);
7685 case ISD::FREM:
7686 C1.mod(C2);
7687 return getConstantFP(C1, DL, VT);
7688 case ISD::FCOPYSIGN:
7689 C1.copySign(C2);
7690 return getConstantFP(C1, DL, VT);
7691 case ISD::FMINNUM:
7692 if (C1.isSignaling() || C2.isSignaling())
7693 return SDValue();
7694 return getConstantFP(minnum(C1, C2), DL, VT);
7695 case ISD::FMAXNUM:
7696 if (C1.isSignaling() || C2.isSignaling())
7697 return SDValue();
7698 return getConstantFP(maxnum(C1, C2), DL, VT);
7699 case ISD::FMINIMUM:
7700 return getConstantFP(minimum(C1, C2), DL, VT);
7701 case ISD::FMAXIMUM:
7702 return getConstantFP(maximum(C1, C2), DL, VT);
7703 case ISD::FMINIMUMNUM:
7704 return getConstantFP(minimumnum(C1, C2), DL, VT);
7705 case ISD::FMAXIMUMNUM:
7706 return getConstantFP(maximumnum(C1, C2), DL, VT);
7707 default: break;
7708 }
7709 }
7710 if (N1CFP && Opcode == ISD::FP_ROUND) {
7711 APFloat C1 = N1CFP->getValueAPF(); // make copy
7712 bool Unused;
7713 // This can return overflow, underflow, or inexact; we don't care.
7714 // FIXME need to be more flexible about rounding mode.
7716 &Unused);
7717 return getConstantFP(C1, DL, VT);
7718 }
7719
7720 switch (Opcode) {
7721 case ISD::FSUB:
7722 // -0.0 - undef --> undef (consistent with "fneg undef")
7723 if (ConstantFPSDNode *N1C = isConstOrConstSplatFP(N1, /*AllowUndefs*/ true))
7724 if (N1C && N1C->getValueAPF().isNegZero() && N2.isUndef())
7725 return getUNDEF(VT);
7726 [[fallthrough]];
7727
7728 case ISD::FADD:
7729 case ISD::FMUL:
7730 case ISD::FDIV:
7731 case ISD::FREM:
7732 // If both operands are undef, the result is undef. If 1 operand is undef,
7733 // the result is NaN. This should match the behavior of the IR optimizer.
7734 if (N1.isUndef() && N2.isUndef())
7735 return getUNDEF(VT);
7736 if (N1.isUndef() || N2.isUndef())
7738 }
7739 return SDValue();
7740}
7741
7743 const SDLoc &DL, EVT DstEltVT) {
7744 EVT SrcEltVT = BV->getValueType(0).getVectorElementType();
7745
7746 // If this is already the right type, we're done.
7747 if (SrcEltVT == DstEltVT)
7748 return SDValue(BV, 0);
7749
7750 unsigned SrcBitSize = SrcEltVT.getSizeInBits();
7751 unsigned DstBitSize = DstEltVT.getSizeInBits();
7752
7753 // If this is a conversion of N elements of one type to N elements of another
7754 // type, convert each element. This handles FP<->INT cases.
7755 if (SrcBitSize == DstBitSize) {
7757 for (SDValue Op : BV->op_values()) {
7758 // If the vector element type is not legal, the BUILD_VECTOR operands
7759 // are promoted and implicitly truncated. Make that explicit here.
7760 if (Op.getValueType() != SrcEltVT)
7761 Op = getNode(ISD::TRUNCATE, DL, SrcEltVT, Op);
7762 Ops.push_back(getBitcast(DstEltVT, Op));
7763 }
7764 EVT VT = EVT::getVectorVT(*getContext(), DstEltVT,
7766 return getBuildVector(VT, DL, Ops);
7767 }
7768
7769 // Otherwise, we're growing or shrinking the elements. To avoid having to
7770 // handle annoying details of growing/shrinking FP values, we convert them to
7771 // int first.
7772 if (SrcEltVT.isFloatingPoint()) {
7773 // Convert the input float vector to a int vector where the elements are the
7774 // same sizes.
7775 EVT IntEltVT = EVT::getIntegerVT(*getContext(), SrcEltVT.getSizeInBits());
7776 if (SDValue Tmp = FoldConstantBuildVector(BV, DL, IntEltVT))
7778 DstEltVT);
7779 return SDValue();
7780 }
7781
7782 // Now we know the input is an integer vector. If the output is a FP type,
7783 // convert to integer first, then to FP of the right size.
7784 if (DstEltVT.isFloatingPoint()) {
7785 EVT IntEltVT = EVT::getIntegerVT(*getContext(), DstEltVT.getSizeInBits());
7786 if (SDValue Tmp = FoldConstantBuildVector(BV, DL, IntEltVT))
7788 DstEltVT);
7789 return SDValue();
7790 }
7791
7792 // Okay, we know the src/dst types are both integers of differing types.
7793 assert(SrcEltVT.isInteger() && DstEltVT.isInteger());
7794
7795 // Extract the constant raw bit data.
7796 BitVector UndefElements;
7797 SmallVector<APInt> RawBits;
7798 bool IsLE = getDataLayout().isLittleEndian();
7799 if (!BV->getConstantRawBits(IsLE, DstBitSize, RawBits, UndefElements))
7800 return SDValue();
7801
7803 for (unsigned I = 0, E = RawBits.size(); I != E; ++I) {
7804 if (UndefElements[I])
7805 Ops.push_back(getUNDEF(DstEltVT));
7806 else
7807 Ops.push_back(getConstant(RawBits[I], DL, DstEltVT));
7808 }
7809
7810 EVT VT = EVT::getVectorVT(*getContext(), DstEltVT, Ops.size());
7811 return getBuildVector(VT, DL, Ops);
7812}
7813
7815 assert(Val.getValueType().isInteger() && "Invalid AssertAlign!");
7816
7817 // There's no need to assert on a byte-aligned pointer. All pointers are at
7818 // least byte aligned.
7819 if (A == Align(1))
7820 return Val;
7821
7822 SDVTList VTs = getVTList(Val.getValueType());
7824 AddNodeIDNode(ID, ISD::AssertAlign, VTs, {Val});
7825 ID.AddInteger(A.value());
7826
7827 void *IP = nullptr;
7828 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP))
7829 return SDValue(E, 0);
7830
7831 auto *N =
7832 newSDNode<AssertAlignSDNode>(DL.getIROrder(), DL.getDebugLoc(), VTs, A);
7833 createOperands(N, {Val});
7834
7835 CSEMap.InsertNode(N, IP);
7836 InsertNode(N);
7837
7838 SDValue V(N, 0);
7839 NewSDValueDbgMsg(V, "Creating new node: ", this);
7840 return V;
7841}
7842
7843SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
7844 SDValue N1, SDValue N2) {
7845 SDNodeFlags Flags;
7846 if (Inserter)
7847 Flags = Inserter->getFlags();
7848 return getNode(Opcode, DL, VT, N1, N2, Flags);
7849}
7850
7852 SDValue &N2) const {
7853 if (!TLI->isCommutativeBinOp(Opcode))
7854 return;
7855
7856 // Canonicalize:
7857 // binop(const, nonconst) -> binop(nonconst, const)
7860 bool N1CFP = isConstantFPBuildVectorOrConstantFP(N1);
7861 bool N2CFP = isConstantFPBuildVectorOrConstantFP(N2);
7862 if ((N1C && !N2C) || (N1CFP && !N2CFP))
7863 std::swap(N1, N2);
7864
7865 // Canonicalize:
7866 // binop(splat(x), step_vector) -> binop(step_vector, splat(x))
7867 else if (N1.getOpcode() == ISD::SPLAT_VECTOR &&
7869 std::swap(N1, N2);
7870}
7871
7872SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
7873 SDValue N1, SDValue N2, const SDNodeFlags Flags) {
7875 N2.getOpcode() != ISD::DELETED_NODE &&
7876 "Operand is DELETED_NODE!");
7877
7878 canonicalizeCommutativeBinop(Opcode, N1, N2);
7879
7880 auto *N1C = dyn_cast<ConstantSDNode>(N1);
7881 auto *N2C = dyn_cast<ConstantSDNode>(N2);
7882
7883 // Don't allow undefs in vector splats - we might be returning N2 when folding
7884 // to zero etc.
7885 ConstantSDNode *N2CV =
7886 isConstOrConstSplat(N2, /*AllowUndefs*/ false, /*AllowTruncation*/ true);
7887
7888 switch (Opcode) {
7889 default: break;
7890 case ISD::TokenFactor:
7891 assert(VT == MVT::Other && N1.getValueType() == MVT::Other &&
7892 N2.getValueType() == MVT::Other && "Invalid token factor!");
7893 // Fold trivial token factors.
7894 if (N1.getOpcode() == ISD::EntryToken) return N2;
7895 if (N2.getOpcode() == ISD::EntryToken) return N1;
7896 if (N1 == N2) return N1;
7897 break;
7898 case ISD::BUILD_VECTOR: {
7899 // Attempt to simplify BUILD_VECTOR.
7900 SDValue Ops[] = {N1, N2};
7901 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
7902 return V;
7903 break;
7904 }
7905 case ISD::CONCAT_VECTORS: {
7906 SDValue Ops[] = {N1, N2};
7907 if (SDValue V = foldCONCAT_VECTORS(DL, VT, Ops, *this))
7908 return V;
7909 break;
7910 }
7911 case ISD::AND:
7912 assert(VT.isInteger() && "This operator does not apply to FP types!");
7913 assert(N1.getValueType() == N2.getValueType() &&
7914 N1.getValueType() == VT && "Binary operator types must match!");
7915 // (X & 0) -> 0. This commonly occurs when legalizing i64 values, so it's
7916 // worth handling here.
7917 if (N2CV && N2CV->isZero())
7918 return N2;
7919 if (N2CV && N2CV->isAllOnes()) // X & -1 -> X
7920 return N1;
7921 break;
7922 case ISD::OR:
7923 case ISD::XOR:
7924 case ISD::ADD:
7925 case ISD::PTRADD:
7926 case ISD::SUB:
7927 assert(VT.isInteger() && "This operator does not apply to FP types!");
7928 assert(N1.getValueType() == N2.getValueType() &&
7929 N1.getValueType() == VT && "Binary operator types must match!");
7930 // The equal operand types requirement is unnecessarily strong for PTRADD.
7931 // However, the SelectionDAGBuilder does not generate PTRADDs with different
7932 // operand types, and we'd need to re-implement GEP's non-standard wrapping
7933 // logic everywhere where PTRADDs may be folded or combined to properly
7934 // support them. If/when we introduce pointer types to the SDAG, we will
7935 // need to relax this constraint.
7936
7937 // (X ^|+- 0) -> X. This commonly occurs when legalizing i64 values, so
7938 // it's worth handling here.
7939 if (N2CV && N2CV->isZero())
7940 return N1;
7941 if ((Opcode == ISD::ADD || Opcode == ISD::SUB) &&
7942 VT.getScalarType() == MVT::i1)
7943 return getNode(ISD::XOR, DL, VT, N1, N2);
7944 // Fold (add (vscale * C0), (vscale * C1)) to (vscale * (C0 + C1)).
7945 if (Opcode == ISD::ADD && N1.getOpcode() == ISD::VSCALE &&
7946 N2.getOpcode() == ISD::VSCALE) {
7947 const APInt &C1 = N1->getConstantOperandAPInt(0);
7948 const APInt &C2 = N2->getConstantOperandAPInt(0);
7949 return getVScale(DL, VT, C1 + C2);
7950 }
7951 break;
7952 case ISD::MUL:
7953 assert(VT.isInteger() && "This operator does not apply to FP types!");
7954 assert(N1.getValueType() == N2.getValueType() &&
7955 N1.getValueType() == VT && "Binary operator types must match!");
7956 if (VT.getScalarType() == MVT::i1)
7957 return getNode(ISD::AND, DL, VT, N1, N2);
7958 if (N2CV && N2CV->isZero())
7959 return N2;
7960 if (N2C && (N1.getOpcode() == ISD::VSCALE) && Flags.hasNoSignedWrap()) {
7961 const APInt &MulImm = N1->getConstantOperandAPInt(0);
7962 const APInt &N2CImm = N2C->getAPIntValue();
7963 return getVScale(DL, VT, MulImm * N2CImm);
7964 }
7965 break;
7966 case ISD::UDIV:
7967 case ISD::UREM:
7968 case ISD::MULHU:
7969 case ISD::MULHS:
7970 case ISD::SDIV:
7971 case ISD::SREM:
7972 case ISD::SADDSAT:
7973 case ISD::SSUBSAT:
7974 case ISD::UADDSAT:
7975 case ISD::USUBSAT:
7976 assert(VT.isInteger() && "This operator does not apply to FP types!");
7977 assert(N1.getValueType() == N2.getValueType() &&
7978 N1.getValueType() == VT && "Binary operator types must match!");
7979 if (VT.getScalarType() == MVT::i1) {
7980 // fold (add_sat x, y) -> (or x, y) for bool types.
7981 if (Opcode == ISD::SADDSAT || Opcode == ISD::UADDSAT)
7982 return getNode(ISD::OR, DL, VT, N1, N2);
7983 // fold (sub_sat x, y) -> (and x, ~y) for bool types.
7984 if (Opcode == ISD::SSUBSAT || Opcode == ISD::USUBSAT)
7985 return getNode(ISD::AND, DL, VT, N1, getNOT(DL, N2, VT));
7986 }
7987 break;
7988 case ISD::SCMP:
7989 case ISD::UCMP:
7990 assert(N1.getValueType() == N2.getValueType() &&
7991 "Types of operands of UCMP/SCMP must match");
7992 assert(N1.getValueType().isVector() == VT.isVector() &&
7993 "Operands and return type of must both be scalars or vectors");
7994 if (VT.isVector())
7997 "Result and operands must have the same number of elements");
7998 break;
7999 case ISD::AVGFLOORS:
8000 case ISD::AVGFLOORU:
8001 case ISD::AVGCEILS:
8002 case ISD::AVGCEILU:
8003 assert(VT.isInteger() && "This operator does not apply to FP types!");
8004 assert(N1.getValueType() == N2.getValueType() &&
8005 N1.getValueType() == VT && "Binary operator types must match!");
8006 break;
8007 case ISD::ABDS:
8008 case ISD::ABDU:
8009 assert(VT.isInteger() && "This operator does not apply to FP types!");
8010 assert(N1.getValueType() == N2.getValueType() &&
8011 N1.getValueType() == VT && "Binary operator types must match!");
8012 if (VT.getScalarType() == MVT::i1)
8013 return getNode(ISD::XOR, DL, VT, N1, N2);
8014 break;
8015 case ISD::SMIN:
8016 case ISD::UMAX:
8017 assert(VT.isInteger() && "This operator does not apply to FP types!");
8018 assert(N1.getValueType() == N2.getValueType() &&
8019 N1.getValueType() == VT && "Binary operator types must match!");
8020 if (VT.getScalarType() == MVT::i1)
8021 return getNode(ISD::OR, DL, VT, N1, N2);
8022 break;
8023 case ISD::SMAX:
8024 case ISD::UMIN:
8025 assert(VT.isInteger() && "This operator does not apply to FP types!");
8026 assert(N1.getValueType() == N2.getValueType() &&
8027 N1.getValueType() == VT && "Binary operator types must match!");
8028 if (VT.getScalarType() == MVT::i1)
8029 return getNode(ISD::AND, DL, VT, N1, N2);
8030 break;
8031 case ISD::FADD:
8032 case ISD::FSUB:
8033 case ISD::FMUL:
8034 case ISD::FDIV:
8035 case ISD::FREM:
8036 assert(VT.isFloatingPoint() && "This operator only applies to FP types!");
8037 assert(N1.getValueType() == N2.getValueType() &&
8038 N1.getValueType() == VT && "Binary operator types must match!");
8039 if (SDValue V = simplifyFPBinop(Opcode, N1, N2, Flags))
8040 return V;
8041 break;
8042 case ISD::FCOPYSIGN: // N1 and result must match. N1/N2 need not match.
8043 assert(N1.getValueType() == VT &&
8046 "Invalid FCOPYSIGN!");
8047 break;
8048 case ISD::SHL:
8049 if (N2C && (N1.getOpcode() == ISD::VSCALE) && Flags.hasNoSignedWrap()) {
8050 const APInt &MulImm = N1->getConstantOperandAPInt(0);
8051 const APInt &ShiftImm = N2C->getAPIntValue();
8052 return getVScale(DL, VT, MulImm << ShiftImm);
8053 }
8054 [[fallthrough]];
8055 case ISD::SRA:
8056 case ISD::SRL:
8057 if (SDValue V = simplifyShift(N1, N2))
8058 return V;
8059 [[fallthrough]];
8060 case ISD::ROTL:
8061 case ISD::ROTR:
8062 case ISD::SSHLSAT:
8063 case ISD::USHLSAT:
8064 assert(VT == N1.getValueType() &&
8065 "Shift operators return type must be the same as their first arg");
8066 assert(VT.isInteger() && N2.getValueType().isInteger() &&
8067 "Shifts only work on integers");
8068 assert((!VT.isVector() || VT == N2.getValueType()) &&
8069 "Vector shift amounts must be in the same as their first arg");
8070 // Verify that the shift amount VT is big enough to hold valid shift
8071 // amounts. This catches things like trying to shift an i1024 value by an
8072 // i8, which is easy to fall into in generic code that uses
8073 // TLI.getShiftAmount().
8076 "Invalid use of small shift amount with oversized value!");
8077
8078 // Always fold shifts of i1 values so the code generator doesn't need to
8079 // handle them. Since we know the size of the shift has to be less than the
8080 // size of the value, the shift/rotate count is guaranteed to be zero.
8081 if (VT == MVT::i1)
8082 return N1;
8083 if (N2CV && N2CV->isZero())
8084 return N1;
8085 break;
8086 case ISD::FP_ROUND:
8088 VT.bitsLE(N1.getValueType()) && N2C &&
8089 (N2C->getZExtValue() == 0 || N2C->getZExtValue() == 1) &&
8090 N2.getOpcode() == ISD::TargetConstant && "Invalid FP_ROUND!");
8091 if (N1.getValueType() == VT) return N1; // noop conversion.
8092 break;
8093 case ISD::AssertNoFPClass: {
8095 "AssertNoFPClass is used for a non-floating type");
8096 assert(isa<ConstantSDNode>(N2) && "NoFPClass is not Constant");
8097 FPClassTest NoFPClass = static_cast<FPClassTest>(N2->getAsZExtVal());
8098 assert(llvm::to_underlying(NoFPClass) <=
8100 "FPClassTest value too large");
8101 (void)NoFPClass;
8102 break;
8103 }
8104 case ISD::AssertSext:
8105 case ISD::AssertZext: {
8106 EVT EVT = cast<VTSDNode>(N2)->getVT();
8107 assert(VT == N1.getValueType() && "Not an inreg extend!");
8108 assert(VT.isInteger() && EVT.isInteger() &&
8109 "Cannot *_EXTEND_INREG FP types");
8110 assert(!EVT.isVector() &&
8111 "AssertSExt/AssertZExt type should be the vector element type "
8112 "rather than the vector type!");
8113 assert(EVT.bitsLE(VT.getScalarType()) && "Not extending!");
8114 if (VT.getScalarType() == EVT) return N1; // noop assertion.
8115 break;
8116 }
8118 EVT EVT = cast<VTSDNode>(N2)->getVT();
8119 assert(VT == N1.getValueType() && "Not an inreg extend!");
8120 assert(VT.isInteger() && EVT.isInteger() &&
8121 "Cannot *_EXTEND_INREG FP types");
8122 assert(EVT.isVector() == VT.isVector() &&
8123 "SIGN_EXTEND_INREG type should be vector iff the operand "
8124 "type is vector!");
8125 assert((!EVT.isVector() ||
8127 "Vector element counts must match in SIGN_EXTEND_INREG");
8128 assert(EVT.getScalarType().bitsLE(VT.getScalarType()) && "Not extending!");
8129 if (EVT == VT) return N1; // Not actually extending
8130 break;
8131 }
8133 case ISD::FP_TO_UINT_SAT: {
8134 assert(VT.isInteger() && cast<VTSDNode>(N2)->getVT().isInteger() &&
8135 N1.getValueType().isFloatingPoint() && "Invalid FP_TO_*INT_SAT");
8136 assert(N1.getValueType().isVector() == VT.isVector() &&
8137 "FP_TO_*INT_SAT type should be vector iff the operand type is "
8138 "vector!");
8139 assert((!VT.isVector() || VT.getVectorElementCount() ==
8141 "Vector element counts must match in FP_TO_*INT_SAT");
8142 assert(!cast<VTSDNode>(N2)->getVT().isVector() &&
8143 "Type to saturate to must be a scalar.");
8144 assert(cast<VTSDNode>(N2)->getVT().bitsLE(VT.getScalarType()) &&
8145 "Not extending!");
8146 break;
8147 }
8150 "The result of EXTRACT_VECTOR_ELT must be at least as wide as the \
8151 element type of the vector.");
8152
8153 // Extract from an undefined value or using an undefined index is undefined.
8154 if (N1.isUndef() || N2.isUndef())
8155 return getUNDEF(VT);
8156
8157 // EXTRACT_VECTOR_ELT of out-of-bounds element is an UNDEF for fixed length
8158 // vectors. For scalable vectors we will provide appropriate support for
8159 // dealing with arbitrary indices.
8160 if (N2C && N1.getValueType().isFixedLengthVector() &&
8161 N2C->getAPIntValue().uge(N1.getValueType().getVectorNumElements()))
8162 return getUNDEF(VT);
8163
8164 // EXTRACT_VECTOR_ELT of CONCAT_VECTORS is often formed while lowering is
8165 // expanding copies of large vectors from registers. This only works for
8166 // fixed length vectors, since we need to know the exact number of
8167 // elements.
8168 if (N2C && N1.getOpcode() == ISD::CONCAT_VECTORS &&
8170 unsigned Factor = N1.getOperand(0).getValueType().getVectorNumElements();
8171 return getExtractVectorElt(DL, VT,
8172 N1.getOperand(N2C->getZExtValue() / Factor),
8173 N2C->getZExtValue() % Factor);
8174 }
8175
8176 // EXTRACT_VECTOR_ELT of BUILD_VECTOR or SPLAT_VECTOR is often formed while
8177 // lowering is expanding large vector constants.
8178 if (N2C && (N1.getOpcode() == ISD::BUILD_VECTOR ||
8179 N1.getOpcode() == ISD::SPLAT_VECTOR)) {
8182 "BUILD_VECTOR used for scalable vectors");
8183 unsigned Index =
8184 N1.getOpcode() == ISD::BUILD_VECTOR ? N2C->getZExtValue() : 0;
8185 SDValue Elt = N1.getOperand(Index);
8186
8187 if (VT != Elt.getValueType())
8188 // If the vector element type is not legal, the BUILD_VECTOR operands
8189 // are promoted and implicitly truncated, and the result implicitly
8190 // extended. Make that explicit here.
8191 Elt = getAnyExtOrTrunc(Elt, DL, VT);
8192
8193 return Elt;
8194 }
8195
8196 // EXTRACT_VECTOR_ELT of INSERT_VECTOR_ELT is often formed when vector
8197 // operations are lowered to scalars.
8198 if (N1.getOpcode() == ISD::INSERT_VECTOR_ELT) {
8199 // If the indices are the same, return the inserted element else
8200 // if the indices are known different, extract the element from
8201 // the original vector.
8202 SDValue N1Op2 = N1.getOperand(2);
8204
8205 if (N1Op2C && N2C) {
8206 if (N1Op2C->getZExtValue() == N2C->getZExtValue()) {
8207 if (VT == N1.getOperand(1).getValueType())
8208 return N1.getOperand(1);
8209 if (VT.isFloatingPoint()) {
8211 return getFPExtendOrRound(N1.getOperand(1), DL, VT);
8212 }
8213 return getSExtOrTrunc(N1.getOperand(1), DL, VT);
8214 }
8215 return getNode(ISD::EXTRACT_VECTOR_ELT, DL, VT, N1.getOperand(0), N2);
8216 }
8217 }
8218
8219 // EXTRACT_VECTOR_ELT of v1iX EXTRACT_SUBVECTOR could be formed
8220 // when vector types are scalarized and v1iX is legal.
8221 // vextract (v1iX extract_subvector(vNiX, Idx)) -> vextract(vNiX,Idx).
8222 // Here we are completely ignoring the extract element index (N2),
8223 // which is fine for fixed width vectors, since any index other than 0
8224 // is undefined anyway. However, this cannot be ignored for scalable
8225 // vectors - in theory we could support this, but we don't want to do this
8226 // without a profitability check.
8227 if (N1.getOpcode() == ISD::EXTRACT_SUBVECTOR &&
8229 N1.getValueType().getVectorNumElements() == 1) {
8230 return getNode(ISD::EXTRACT_VECTOR_ELT, DL, VT, N1.getOperand(0),
8231 N1.getOperand(1));
8232 }
8233 break;
8235 assert(N2C && (unsigned)N2C->getZExtValue() < 2 && "Bad EXTRACT_ELEMENT!");
8236 assert(!N1.getValueType().isVector() && !VT.isVector() &&
8237 (N1.getValueType().isInteger() == VT.isInteger()) &&
8238 N1.getValueType() != VT &&
8239 "Wrong types for EXTRACT_ELEMENT!");
8240
8241 // EXTRACT_ELEMENT of BUILD_PAIR is often formed while legalize is expanding
8242 // 64-bit integers into 32-bit parts. Instead of building the extract of
8243 // the BUILD_PAIR, only to have legalize rip it apart, just do it now.
8244 if (N1.getOpcode() == ISD::BUILD_PAIR)
8245 return N1.getOperand(N2C->getZExtValue());
8246
8247 // EXTRACT_ELEMENT of a constant int is also very common.
8248 if (N1C) {
8249 unsigned ElementSize = VT.getSizeInBits();
8250 unsigned Shift = ElementSize * N2C->getZExtValue();
8251 const APInt &Val = N1C->getAPIntValue();
8252 return getConstant(Val.extractBits(ElementSize, Shift), DL, VT);
8253 }
8254 break;
8256 EVT N1VT = N1.getValueType();
8257 assert(VT.isVector() && N1VT.isVector() &&
8258 "Extract subvector VTs must be vectors!");
8260 "Extract subvector VTs must have the same element type!");
8261 assert((VT.isFixedLengthVector() || N1VT.isScalableVector()) &&
8262 "Cannot extract a scalable vector from a fixed length vector!");
8263 assert((VT.isScalableVector() != N1VT.isScalableVector() ||
8265 "Extract subvector must be from larger vector to smaller vector!");
8266 assert(N2C && "Extract subvector index must be a constant");
8267 assert((VT.isScalableVector() != N1VT.isScalableVector() ||
8268 (VT.getVectorMinNumElements() + N2C->getZExtValue()) <=
8269 N1VT.getVectorMinNumElements()) &&
8270 "Extract subvector overflow!");
8271 assert(N2C->getAPIntValue().getBitWidth() ==
8272 TLI->getVectorIdxWidth(getDataLayout()) &&
8273 "Constant index for EXTRACT_SUBVECTOR has an invalid size");
8274 assert(N2C->getZExtValue() % VT.getVectorMinNumElements() == 0 &&
8275 "Extract index is not a multiple of the output vector length");
8276
8277 // Trivial extraction.
8278 if (VT == N1VT)
8279 return N1;
8280
8281 // EXTRACT_SUBVECTOR of an UNDEF is an UNDEF.
8282 if (N1.isUndef())
8283 return getUNDEF(VT);
8284
8285 // EXTRACT_SUBVECTOR of CONCAT_VECTOR can be simplified if the pieces of
8286 // the concat have the same type as the extract.
8287 if (N1.getOpcode() == ISD::CONCAT_VECTORS &&
8288 VT == N1.getOperand(0).getValueType()) {
8289 unsigned Factor = VT.getVectorMinNumElements();
8290 return N1.getOperand(N2C->getZExtValue() / Factor);
8291 }
8292
8293 // EXTRACT_SUBVECTOR of INSERT_SUBVECTOR is often created
8294 // during shuffle legalization.
8295 if (N1.getOpcode() == ISD::INSERT_SUBVECTOR && N2 == N1.getOperand(2) &&
8296 VT == N1.getOperand(1).getValueType())
8297 return N1.getOperand(1);
8298 break;
8299 }
8300 }
8301
8302 if (N1.getOpcode() == ISD::POISON || N2.getOpcode() == ISD::POISON) {
8303 switch (Opcode) {
8304 case ISD::XOR:
8305 case ISD::ADD:
8306 case ISD::PTRADD:
8307 case ISD::SUB:
8309 case ISD::UDIV:
8310 case ISD::SDIV:
8311 case ISD::UREM:
8312 case ISD::SREM:
8313 case ISD::MUL:
8314 case ISD::AND:
8315 case ISD::SSUBSAT:
8316 case ISD::USUBSAT:
8317 case ISD::UMIN:
8318 case ISD::OR:
8319 case ISD::SADDSAT:
8320 case ISD::UADDSAT:
8321 case ISD::UMAX:
8322 case ISD::SMAX:
8323 case ISD::SMIN:
8324 // fold op(arg1, poison) -> poison, fold op(poison, arg2) -> poison.
8325 return N2.getOpcode() == ISD::POISON ? N2 : N1;
8326 }
8327 }
8328
8329 // Canonicalize an UNDEF to the RHS, even over a constant.
8330 if (N1.getOpcode() == ISD::UNDEF && N2.getOpcode() != ISD::UNDEF) {
8331 if (TLI->isCommutativeBinOp(Opcode)) {
8332 std::swap(N1, N2);
8333 } else {
8334 switch (Opcode) {
8335 case ISD::PTRADD:
8336 case ISD::SUB:
8337 // fold op(undef, non_undef_arg2) -> undef.
8338 return N1;
8340 case ISD::UDIV:
8341 case ISD::SDIV:
8342 case ISD::UREM:
8343 case ISD::SREM:
8344 case ISD::SSUBSAT:
8345 case ISD::USUBSAT:
8346 // fold op(undef, non_undef_arg2) -> 0.
8347 return getConstant(0, DL, VT);
8348 }
8349 }
8350 }
8351
8352 // Fold a bunch of operators when the RHS is undef.
8353 if (N2.getOpcode() == ISD::UNDEF) {
8354 switch (Opcode) {
8355 case ISD::XOR:
8356 if (N1.getOpcode() == ISD::UNDEF)
8357 // Handle undef ^ undef -> 0 special case. This is a common
8358 // idiom (misuse).
8359 return getConstant(0, DL, VT);
8360 [[fallthrough]];
8361 case ISD::ADD:
8362 case ISD::PTRADD:
8363 case ISD::SUB:
8364 // fold op(arg1, undef) -> undef.
8365 return N2;
8366 case ISD::UDIV:
8367 case ISD::SDIV:
8368 case ISD::UREM:
8369 case ISD::SREM:
8370 // fold op(arg1, undef) -> poison.
8371 return getPOISON(VT);
8372 case ISD::MUL:
8373 case ISD::AND:
8374 case ISD::SSUBSAT:
8375 case ISD::USUBSAT:
8376 case ISD::UMIN:
8377 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> 0.
8378 return N1.getOpcode() == ISD::UNDEF ? N2 : getConstant(0, DL, VT);
8379 case ISD::OR:
8380 case ISD::SADDSAT:
8381 case ISD::UADDSAT:
8382 case ISD::UMAX:
8383 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> -1.
8384 return N1.getOpcode() == ISD::UNDEF ? N2 : getAllOnesConstant(DL, VT);
8385 case ISD::SMAX:
8386 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> MAX_INT.
8387 return N1.getOpcode() == ISD::UNDEF
8388 ? N2
8389 : getConstant(
8391 VT);
8392 case ISD::SMIN:
8393 // fold op(undef, undef) -> undef, fold op(arg1, undef) -> MIN_INT.
8394 return N1.getOpcode() == ISD::UNDEF
8395 ? N2
8396 : getConstant(
8398 VT);
8399 }
8400 }
8401
8402 // Perform trivial constant folding.
8403 if (SDValue SV = FoldConstantArithmetic(Opcode, DL, VT, {N1, N2}, Flags))
8404 return SV;
8405
8406 // Memoize this node if possible.
8407 SDNode *N;
8408 SDVTList VTs = getVTList(VT);
8409 SDValue Ops[] = {N1, N2};
8410 if (VT != MVT::Glue) {
8412 AddNodeIDNode(ID, Opcode, VTs, Ops);
8413 void *IP = nullptr;
8414 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
8415 E->intersectFlagsWith(Flags);
8416 return SDValue(E, 0);
8417 }
8418
8419 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8420 N->setFlags(Flags);
8421 createOperands(N, Ops);
8422 CSEMap.InsertNode(N, IP);
8423 } else {
8424 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8425 createOperands(N, Ops);
8426 }
8427
8428 InsertNode(N);
8429 SDValue V = SDValue(N, 0);
8430 NewSDValueDbgMsg(V, "Creating new node: ", this);
8431 return V;
8432}
8433
8434SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8435 SDValue N1, SDValue N2, SDValue N3) {
8436 SDNodeFlags Flags;
8437 if (Inserter)
8438 Flags = Inserter->getFlags();
8439 return getNode(Opcode, DL, VT, N1, N2, N3, Flags);
8440}
8441
8442SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8443 SDValue N1, SDValue N2, SDValue N3,
8444 const SDNodeFlags Flags) {
8446 N2.getOpcode() != ISD::DELETED_NODE &&
8447 N3.getOpcode() != ISD::DELETED_NODE &&
8448 "Operand is DELETED_NODE!");
8449 // Perform various simplifications.
8450 switch (Opcode) {
8451 case ISD::BUILD_VECTOR: {
8452 // Attempt to simplify BUILD_VECTOR.
8453 SDValue Ops[] = {N1, N2, N3};
8454 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
8455 return V;
8456 break;
8457 }
8458 case ISD::CONCAT_VECTORS: {
8459 SDValue Ops[] = {N1, N2, N3};
8460 if (SDValue V = foldCONCAT_VECTORS(DL, VT, Ops, *this))
8461 return V;
8462 break;
8463 }
8464 case ISD::SETCC: {
8465 assert(VT.isInteger() && "SETCC result type must be an integer!");
8466 assert(N1.getValueType() == N2.getValueType() &&
8467 "SETCC operands must have the same type!");
8468 assert(VT.isVector() == N1.getValueType().isVector() &&
8469 "SETCC type should be vector iff the operand type is vector!");
8470 assert((!VT.isVector() || VT.getVectorElementCount() ==
8472 "SETCC vector element counts must match!");
8473 // Use FoldSetCC to simplify SETCC's.
8474 if (SDValue V =
8475 FoldSetCC(VT, N1, N2, cast<CondCodeSDNode>(N3)->get(), DL, Flags))
8476 return V;
8477 break;
8478 }
8479 case ISD::SELECT:
8480 case ISD::VSELECT:
8481 if (SDValue V = simplifySelect(N1, N2, N3))
8482 return V;
8483 break;
8485 llvm_unreachable("should use getVectorShuffle constructor!");
8487 if (isNullConstant(N3))
8488 return N1;
8489 break;
8491 if (isNullConstant(N3))
8492 return N2;
8493 break;
8495 assert(VT.isVector() && VT == N1.getValueType() &&
8496 "INSERT_VECTOR_ELT vector type mismatch");
8498 "INSERT_VECTOR_ELT scalar fp/int mismatch");
8499 assert((!VT.isFloatingPoint() ||
8500 VT.getVectorElementType() == N2.getValueType()) &&
8501 "INSERT_VECTOR_ELT fp scalar type mismatch");
8502 assert((!VT.isInteger() ||
8504 "INSERT_VECTOR_ELT int scalar size mismatch");
8505
8506 auto *N3C = dyn_cast<ConstantSDNode>(N3);
8507 // INSERT_VECTOR_ELT into out-of-bounds element is an UNDEF, except
8508 // for scalable vectors where we will generate appropriate code to
8509 // deal with out-of-bounds cases correctly.
8510 if (N3C && VT.isFixedLengthVector() &&
8511 N3C->getZExtValue() >= VT.getVectorNumElements())
8512 return getUNDEF(VT);
8513
8514 // Undefined index can be assumed out-of-bounds, so that's UNDEF too.
8515 if (N3.isUndef())
8516 return getUNDEF(VT);
8517
8518 // If inserting poison, just use the input vector.
8519 if (N2.getOpcode() == ISD::POISON)
8520 return N1;
8521
8522 // Inserting undef into undef/poison is still undef.
8523 if (N2.getOpcode() == ISD::UNDEF && N1.isUndef())
8524 return getUNDEF(VT);
8525
8526 // If the inserted element is an UNDEF, just use the input vector.
8527 // But not if skipping the insert could make the result more poisonous.
8528 if (N2.isUndef()) {
8529 if (N3C && VT.isFixedLengthVector()) {
8530 APInt EltMask =
8531 APInt::getOneBitSet(VT.getVectorNumElements(), N3C->getZExtValue());
8532 if (isGuaranteedNotToBePoison(N1, EltMask))
8533 return N1;
8534 } else if (isGuaranteedNotToBePoison(N1))
8535 return N1;
8536 }
8537 break;
8538 }
8539 case ISD::INSERT_SUBVECTOR: {
8540 // If inserting poison, just use the input vector,
8541 if (N2.getOpcode() == ISD::POISON)
8542 return N1;
8543
8544 // Inserting undef into undef/poison is still undef.
8545 if (N2.getOpcode() == ISD::UNDEF && N1.isUndef())
8546 return getUNDEF(VT);
8547
8548 EVT N2VT = N2.getValueType();
8549 assert(VT == N1.getValueType() &&
8550 "Dest and insert subvector source types must match!");
8551 assert(VT.isVector() && N2VT.isVector() &&
8552 "Insert subvector VTs must be vectors!");
8554 "Insert subvector VTs must have the same element type!");
8555 assert((VT.isScalableVector() || N2VT.isFixedLengthVector()) &&
8556 "Cannot insert a scalable vector into a fixed length vector!");
8557 assert((VT.isScalableVector() != N2VT.isScalableVector() ||
8559 "Insert subvector must be from smaller vector to larger vector!");
8561 "Insert subvector index must be constant");
8562 assert((VT.isScalableVector() != N2VT.isScalableVector() ||
8563 (N2VT.getVectorMinNumElements() + N3->getAsZExtVal()) <=
8565 "Insert subvector overflow!");
8567 TLI->getVectorIdxWidth(getDataLayout()) &&
8568 "Constant index for INSERT_SUBVECTOR has an invalid size");
8569
8570 // Trivial insertion.
8571 if (VT == N2VT)
8572 return N2;
8573
8574 // If this is an insert of an extracted vector into an undef/poison vector,
8575 // we can just use the input to the extract. But not if skipping the
8576 // extract+insert could make the result more poisonous.
8577 if (N1.isUndef() && N2.getOpcode() == ISD::EXTRACT_SUBVECTOR &&
8578 N2.getOperand(1) == N3 && N2.getOperand(0).getValueType() == VT) {
8579 if (N1.getOpcode() == ISD::POISON)
8580 return N2.getOperand(0);
8581 if (VT.isFixedLengthVector() && N2VT.isFixedLengthVector()) {
8582 unsigned LoBit = N3->getAsZExtVal();
8583 unsigned HiBit = LoBit + N2VT.getVectorNumElements();
8584 APInt EltMask =
8585 APInt::getBitsSet(VT.getVectorNumElements(), LoBit, HiBit);
8586 if (isGuaranteedNotToBePoison(N2.getOperand(0), ~EltMask))
8587 return N2.getOperand(0);
8588 } else if (isGuaranteedNotToBePoison(N2.getOperand(0)))
8589 return N2.getOperand(0);
8590 }
8591
8592 // If the inserted subvector is UNDEF, just use the input vector.
8593 // But not if skipping the insert could make the result more poisonous.
8594 if (N2.isUndef()) {
8595 if (VT.isFixedLengthVector()) {
8596 unsigned LoBit = N3->getAsZExtVal();
8597 unsigned HiBit = LoBit + N2VT.getVectorNumElements();
8598 APInt EltMask =
8599 APInt::getBitsSet(VT.getVectorNumElements(), LoBit, HiBit);
8600 if (isGuaranteedNotToBePoison(N1, EltMask))
8601 return N1;
8602 } else if (isGuaranteedNotToBePoison(N1))
8603 return N1;
8604 }
8605 break;
8606 }
8607 case ISD::BITCAST:
8608 // Fold bit_convert nodes from a type to themselves.
8609 if (N1.getValueType() == VT)
8610 return N1;
8611 break;
8612 case ISD::VP_TRUNCATE:
8613 case ISD::VP_SIGN_EXTEND:
8614 case ISD::VP_ZERO_EXTEND:
8615 // Don't create noop casts.
8616 if (N1.getValueType() == VT)
8617 return N1;
8618 break;
8619 case ISD::VECTOR_COMPRESS: {
8620 [[maybe_unused]] EVT VecVT = N1.getValueType();
8621 [[maybe_unused]] EVT MaskVT = N2.getValueType();
8622 [[maybe_unused]] EVT PassthruVT = N3.getValueType();
8623 assert(VT == VecVT && "Vector and result type don't match.");
8624 assert(VecVT.isVector() && MaskVT.isVector() && PassthruVT.isVector() &&
8625 "All inputs must be vectors.");
8626 assert(VecVT == PassthruVT && "Vector and passthru types don't match.");
8628 "Vector and mask must have same number of elements.");
8629
8630 if (N1.isUndef() || N2.isUndef())
8631 return N3;
8632
8633 break;
8634 }
8639 [[maybe_unused]] EVT AccVT = N1.getValueType();
8640 [[maybe_unused]] EVT Input1VT = N2.getValueType();
8641 [[maybe_unused]] EVT Input2VT = N3.getValueType();
8642 assert(Input1VT.isVector() && Input1VT == Input2VT &&
8643 "Expected the second and third operands of the PARTIAL_REDUCE_MLA "
8644 "node to have the same type!");
8645 assert(VT.isVector() && VT == AccVT &&
8646 "Expected the first operand of the PARTIAL_REDUCE_MLA node to have "
8647 "the same type as its result!");
8649 AccVT.getVectorElementCount()) &&
8650 "Expected the element count of the second and third operands of the "
8651 "PARTIAL_REDUCE_MLA node to be a positive integer multiple of the "
8652 "element count of the first operand and the result!");
8654 "Expected the second and third operands of the PARTIAL_REDUCE_MLA "
8655 "node to have an element type which is the same as or smaller than "
8656 "the element type of the first operand and result!");
8657 break;
8658 }
8659 }
8660
8661 // Perform trivial constant folding for arithmetic operators.
8662 switch (Opcode) {
8663 case ISD::FMA:
8664 case ISD::FMAD:
8665 case ISD::SETCC:
8666 case ISD::FSHL:
8667 case ISD::FSHR:
8668 if (SDValue SV =
8669 FoldConstantArithmetic(Opcode, DL, VT, {N1, N2, N3}, Flags))
8670 return SV;
8671 break;
8672 }
8673
8674 // Memoize node if it doesn't produce a glue result.
8675 SDNode *N;
8676 SDVTList VTs = getVTList(VT);
8677 SDValue Ops[] = {N1, N2, N3};
8678 if (VT != MVT::Glue) {
8680 AddNodeIDNode(ID, Opcode, VTs, Ops);
8681 void *IP = nullptr;
8682 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
8683 E->intersectFlagsWith(Flags);
8684 return SDValue(E, 0);
8685 }
8686
8687 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8688 N->setFlags(Flags);
8689 createOperands(N, Ops);
8690 CSEMap.InsertNode(N, IP);
8691 } else {
8692 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
8693 createOperands(N, Ops);
8694 }
8695
8696 InsertNode(N);
8697 SDValue V = SDValue(N, 0);
8698 NewSDValueDbgMsg(V, "Creating new node: ", this);
8699 return V;
8700}
8701
8702SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8703 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
8704 const SDNodeFlags Flags) {
8705 SDValue Ops[] = { N1, N2, N3, N4 };
8706 return getNode(Opcode, DL, VT, Ops, Flags);
8707}
8708
8709SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8710 SDValue N1, SDValue N2, SDValue N3, SDValue N4) {
8711 SDNodeFlags Flags;
8712 if (Inserter)
8713 Flags = Inserter->getFlags();
8714 return getNode(Opcode, DL, VT, N1, N2, N3, N4, Flags);
8715}
8716
8717SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8718 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
8719 SDValue N5, const SDNodeFlags Flags) {
8720 SDValue Ops[] = { N1, N2, N3, N4, N5 };
8721 return getNode(Opcode, DL, VT, Ops, Flags);
8722}
8723
8724SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
8725 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
8726 SDValue N5) {
8727 SDNodeFlags Flags;
8728 if (Inserter)
8729 Flags = Inserter->getFlags();
8730 return getNode(Opcode, DL, VT, N1, N2, N3, N4, N5, Flags);
8731}
8732
8733/// getStackArgumentTokenFactor - Compute a TokenFactor to force all
8734/// the incoming stack arguments to be loaded from the stack.
8736 SmallVector<SDValue, 8> ArgChains;
8737
8738 // Include the original chain at the beginning of the list. When this is
8739 // used by target LowerCall hooks, this helps legalize find the
8740 // CALLSEQ_BEGIN node.
8741 ArgChains.push_back(Chain);
8742
8743 // Add a chain value for each stack argument.
8744 for (SDNode *U : getEntryNode().getNode()->users())
8745 if (LoadSDNode *L = dyn_cast<LoadSDNode>(U))
8746 if (FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(L->getBasePtr()))
8747 if (FI->getIndex() < 0)
8748 ArgChains.push_back(SDValue(L, 1));
8749
8750 // Build a tokenfactor for all the chains.
8751 return getNode(ISD::TokenFactor, SDLoc(Chain), MVT::Other, ArgChains);
8752}
8753
8754/// getMemsetValue - Vectorized representation of the memset value
8755/// operand.
8757 const SDLoc &dl) {
8758 assert(!Value.isUndef());
8759
8760 unsigned NumBits = VT.getScalarSizeInBits();
8762 assert(C->getAPIntValue().getBitWidth() == 8);
8763 APInt Val = APInt::getSplat(NumBits, C->getAPIntValue());
8764 if (VT.isInteger()) {
8765 bool IsOpaque = VT.getSizeInBits() > 64 ||
8766 !DAG.getTargetLoweringInfo().isLegalStoreImmediate(C->getSExtValue());
8767 return DAG.getConstant(Val, dl, VT, false, IsOpaque);
8768 }
8769 return DAG.getConstantFP(APFloat(VT.getFltSemantics(), Val), dl, VT);
8770 }
8771
8772 assert(Value.getValueType() == MVT::i8 && "memset with non-byte fill value?");
8773 EVT IntVT = VT.getScalarType();
8774 if (!IntVT.isInteger())
8775 IntVT = EVT::getIntegerVT(*DAG.getContext(), IntVT.getSizeInBits());
8776
8777 Value = DAG.getNode(ISD::ZERO_EXTEND, dl, IntVT, Value);
8778 if (NumBits > 8) {
8779 // Use a multiplication with 0x010101... to extend the input to the
8780 // required length.
8781 APInt Magic = APInt::getSplat(NumBits, APInt(8, 0x01));
8782 Value = DAG.getNode(ISD::MUL, dl, IntVT, Value,
8783 DAG.getConstant(Magic, dl, IntVT));
8784 }
8785
8786 if (VT != Value.getValueType() && !VT.isInteger())
8787 Value = DAG.getBitcast(VT.getScalarType(), Value);
8788 if (VT != Value.getValueType())
8789 Value = DAG.getSplatBuildVector(VT, dl, Value);
8790
8791 return Value;
8792}
8793
8794/// getMemsetStringVal - Similar to getMemsetValue. Except this is only
8795/// used when a memcpy is turned into a memset when the source is a constant
8796/// string ptr.
8798 const TargetLowering &TLI,
8799 const ConstantDataArraySlice &Slice) {
8800 // Handle vector with all elements zero.
8801 if (Slice.Array == nullptr) {
8802 if (VT.isInteger())
8803 return DAG.getConstant(0, dl, VT);
8804 return DAG.getNode(ISD::BITCAST, dl, VT,
8805 DAG.getConstant(0, dl, VT.changeTypeToInteger()));
8806 }
8807
8808 assert(!VT.isVector() && "Can't handle vector type here!");
8809 unsigned NumVTBits = VT.getSizeInBits();
8810 unsigned NumVTBytes = NumVTBits / 8;
8811 unsigned NumBytes = std::min(NumVTBytes, unsigned(Slice.Length));
8812
8813 APInt Val(NumVTBits, 0);
8814 if (DAG.getDataLayout().isLittleEndian()) {
8815 for (unsigned i = 0; i != NumBytes; ++i)
8816 Val |= (uint64_t)(unsigned char)Slice[i] << i*8;
8817 } else {
8818 for (unsigned i = 0; i != NumBytes; ++i)
8819 Val |= (uint64_t)(unsigned char)Slice[i] << (NumVTBytes-i-1)*8;
8820 }
8821
8822 // If the "cost" of materializing the integer immediate is less than the cost
8823 // of a load, then it is cost effective to turn the load into the immediate.
8824 Type *Ty = VT.getTypeForEVT(*DAG.getContext());
8825 if (TLI.shouldConvertConstantLoadToIntImm(Val, Ty))
8826 return DAG.getConstant(Val, dl, VT);
8827 return SDValue();
8828}
8829
8831 const SDLoc &DL,
8832 const SDNodeFlags Flags) {
8833 SDValue Index = getTypeSize(DL, Base.getValueType(), Offset);
8834 return getMemBasePlusOffset(Base, Index, DL, Flags);
8835}
8836
8838 const SDLoc &DL,
8839 const SDNodeFlags Flags) {
8840 assert(Offset.getValueType().isInteger());
8841 EVT BasePtrVT = Ptr.getValueType();
8842 if (TLI->shouldPreservePtrArith(this->getMachineFunction().getFunction(),
8843 BasePtrVT))
8844 return getNode(ISD::PTRADD, DL, BasePtrVT, Ptr, Offset, Flags);
8845 // InBounds only applies to PTRADD, don't set it if we generate ADD.
8846 SDNodeFlags AddFlags = Flags;
8847 AddFlags.setInBounds(false);
8848 return getNode(ISD::ADD, DL, BasePtrVT, Ptr, Offset, AddFlags);
8849}
8850
8851/// Returns true if memcpy source is constant data.
8853 uint64_t SrcDelta = 0;
8854 GlobalAddressSDNode *G = nullptr;
8855 if (Src.getOpcode() == ISD::GlobalAddress)
8857 else if (Src->isAnyAdd() &&
8858 Src.getOperand(0).getOpcode() == ISD::GlobalAddress &&
8859 Src.getOperand(1).getOpcode() == ISD::Constant) {
8860 G = cast<GlobalAddressSDNode>(Src.getOperand(0));
8861 SrcDelta = Src.getConstantOperandVal(1);
8862 }
8863 if (!G)
8864 return false;
8865
8866 return getConstantDataArrayInfo(G->getGlobal(), Slice, 8,
8867 SrcDelta + G->getOffset());
8868}
8869
8871 SelectionDAG &DAG) {
8872 // On Darwin, -Os means optimize for size without hurting performance, so
8873 // only really optimize for size when -Oz (MinSize) is used.
8875 return MF.getFunction().hasMinSize();
8876 return DAG.shouldOptForSize();
8877}
8878
8880 SmallVector<SDValue, 32> &OutChains, unsigned From,
8881 unsigned To, SmallVector<SDValue, 16> &OutLoadChains,
8882 SmallVector<SDValue, 16> &OutStoreChains) {
8883 assert(OutLoadChains.size() && "Missing loads in memcpy inlining");
8884 assert(OutStoreChains.size() && "Missing stores in memcpy inlining");
8885 SmallVector<SDValue, 16> GluedLoadChains;
8886 for (unsigned i = From; i < To; ++i) {
8887 OutChains.push_back(OutLoadChains[i]);
8888 GluedLoadChains.push_back(OutLoadChains[i]);
8889 }
8890
8891 // Chain for all loads.
8892 SDValue LoadToken = DAG.getNode(ISD::TokenFactor, dl, MVT::Other,
8893 GluedLoadChains);
8894
8895 for (unsigned i = From; i < To; ++i) {
8896 StoreSDNode *ST = dyn_cast<StoreSDNode>(OutStoreChains[i]);
8897 SDValue NewStore = DAG.getTruncStore(LoadToken, dl, ST->getValue(),
8898 ST->getBasePtr(), ST->getMemoryVT(),
8899 ST->getMemOperand());
8900 OutChains.push_back(NewStore);
8901 }
8902}
8903
8905 SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src,
8906 uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline,
8907 MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo,
8908 const AAMDNodes &AAInfo, BatchAAResults *BatchAA) {
8909 // Turn a memcpy of undef to nop.
8910 // FIXME: We need to honor volatile even is Src is undef.
8911 if (Src.isUndef())
8912 return Chain;
8913
8914 // Expand memcpy to a series of load and store ops if the size operand falls
8915 // below a certain threshold.
8916 // TODO: In the AlwaysInline case, if the size is big then generate a loop
8917 // rather than maybe a humongous number of loads and stores.
8918 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
8919 const DataLayout &DL = DAG.getDataLayout();
8920 LLVMContext &C = *DAG.getContext();
8921 std::vector<EVT> MemOps;
8922 bool DstAlignCanChange = false;
8924 MachineFrameInfo &MFI = MF.getFrameInfo();
8925 bool OptSize = shouldLowerMemFuncForSize(MF, DAG);
8927 if (FI && !MFI.isFixedObjectIndex(FI->getIndex()))
8928 DstAlignCanChange = true;
8929 MaybeAlign SrcAlign = DAG.InferPtrAlign(Src);
8930 if (!SrcAlign || Alignment > *SrcAlign)
8931 SrcAlign = Alignment;
8932 assert(SrcAlign && "SrcAlign must be set");
8934 // If marked as volatile, perform a copy even when marked as constant.
8935 bool CopyFromConstant = !isVol && isMemSrcFromConstant(Src, Slice);
8936 bool isZeroConstant = CopyFromConstant && Slice.Array == nullptr;
8937 unsigned Limit = AlwaysInline ? ~0U : TLI.getMaxStoresPerMemcpy(OptSize);
8938 const MemOp Op = isZeroConstant
8939 ? MemOp::Set(Size, DstAlignCanChange, Alignment,
8940 /*IsZeroMemset*/ true, isVol)
8941 : MemOp::Copy(Size, DstAlignCanChange, Alignment,
8942 *SrcAlign, isVol, CopyFromConstant);
8943 if (!TLI.findOptimalMemOpLowering(
8944 C, MemOps, Limit, Op, DstPtrInfo.getAddrSpace(),
8945 SrcPtrInfo.getAddrSpace(), MF.getFunction().getAttributes(), nullptr))
8946 return SDValue();
8947
8948 if (DstAlignCanChange) {
8949 Type *Ty = MemOps[0].getTypeForEVT(C);
8950 Align NewAlign = DL.getABITypeAlign(Ty);
8951
8952 // Don't promote to an alignment that would require dynamic stack
8953 // realignment which may conflict with optimizations such as tail call
8954 // optimization.
8956 if (!TRI->hasStackRealignment(MF))
8957 if (MaybeAlign StackAlign = DL.getStackAlignment())
8958 NewAlign = std::min(NewAlign, *StackAlign);
8959
8960 if (NewAlign > Alignment) {
8961 // Give the stack frame object a larger alignment if needed.
8962 if (MFI.getObjectAlign(FI->getIndex()) < NewAlign)
8963 MFI.setObjectAlignment(FI->getIndex(), NewAlign);
8964 Alignment = NewAlign;
8965 }
8966 }
8967
8968 // Prepare AAInfo for loads/stores after lowering this memcpy.
8969 AAMDNodes NewAAInfo = AAInfo;
8970 NewAAInfo.TBAA = NewAAInfo.TBAAStruct = nullptr;
8971
8972 const Value *SrcVal = dyn_cast_if_present<const Value *>(SrcPtrInfo.V);
8973 bool isConstant =
8974 BatchAA && SrcVal &&
8975 BatchAA->pointsToConstantMemory(MemoryLocation(SrcVal, Size, AAInfo));
8976
8977 MachineMemOperand::Flags MMOFlags =
8979 SmallVector<SDValue, 16> OutLoadChains;
8980 SmallVector<SDValue, 16> OutStoreChains;
8981 SmallVector<SDValue, 32> OutChains;
8982 unsigned NumMemOps = MemOps.size();
8983 uint64_t SrcOff = 0, DstOff = 0;
8984 for (unsigned i = 0; i != NumMemOps; ++i) {
8985 EVT VT = MemOps[i];
8986 unsigned VTSize = VT.getSizeInBits() / 8;
8987 SDValue Value, Store;
8988
8989 if (VTSize > Size) {
8990 // Issuing an unaligned load / store pair that overlaps with the previous
8991 // pair. Adjust the offset accordingly.
8992 assert(i == NumMemOps-1 && i != 0);
8993 SrcOff -= VTSize - Size;
8994 DstOff -= VTSize - Size;
8995 }
8996
8997 if (CopyFromConstant &&
8998 (isZeroConstant || (VT.isInteger() && !VT.isVector()))) {
8999 // It's unlikely a store of a vector immediate can be done in a single
9000 // instruction. It would require a load from a constantpool first.
9001 // We only handle zero vectors here.
9002 // FIXME: Handle other cases where store of vector immediate is done in
9003 // a single instruction.
9004 ConstantDataArraySlice SubSlice;
9005 if (SrcOff < Slice.Length) {
9006 SubSlice = Slice;
9007 SubSlice.move(SrcOff);
9008 } else {
9009 // This is an out-of-bounds access and hence UB. Pretend we read zero.
9010 SubSlice.Array = nullptr;
9011 SubSlice.Offset = 0;
9012 SubSlice.Length = VTSize;
9013 }
9014 Value = getMemsetStringVal(VT, dl, DAG, TLI, SubSlice);
9015 if (Value.getNode()) {
9016 Store = DAG.getStore(
9017 Chain, dl, Value,
9018 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9019 DstPtrInfo.getWithOffset(DstOff), Alignment, MMOFlags, NewAAInfo);
9020 OutChains.push_back(Store);
9021 }
9022 }
9023
9024 if (!Store.getNode()) {
9025 // The type might not be legal for the target. This should only happen
9026 // if the type is smaller than a legal type, as on PPC, so the right
9027 // thing to do is generate a LoadExt/StoreTrunc pair. These simplify
9028 // to Load/Store if NVT==VT.
9029 // FIXME does the case above also need this?
9030 EVT NVT = TLI.getTypeToTransformTo(C, VT);
9031 assert(NVT.bitsGE(VT));
9032
9033 bool isDereferenceable =
9034 SrcPtrInfo.getWithOffset(SrcOff).isDereferenceable(VTSize, C, DL);
9035 MachineMemOperand::Flags SrcMMOFlags = MMOFlags;
9036 if (isDereferenceable)
9038 if (isConstant)
9039 SrcMMOFlags |= MachineMemOperand::MOInvariant;
9040
9041 Value = DAG.getExtLoad(
9042 ISD::EXTLOAD, dl, NVT, Chain,
9043 DAG.getObjectPtrOffset(dl, Src, TypeSize::getFixed(SrcOff)),
9044 SrcPtrInfo.getWithOffset(SrcOff), VT,
9045 commonAlignment(*SrcAlign, SrcOff), SrcMMOFlags, NewAAInfo);
9046 OutLoadChains.push_back(Value.getValue(1));
9047
9048 Store = DAG.getTruncStore(
9049 Chain, dl, Value,
9050 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9051 DstPtrInfo.getWithOffset(DstOff), VT, Alignment, MMOFlags, NewAAInfo);
9052 OutStoreChains.push_back(Store);
9053 }
9054 SrcOff += VTSize;
9055 DstOff += VTSize;
9056 Size -= VTSize;
9057 }
9058
9059 unsigned GluedLdStLimit = MaxLdStGlue == 0 ?
9061 unsigned NumLdStInMemcpy = OutStoreChains.size();
9062
9063 if (NumLdStInMemcpy) {
9064 // It may be that memcpy might be converted to memset if it's memcpy
9065 // of constants. In such a case, we won't have loads and stores, but
9066 // just stores. In the absence of loads, there is nothing to gang up.
9067 if ((GluedLdStLimit <= 1) || !EnableMemCpyDAGOpt) {
9068 // If target does not care, just leave as it.
9069 for (unsigned i = 0; i < NumLdStInMemcpy; ++i) {
9070 OutChains.push_back(OutLoadChains[i]);
9071 OutChains.push_back(OutStoreChains[i]);
9072 }
9073 } else {
9074 // Ld/St less than/equal limit set by target.
9075 if (NumLdStInMemcpy <= GluedLdStLimit) {
9076 chainLoadsAndStoresForMemcpy(DAG, dl, OutChains, 0,
9077 NumLdStInMemcpy, OutLoadChains,
9078 OutStoreChains);
9079 } else {
9080 unsigned NumberLdChain = NumLdStInMemcpy / GluedLdStLimit;
9081 unsigned RemainingLdStInMemcpy = NumLdStInMemcpy % GluedLdStLimit;
9082 unsigned GlueIter = 0;
9083
9084 // Residual ld/st.
9085 if (RemainingLdStInMemcpy) {
9087 DAG, dl, OutChains, NumLdStInMemcpy - RemainingLdStInMemcpy,
9088 NumLdStInMemcpy, OutLoadChains, OutStoreChains);
9089 }
9090
9091 for (unsigned cnt = 0; cnt < NumberLdChain; ++cnt) {
9092 unsigned IndexFrom = NumLdStInMemcpy - RemainingLdStInMemcpy -
9093 GlueIter - GluedLdStLimit;
9094 unsigned IndexTo = NumLdStInMemcpy - RemainingLdStInMemcpy - GlueIter;
9095 chainLoadsAndStoresForMemcpy(DAG, dl, OutChains, IndexFrom, IndexTo,
9096 OutLoadChains, OutStoreChains);
9097 GlueIter += GluedLdStLimit;
9098 }
9099 }
9100 }
9101 }
9102 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OutChains);
9103}
9104
9106 SDValue Chain, SDValue Dst, SDValue Src,
9107 uint64_t Size, Align Alignment,
9108 bool isVol, bool AlwaysInline,
9109 MachinePointerInfo DstPtrInfo,
9110 MachinePointerInfo SrcPtrInfo,
9111 const AAMDNodes &AAInfo) {
9112 // Turn a memmove of undef to nop.
9113 // FIXME: We need to honor volatile even is Src is undef.
9114 if (Src.isUndef())
9115 return Chain;
9116
9117 // Expand memmove to a series of load and store ops if the size operand falls
9118 // below a certain threshold.
9119 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
9120 const DataLayout &DL = DAG.getDataLayout();
9121 LLVMContext &C = *DAG.getContext();
9122 std::vector<EVT> MemOps;
9123 bool DstAlignCanChange = false;
9125 MachineFrameInfo &MFI = MF.getFrameInfo();
9126 bool OptSize = shouldLowerMemFuncForSize(MF, DAG);
9128 if (FI && !MFI.isFixedObjectIndex(FI->getIndex()))
9129 DstAlignCanChange = true;
9130 MaybeAlign SrcAlign = DAG.InferPtrAlign(Src);
9131 if (!SrcAlign || Alignment > *SrcAlign)
9132 SrcAlign = Alignment;
9133 assert(SrcAlign && "SrcAlign must be set");
9134 unsigned Limit = AlwaysInline ? ~0U : TLI.getMaxStoresPerMemmove(OptSize);
9135 if (!TLI.findOptimalMemOpLowering(
9136 C, MemOps, Limit,
9137 MemOp::Copy(Size, DstAlignCanChange, Alignment, *SrcAlign,
9138 /*IsVolatile*/ true),
9139 DstPtrInfo.getAddrSpace(), SrcPtrInfo.getAddrSpace(),
9140 MF.getFunction().getAttributes(), nullptr))
9141 return SDValue();
9142
9143 if (DstAlignCanChange) {
9144 Type *Ty = MemOps[0].getTypeForEVT(C);
9145 Align NewAlign = DL.getABITypeAlign(Ty);
9146
9147 // Don't promote to an alignment that would require dynamic stack
9148 // realignment which may conflict with optimizations such as tail call
9149 // optimization.
9151 if (!TRI->hasStackRealignment(MF))
9152 if (MaybeAlign StackAlign = DL.getStackAlignment())
9153 NewAlign = std::min(NewAlign, *StackAlign);
9154
9155 if (NewAlign > Alignment) {
9156 // Give the stack frame object a larger alignment if needed.
9157 if (MFI.getObjectAlign(FI->getIndex()) < NewAlign)
9158 MFI.setObjectAlignment(FI->getIndex(), NewAlign);
9159 Alignment = NewAlign;
9160 }
9161 }
9162
9163 // Prepare AAInfo for loads/stores after lowering this memmove.
9164 AAMDNodes NewAAInfo = AAInfo;
9165 NewAAInfo.TBAA = NewAAInfo.TBAAStruct = nullptr;
9166
9167 MachineMemOperand::Flags MMOFlags =
9169 uint64_t SrcOff = 0, DstOff = 0;
9170 SmallVector<SDValue, 8> LoadValues;
9171 SmallVector<SDValue, 8> LoadChains;
9172 SmallVector<SDValue, 8> OutChains;
9173 unsigned NumMemOps = MemOps.size();
9174 for (unsigned i = 0; i < NumMemOps; i++) {
9175 EVT VT = MemOps[i];
9176 unsigned VTSize = VT.getSizeInBits() / 8;
9177 SDValue Value;
9178
9179 bool isDereferenceable =
9180 SrcPtrInfo.getWithOffset(SrcOff).isDereferenceable(VTSize, C, DL);
9181 MachineMemOperand::Flags SrcMMOFlags = MMOFlags;
9182 if (isDereferenceable)
9184
9185 Value = DAG.getLoad(
9186 VT, dl, Chain,
9187 DAG.getObjectPtrOffset(dl, Src, TypeSize::getFixed(SrcOff)),
9188 SrcPtrInfo.getWithOffset(SrcOff), *SrcAlign, SrcMMOFlags, NewAAInfo);
9189 LoadValues.push_back(Value);
9190 LoadChains.push_back(Value.getValue(1));
9191 SrcOff += VTSize;
9192 }
9193 Chain = DAG.getNode(ISD::TokenFactor, dl, MVT::Other, LoadChains);
9194 OutChains.clear();
9195 for (unsigned i = 0; i < NumMemOps; i++) {
9196 EVT VT = MemOps[i];
9197 unsigned VTSize = VT.getSizeInBits() / 8;
9198 SDValue Store;
9199
9200 Store = DAG.getStore(
9201 Chain, dl, LoadValues[i],
9202 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9203 DstPtrInfo.getWithOffset(DstOff), Alignment, MMOFlags, NewAAInfo);
9204 OutChains.push_back(Store);
9205 DstOff += VTSize;
9206 }
9207
9208 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OutChains);
9209}
9210
9211/// Lower the call to 'memset' intrinsic function into a series of store
9212/// operations.
9213///
9214/// \param DAG Selection DAG where lowered code is placed.
9215/// \param dl Link to corresponding IR location.
9216/// \param Chain Control flow dependency.
9217/// \param Dst Pointer to destination memory location.
9218/// \param Src Value of byte to write into the memory.
9219/// \param Size Number of bytes to write.
9220/// \param Alignment Alignment of the destination in bytes.
9221/// \param isVol True if destination is volatile.
9222/// \param AlwaysInline Makes sure no function call is generated.
9223/// \param DstPtrInfo IR information on the memory pointer.
9224/// \returns New head in the control flow, if lowering was successful, empty
9225/// SDValue otherwise.
9226///
9227/// The function tries to replace 'llvm.memset' intrinsic with several store
9228/// operations and value calculation code. This is usually profitable for small
9229/// memory size or when the semantic requires inlining.
9231 SDValue Chain, SDValue Dst, SDValue Src,
9232 uint64_t Size, Align Alignment, bool isVol,
9233 bool AlwaysInline, MachinePointerInfo DstPtrInfo,
9234 const AAMDNodes &AAInfo) {
9235 // Turn a memset of undef to nop.
9236 // FIXME: We need to honor volatile even is Src is undef.
9237 if (Src.isUndef())
9238 return Chain;
9239
9240 // Expand memset to a series of load/store ops if the size operand
9241 // falls below a certain threshold.
9242 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
9243 std::vector<EVT> MemOps;
9244 bool DstAlignCanChange = false;
9245 LLVMContext &C = *DAG.getContext();
9247 MachineFrameInfo &MFI = MF.getFrameInfo();
9248 bool OptSize = shouldLowerMemFuncForSize(MF, DAG);
9250 if (FI && !MFI.isFixedObjectIndex(FI->getIndex()))
9251 DstAlignCanChange = true;
9252 bool IsZeroVal = isNullConstant(Src);
9253 unsigned Limit = AlwaysInline ? ~0 : TLI.getMaxStoresPerMemset(OptSize);
9254
9255 EVT LargestVT;
9256 if (!TLI.findOptimalMemOpLowering(
9257 C, MemOps, Limit,
9258 MemOp::Set(Size, DstAlignCanChange, Alignment, IsZeroVal, isVol),
9259 DstPtrInfo.getAddrSpace(), ~0u, MF.getFunction().getAttributes(),
9260 &LargestVT))
9261 return SDValue();
9262
9263 if (DstAlignCanChange) {
9264 Type *Ty = MemOps[0].getTypeForEVT(*DAG.getContext());
9265 const DataLayout &DL = DAG.getDataLayout();
9266 Align NewAlign = DL.getABITypeAlign(Ty);
9267
9268 // Don't promote to an alignment that would require dynamic stack
9269 // realignment which may conflict with optimizations such as tail call
9270 // optimization.
9272 if (!TRI->hasStackRealignment(MF))
9273 if (MaybeAlign StackAlign = DL.getStackAlignment())
9274 NewAlign = std::min(NewAlign, *StackAlign);
9275
9276 if (NewAlign > Alignment) {
9277 // Give the stack frame object a larger alignment if needed.
9278 if (MFI.getObjectAlign(FI->getIndex()) < NewAlign)
9279 MFI.setObjectAlignment(FI->getIndex(), NewAlign);
9280 Alignment = NewAlign;
9281 }
9282 }
9283
9284 SmallVector<SDValue, 8> OutChains;
9285 uint64_t DstOff = 0;
9286 unsigned NumMemOps = MemOps.size();
9287
9288 // Find the largest store and generate the bit pattern for it.
9289 // If target didn't set LargestVT, compute it from MemOps.
9290 if (!LargestVT.isSimple()) {
9291 LargestVT = MemOps[0];
9292 for (unsigned i = 1; i < NumMemOps; i++)
9293 if (MemOps[i].bitsGT(LargestVT))
9294 LargestVT = MemOps[i];
9295 }
9296 SDValue MemSetValue = getMemsetValue(Src, LargestVT, DAG, dl);
9297
9298 // Prepare AAInfo for loads/stores after lowering this memset.
9299 AAMDNodes NewAAInfo = AAInfo;
9300 NewAAInfo.TBAA = NewAAInfo.TBAAStruct = nullptr;
9301
9302 for (unsigned i = 0; i < NumMemOps; i++) {
9303 EVT VT = MemOps[i];
9304 unsigned VTSize = VT.getSizeInBits() / 8;
9305 // The target should specify store types that exactly cover the memset size
9306 // (with the last store potentially being oversized for overlapping stores).
9307 assert(Size > 0 && "Target specified more stores than needed in "
9308 "findOptimalMemOpLowering");
9309 if (VTSize > Size) {
9310 // Issuing an unaligned load / store pair that overlaps with the previous
9311 // pair. Adjust the offset accordingly.
9312 assert(i == NumMemOps-1 && i != 0);
9313 DstOff -= VTSize - Size;
9314 }
9315
9316 // If this store is smaller than the largest store see whether we can get
9317 // the smaller value for free with a truncate or extract vector element and
9318 // then store.
9319 SDValue Value = MemSetValue;
9320 if (VT.bitsLT(LargestVT)) {
9321 unsigned Index;
9322 unsigned NElts = LargestVT.getSizeInBits() / VT.getSizeInBits();
9323 EVT SVT = EVT::getVectorVT(*DAG.getContext(), VT.getScalarType(), NElts);
9324 if (!LargestVT.isVector() && !VT.isVector() &&
9325 TLI.isTruncateFree(LargestVT, VT))
9326 Value = DAG.getNode(ISD::TRUNCATE, dl, VT, MemSetValue);
9327 else if (LargestVT.isVector() && !VT.isVector() &&
9329 LargestVT.getTypeForEVT(*DAG.getContext()),
9330 VT.getSizeInBits(), Index) &&
9331 TLI.isTypeLegal(SVT) &&
9332 LargestVT.getSizeInBits() == SVT.getSizeInBits()) {
9333 // Target which can combine store(extractelement VectorTy, Idx) can get
9334 // the smaller value for free.
9335 SDValue TailValue = DAG.getNode(ISD::BITCAST, dl, SVT, MemSetValue);
9336 Value = DAG.getExtractVectorElt(dl, VT, TailValue, Index);
9337 } else
9338 Value = getMemsetValue(Src, VT, DAG, dl);
9339 }
9340 assert(Value.getValueType() == VT && "Value with wrong type.");
9341 SDValue Store = DAG.getStore(
9342 Chain, dl, Value,
9343 DAG.getObjectPtrOffset(dl, Dst, TypeSize::getFixed(DstOff)),
9344 DstPtrInfo.getWithOffset(DstOff), Alignment,
9346 NewAAInfo);
9347 OutChains.push_back(Store);
9348 DstOff += VT.getSizeInBits() / 8;
9349 // For oversized overlapping stores, only subtract the remaining bytes.
9350 // For normal stores, subtract the full store size.
9351 if (VTSize > Size) {
9352 Size = 0;
9353 } else {
9354 Size -= VTSize;
9355 }
9356 }
9357
9358 // After processing all stores, Size should be exactly 0. Any remaining bytes
9359 // indicate a bug in the target's findOptimalMemOpLowering implementation.
9360 assert(Size == 0 && "Target's findOptimalMemOpLowering did not specify "
9361 "stores that exactly cover the memset size");
9362
9363 return DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OutChains);
9364}
9365
9367 unsigned AS) {
9368 // Lowering memcpy / memset / memmove intrinsics to calls is only valid if all
9369 // pointer operands can be losslessly bitcasted to pointers of address space 0
9370 if (AS != 0 && !TLI->getTargetMachine().isNoopAddrSpaceCast(AS, 0)) {
9371 report_fatal_error("cannot lower memory intrinsic in address space " +
9372 Twine(AS));
9373 }
9374}
9375
9377 const SelectionDAG *SelDAG,
9378 bool AllowReturnsFirstArg) {
9379 if (!CI || !CI->isTailCall())
9380 return false;
9381 // TODO: Fix "returns-first-arg" determination so it doesn't depend on which
9382 // helper symbol we lower to.
9383 return isInTailCallPosition(*CI, SelDAG->getTarget(),
9384 AllowReturnsFirstArg &&
9386}
9387
9388static std::pair<SDValue, SDValue>
9391 const CallInst *CI, RTLIB::Libcall Call,
9392 SelectionDAG *DAG, const TargetLowering *TLI) {
9393 RTLIB::LibcallImpl LCImpl = DAG->getLibcalls().getLibcallImpl(Call);
9394
9395 if (LCImpl == RTLIB::Unsupported)
9396 return {};
9397
9399 bool IsTailCall =
9400 isInTailCallPositionWrapper(CI, DAG, /*AllowReturnsFirstArg=*/true);
9401 SDValue Callee =
9402 DAG->getExternalSymbol(LCImpl, TLI->getPointerTy(DAG->getDataLayout()));
9403
9404 CLI.setDebugLoc(dl)
9405 .setChain(Chain)
9407 CI->getType(), Callee, std::move(Args))
9408 .setTailCall(IsTailCall);
9409
9410 return TLI->LowerCallTo(CLI);
9411}
9412
9413std::pair<SDValue, SDValue> SelectionDAG::getStrcmp(SDValue Chain,
9414 const SDLoc &dl, SDValue S1,
9415 SDValue S2,
9416 const CallInst *CI) {
9418 TargetLowering::ArgListTy Args = {{S1, PT}, {S2, PT}};
9419 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9420 RTLIB::STRCMP, this, TLI);
9421}
9422
9423std::pair<SDValue, SDValue> SelectionDAG::getStrstr(SDValue Chain,
9424 const SDLoc &dl, SDValue S1,
9425 SDValue S2,
9426 const CallInst *CI) {
9428 TargetLowering::ArgListTy Args = {{S1, PT}, {S2, PT}};
9429 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9430 RTLIB::STRSTR, this, TLI);
9431}
9432
9433std::pair<SDValue, SDValue> SelectionDAG::getMemccpy(SDValue Chain,
9434 const SDLoc &dl,
9435 SDValue Dst, SDValue Src,
9437 const CallInst *CI) {
9439
9441 {Dst, PT},
9442 {Src, PT},
9445 return getRuntimeCallSDValueHelper(Chain, dl, std::move(Args), CI,
9446 RTLIB::MEMCCPY, this, TLI);
9447}
9448
9449std::pair<SDValue, SDValue>
9451 SDValue Mem1, SDValue Size, const CallInst *CI) {
9452 RTLIB::LibcallImpl MemcmpImpl = Libcalls->getLibcallImpl(RTLIB::MEMCMP);
9453 if (MemcmpImpl == RTLIB::Unsupported)
9454 return {};
9455
9458 {Mem0, PT},
9459 {Mem1, PT},
9461
9463 bool IsTailCall =
9464 isInTailCallPositionWrapper(CI, this, /*AllowReturnsFirstArg*/ true);
9465
9466 CLI.setDebugLoc(dl)
9467 .setChain(Chain)
9468 .setLibCallee(
9469 Libcalls->getLibcallImplCallingConv(MemcmpImpl),
9471 getExternalSymbol(MemcmpImpl, TLI->getPointerTy(getDataLayout())),
9472 std::move(Args))
9473 .setTailCall(IsTailCall);
9474
9475 return TLI->LowerCallTo(CLI);
9476}
9477
9478std::pair<SDValue, SDValue> SelectionDAG::getStrcpy(SDValue Chain,
9479 const SDLoc &dl,
9480 SDValue Dst, SDValue Src,
9481 const CallInst *CI) {
9482 RTLIB::LibcallImpl LCImpl = Libcalls->getLibcallImpl(RTLIB::STRCPY);
9483 if (LCImpl == RTLIB::Unsupported)
9484 return {};
9485
9487 TargetLowering::ArgListTy Args = {{Dst, PT}, {Src, PT}};
9488
9490 bool IsTailCall =
9491 isInTailCallPositionWrapper(CI, this, /*AllowReturnsFirstArg=*/true);
9492
9493 CLI.setDebugLoc(dl)
9494 .setChain(Chain)
9495 .setLibCallee(
9496 Libcalls->getLibcallImplCallingConv(LCImpl), CI->getType(),
9497 getExternalSymbol(LCImpl, TLI->getPointerTy(getDataLayout())),
9498 std::move(Args))
9499 .setTailCall(IsTailCall);
9500
9501 return TLI->LowerCallTo(CLI);
9502}
9503
9504std::pair<SDValue, SDValue> SelectionDAG::getStrlen(SDValue Chain,
9505 const SDLoc &dl,
9506 SDValue Src,
9507 const CallInst *CI) {
9508 RTLIB::LibcallImpl StrlenImpl = Libcalls->getLibcallImpl(RTLIB::STRLEN);
9509 if (StrlenImpl == RTLIB::Unsupported)
9510 return {};
9511
9512 // Emit a library call.
9515
9517 bool IsTailCall =
9518 isInTailCallPositionWrapper(CI, this, /*AllowReturnsFirstArg*/ true);
9519
9520 CLI.setDebugLoc(dl)
9521 .setChain(Chain)
9522 .setLibCallee(Libcalls->getLibcallImplCallingConv(StrlenImpl),
9523 CI->getType(),
9525 StrlenImpl, TLI->getProgramPointerTy(getDataLayout())),
9526 std::move(Args))
9527 .setTailCall(IsTailCall);
9528
9529 return TLI->LowerCallTo(CLI);
9530}
9531
9533 SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size,
9534 Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI,
9535 std::optional<bool> OverrideTailCall, MachinePointerInfo DstPtrInfo,
9536 MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo,
9537 BatchAAResults *BatchAA) {
9538 // Check to see if we should lower the memcpy to loads and stores first.
9539 // For cases within the target-specified limits, this is the best choice.
9541 if (ConstantSize) {
9542 // Memcpy with size zero? Just return the original chain.
9543 if (ConstantSize->isZero())
9544 return Chain;
9545
9547 *this, dl, Chain, Dst, Src, ConstantSize->getZExtValue(), Alignment,
9548 isVol, false, DstPtrInfo, SrcPtrInfo, AAInfo, BatchAA);
9549 if (Result.getNode())
9550 return Result;
9551 }
9552
9553 // Then check to see if we should lower the memcpy with target-specific
9554 // code. If the target chooses to do this, this is the next best.
9555 if (TSI) {
9556 SDValue Result = TSI->EmitTargetCodeForMemcpy(
9557 *this, dl, Chain, Dst, Src, Size, Alignment, isVol, AlwaysInline,
9558 DstPtrInfo, SrcPtrInfo);
9559 if (Result.getNode())
9560 return Result;
9561 }
9562
9563 // If we really need inline code and the target declined to provide it,
9564 // use a (potentially long) sequence of loads and stores.
9565 if (AlwaysInline) {
9566 assert(ConstantSize && "AlwaysInline requires a constant size!");
9568 *this, dl, Chain, Dst, Src, ConstantSize->getZExtValue(), Alignment,
9569 isVol, true, DstPtrInfo, SrcPtrInfo, AAInfo, BatchAA);
9570 }
9571
9574
9575 // FIXME: If the memcpy is volatile (isVol), lowering it to a plain libc
9576 // memcpy is not guaranteed to be safe. libc memcpys aren't required to
9577 // respect volatile, so they may do things like read or write memory
9578 // beyond the given memory regions. But fixing this isn't easy, and most
9579 // people don't care.
9580
9581 // Emit a library call.
9584 Args.emplace_back(Dst, PtrTy);
9585 Args.emplace_back(Src, PtrTy);
9586 Args.emplace_back(Size, getDataLayout().getIntPtrType(*getContext()));
9587 // FIXME: pass in SDLoc
9589 bool IsTailCall = false;
9590 RTLIB::LibcallImpl MemCpyImpl = TLI->getMemcpyImpl();
9591
9592 if (OverrideTailCall.has_value()) {
9593 IsTailCall = *OverrideTailCall;
9594 } else {
9595 bool LowersToMemcpy = MemCpyImpl == RTLIB::impl_memcpy;
9596 IsTailCall = isInTailCallPositionWrapper(CI, this, LowersToMemcpy);
9597 }
9598
9599 CLI.setDebugLoc(dl)
9600 .setChain(Chain)
9601 .setLibCallee(
9602 Libcalls->getLibcallImplCallingConv(MemCpyImpl),
9603 Dst.getValueType().getTypeForEVT(*getContext()),
9604 getExternalSymbol(MemCpyImpl, TLI->getPointerTy(getDataLayout())),
9605 std::move(Args))
9607 .setTailCall(IsTailCall);
9608
9609 std::pair<SDValue,SDValue> CallResult = TLI->LowerCallTo(CLI);
9610 return CallResult.second;
9611}
9612
9614 SDValue Dst, SDValue Src, SDValue Size,
9615 Type *SizeTy, unsigned ElemSz,
9616 bool isTailCall,
9617 MachinePointerInfo DstPtrInfo,
9618 MachinePointerInfo SrcPtrInfo) {
9619 // Emit a library call.
9622 Args.emplace_back(Dst, ArgTy);
9623 Args.emplace_back(Src, ArgTy);
9624 Args.emplace_back(Size, SizeTy);
9625
9626 RTLIB::Libcall LibraryCall =
9628 RTLIB::LibcallImpl LibcallImpl = Libcalls->getLibcallImpl(LibraryCall);
9629 if (LibcallImpl == RTLIB::Unsupported)
9630 report_fatal_error("Unsupported element size");
9631
9633 CLI.setDebugLoc(dl)
9634 .setChain(Chain)
9635 .setLibCallee(
9636 Libcalls->getLibcallImplCallingConv(LibcallImpl),
9638 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout())),
9639 std::move(Args))
9641 .setTailCall(isTailCall);
9642
9643 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9644 return CallResult.second;
9645}
9646
9648 SDValue Src, SDValue Size, Align Alignment,
9649 bool isVol, const CallInst *CI,
9650 std::optional<bool> OverrideTailCall,
9651 MachinePointerInfo DstPtrInfo,
9652 MachinePointerInfo SrcPtrInfo,
9653 const AAMDNodes &AAInfo,
9654 BatchAAResults *BatchAA) {
9655 // Check to see if we should lower the memmove to loads and stores first.
9656 // For cases within the target-specified limits, this is the best choice.
9658 if (ConstantSize) {
9659 // Memmove with size zero? Just return the original chain.
9660 if (ConstantSize->isZero())
9661 return Chain;
9662
9664 *this, dl, Chain, Dst, Src, ConstantSize->getZExtValue(), Alignment,
9665 isVol, false, DstPtrInfo, SrcPtrInfo, AAInfo);
9666 if (Result.getNode())
9667 return Result;
9668 }
9669
9670 // Then check to see if we should lower the memmove with target-specific
9671 // code. If the target chooses to do this, this is the next best.
9672 if (TSI) {
9673 SDValue Result =
9674 TSI->EmitTargetCodeForMemmove(*this, dl, Chain, Dst, Src, Size,
9675 Alignment, isVol, DstPtrInfo, SrcPtrInfo);
9676 if (Result.getNode())
9677 return Result;
9678 }
9679
9682
9683 // FIXME: If the memmove is volatile, lowering it to plain libc memmove may
9684 // not be safe. See memcpy above for more details.
9685
9686 // Emit a library call.
9689 Args.emplace_back(Dst, PtrTy);
9690 Args.emplace_back(Src, PtrTy);
9691 Args.emplace_back(Size, getDataLayout().getIntPtrType(*getContext()));
9692 // FIXME: pass in SDLoc
9694
9695 RTLIB::LibcallImpl MemmoveImpl = Libcalls->getLibcallImpl(RTLIB::MEMMOVE);
9696
9697 bool IsTailCall = false;
9698 if (OverrideTailCall.has_value()) {
9699 IsTailCall = *OverrideTailCall;
9700 } else {
9701 bool LowersToMemmove = MemmoveImpl == RTLIB::impl_memmove;
9702 IsTailCall = isInTailCallPositionWrapper(CI, this, LowersToMemmove);
9703 }
9704
9705 CLI.setDebugLoc(dl)
9706 .setChain(Chain)
9707 .setLibCallee(
9708 Libcalls->getLibcallImplCallingConv(MemmoveImpl),
9709 Dst.getValueType().getTypeForEVT(*getContext()),
9710 getExternalSymbol(MemmoveImpl, TLI->getPointerTy(getDataLayout())),
9711 std::move(Args))
9713 .setTailCall(IsTailCall);
9714
9715 std::pair<SDValue,SDValue> CallResult = TLI->LowerCallTo(CLI);
9716 return CallResult.second;
9717}
9718
9720 SDValue Dst, SDValue Src, SDValue Size,
9721 Type *SizeTy, unsigned ElemSz,
9722 bool isTailCall,
9723 MachinePointerInfo DstPtrInfo,
9724 MachinePointerInfo SrcPtrInfo) {
9725 // Emit a library call.
9727 Type *IntPtrTy = getDataLayout().getIntPtrType(*getContext());
9728 Args.emplace_back(Dst, IntPtrTy);
9729 Args.emplace_back(Src, IntPtrTy);
9730 Args.emplace_back(Size, SizeTy);
9731
9732 RTLIB::Libcall LibraryCall =
9734 RTLIB::LibcallImpl LibcallImpl = Libcalls->getLibcallImpl(LibraryCall);
9735 if (LibcallImpl == RTLIB::Unsupported)
9736 report_fatal_error("Unsupported element size");
9737
9739 CLI.setDebugLoc(dl)
9740 .setChain(Chain)
9741 .setLibCallee(
9742 Libcalls->getLibcallImplCallingConv(LibcallImpl),
9744 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout())),
9745 std::move(Args))
9747 .setTailCall(isTailCall);
9748
9749 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9750 return CallResult.second;
9751}
9752
9754 SDValue Src, SDValue Size, Align Alignment,
9755 bool isVol, bool AlwaysInline,
9756 const CallInst *CI,
9757 MachinePointerInfo DstPtrInfo,
9758 const AAMDNodes &AAInfo) {
9759 // Check to see if we should lower the memset to stores first.
9760 // For cases within the target-specified limits, this is the best choice.
9762 if (ConstantSize) {
9763 // Memset with size zero? Just return the original chain.
9764 if (ConstantSize->isZero())
9765 return Chain;
9766
9767 SDValue Result = getMemsetStores(*this, dl, Chain, Dst, Src,
9768 ConstantSize->getZExtValue(), Alignment,
9769 isVol, false, DstPtrInfo, AAInfo);
9770
9771 if (Result.getNode())
9772 return Result;
9773 }
9774
9775 // Then check to see if we should lower the memset with target-specific
9776 // code. If the target chooses to do this, this is the next best.
9777 if (TSI) {
9778 SDValue Result = TSI->EmitTargetCodeForMemset(
9779 *this, dl, Chain, Dst, Src, Size, Alignment, isVol, AlwaysInline, DstPtrInfo);
9780 if (Result.getNode())
9781 return Result;
9782 }
9783
9784 // If we really need inline code and the target declined to provide it,
9785 // use a (potentially long) sequence of loads and stores.
9786 if (AlwaysInline) {
9787 assert(ConstantSize && "AlwaysInline requires a constant size!");
9788 SDValue Result = getMemsetStores(*this, dl, Chain, Dst, Src,
9789 ConstantSize->getZExtValue(), Alignment,
9790 isVol, true, DstPtrInfo, AAInfo);
9791 assert(Result &&
9792 "getMemsetStores must return a valid sequence when AlwaysInline");
9793 return Result;
9794 }
9795
9797
9798 // Emit a library call.
9799 auto &Ctx = *getContext();
9800 const auto& DL = getDataLayout();
9801
9803 // FIXME: pass in SDLoc
9804 CLI.setDebugLoc(dl).setChain(Chain);
9805
9806 RTLIB::LibcallImpl BzeroImpl = Libcalls->getLibcallImpl(RTLIB::BZERO);
9807 bool UseBZero = BzeroImpl != RTLIB::Unsupported && isNullConstant(Src);
9808
9809 // If zeroing out and bzero is present, use it.
9810 if (UseBZero) {
9812 Args.emplace_back(Dst, PointerType::getUnqual(Ctx));
9813 Args.emplace_back(Size, DL.getIntPtrType(Ctx));
9814 CLI.setLibCallee(
9815 Libcalls->getLibcallImplCallingConv(BzeroImpl), Type::getVoidTy(Ctx),
9816 getExternalSymbol(BzeroImpl, TLI->getPointerTy(DL)), std::move(Args));
9817 } else {
9818 RTLIB::LibcallImpl MemsetImpl = Libcalls->getLibcallImpl(RTLIB::MEMSET);
9819
9821 Args.emplace_back(Dst, PointerType::getUnqual(Ctx));
9822 Args.emplace_back(Src, Src.getValueType().getTypeForEVT(Ctx));
9823 Args.emplace_back(Size, DL.getIntPtrType(Ctx));
9824 CLI.setLibCallee(Libcalls->getLibcallImplCallingConv(MemsetImpl),
9825 Dst.getValueType().getTypeForEVT(Ctx),
9826 getExternalSymbol(MemsetImpl, TLI->getPointerTy(DL)),
9827 std::move(Args));
9828 }
9829
9830 RTLIB::LibcallImpl MemsetImpl = Libcalls->getLibcallImpl(RTLIB::MEMSET);
9831 bool LowersToMemset = MemsetImpl == RTLIB::impl_memset;
9832
9833 // If we're going to use bzero, make sure not to tail call unless the
9834 // subsequent return doesn't need a value, as bzero doesn't return the first
9835 // arg unlike memset.
9836 bool ReturnsFirstArg = CI && funcReturnsFirstArgOfCall(*CI) && !UseBZero;
9837 bool IsTailCall =
9838 CI && CI->isTailCall() &&
9839 isInTailCallPosition(*CI, getTarget(), ReturnsFirstArg && LowersToMemset);
9840 CLI.setDiscardResult().setTailCall(IsTailCall);
9841
9842 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9843 return CallResult.second;
9844}
9845
9848 Type *SizeTy, unsigned ElemSz,
9849 bool isTailCall,
9850 MachinePointerInfo DstPtrInfo) {
9851 // Emit a library call.
9853 Args.emplace_back(Dst, getDataLayout().getIntPtrType(*getContext()));
9854 Args.emplace_back(Value, Type::getInt8Ty(*getContext()));
9855 Args.emplace_back(Size, SizeTy);
9856
9857 RTLIB::Libcall LibraryCall =
9859 RTLIB::LibcallImpl LibcallImpl = Libcalls->getLibcallImpl(LibraryCall);
9860 if (LibcallImpl == RTLIB::Unsupported)
9861 report_fatal_error("Unsupported element size");
9862
9864 CLI.setDebugLoc(dl)
9865 .setChain(Chain)
9866 .setLibCallee(
9867 Libcalls->getLibcallImplCallingConv(LibcallImpl),
9869 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout())),
9870 std::move(Args))
9872 .setTailCall(isTailCall);
9873
9874 std::pair<SDValue, SDValue> CallResult = TLI->LowerCallTo(CLI);
9875 return CallResult.second;
9876}
9877
9878SDValue SelectionDAG::getAtomic(unsigned Opcode, const SDLoc &dl, EVT MemVT,
9880 MachineMemOperand *MMO,
9881 ISD::LoadExtType ExtType) {
9883 AddNodeIDNode(ID, Opcode, VTList, Ops);
9884 ID.AddInteger(MemVT.getRawBits());
9885 ID.AddInteger(getSyntheticNodeSubclassData<AtomicSDNode>(
9886 dl.getIROrder(), Opcode, VTList, MemVT, MMO, ExtType));
9887 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
9888 ID.AddInteger(MMO->getFlags());
9889 void* IP = nullptr;
9890 if (auto *E = cast_or_null<AtomicSDNode>(FindNodeOrInsertPos(ID, dl, IP))) {
9891 E->refineAlignment(MMO);
9892 E->refineRanges(MMO);
9893 return SDValue(E, 0);
9894 }
9895
9896 auto *N = newSDNode<AtomicSDNode>(dl.getIROrder(), dl.getDebugLoc(), Opcode,
9897 VTList, MemVT, MMO, ExtType);
9898 createOperands(N, Ops);
9899
9900 CSEMap.InsertNode(N, IP);
9901 InsertNode(N);
9902 SDValue V(N, 0);
9903 NewSDValueDbgMsg(V, "Creating new node: ", this);
9904 return V;
9905}
9906
9908 EVT MemVT, SDVTList VTs, SDValue Chain,
9909 SDValue Ptr, SDValue Cmp, SDValue Swp,
9910 MachineMemOperand *MMO) {
9911 assert(Opcode == ISD::ATOMIC_CMP_SWAP ||
9913 assert(Cmp.getValueType() == Swp.getValueType() && "Invalid Atomic Op Types");
9914
9915 SDValue Ops[] = {Chain, Ptr, Cmp, Swp};
9916 return getAtomic(Opcode, dl, MemVT, VTs, Ops, MMO);
9917}
9918
9919SDValue SelectionDAG::getAtomic(unsigned Opcode, const SDLoc &dl, EVT MemVT,
9920 SDValue Chain, SDValue Ptr, SDValue Val,
9921 MachineMemOperand *MMO) {
9922 assert((Opcode == ISD::ATOMIC_LOAD_ADD || Opcode == ISD::ATOMIC_LOAD_SUB ||
9923 Opcode == ISD::ATOMIC_LOAD_AND || Opcode == ISD::ATOMIC_LOAD_CLR ||
9924 Opcode == ISD::ATOMIC_LOAD_OR || Opcode == ISD::ATOMIC_LOAD_XOR ||
9925 Opcode == ISD::ATOMIC_LOAD_NAND || Opcode == ISD::ATOMIC_LOAD_MIN ||
9926 Opcode == ISD::ATOMIC_LOAD_MAX || Opcode == ISD::ATOMIC_LOAD_UMIN ||
9927 Opcode == ISD::ATOMIC_LOAD_UMAX || Opcode == ISD::ATOMIC_LOAD_FADD ||
9928 Opcode == ISD::ATOMIC_LOAD_FSUB || Opcode == ISD::ATOMIC_LOAD_FMAX ||
9929 Opcode == ISD::ATOMIC_LOAD_FMIN ||
9930 Opcode == ISD::ATOMIC_LOAD_FMINIMUM ||
9931 Opcode == ISD::ATOMIC_LOAD_FMAXIMUM ||
9932 Opcode == ISD::ATOMIC_LOAD_UINC_WRAP ||
9933 Opcode == ISD::ATOMIC_LOAD_UDEC_WRAP ||
9934 Opcode == ISD::ATOMIC_LOAD_USUB_COND ||
9935 Opcode == ISD::ATOMIC_LOAD_USUB_SAT || Opcode == ISD::ATOMIC_SWAP ||
9936 Opcode == ISD::ATOMIC_STORE) &&
9937 "Invalid Atomic Op");
9938
9939 EVT VT = Val.getValueType();
9940
9941 SDVTList VTs = Opcode == ISD::ATOMIC_STORE ? getVTList(MVT::Other) :
9942 getVTList(VT, MVT::Other);
9943 SDValue Ops[] = {Chain, Ptr, Val};
9944 return getAtomic(Opcode, dl, MemVT, VTs, Ops, MMO);
9945}
9946
9948 EVT MemVT, EVT VT, SDValue Chain,
9949 SDValue Ptr, MachineMemOperand *MMO) {
9950 SDVTList VTs = getVTList(VT, MVT::Other);
9951 SDValue Ops[] = {Chain, Ptr};
9952 return getAtomic(ISD::ATOMIC_LOAD, dl, MemVT, VTs, Ops, MMO, ExtType);
9953}
9954
9955/// getMergeValues - Create a MERGE_VALUES node from the given operands.
9957 if (Ops.size() == 1)
9958 return Ops[0];
9959
9961 VTs.reserve(Ops.size());
9962 for (const SDValue &Op : Ops)
9963 VTs.push_back(Op.getValueType());
9964 return getNode(ISD::MERGE_VALUES, dl, getVTList(VTs), Ops);
9965}
9966
9968 unsigned Opcode, const SDLoc &dl, SDVTList VTList, ArrayRef<SDValue> Ops,
9969 EVT MemVT, MachinePointerInfo PtrInfo, Align Alignment,
9971 const AAMDNodes &AAInfo) {
9972 if (Size.hasValue() && !Size.getValue())
9974
9976 MachineMemOperand *MMO =
9977 MF.getMachineMemOperand(PtrInfo, Flags, Size, Alignment, AAInfo);
9978
9979 return getMemIntrinsicNode(Opcode, dl, VTList, Ops, MemVT, MMO);
9980}
9981
9983 SDVTList VTList,
9984 ArrayRef<SDValue> Ops, EVT MemVT,
9985 MachineMemOperand *MMO) {
9986 return getMemIntrinsicNode(Opcode, dl, VTList, Ops, MemVT, ArrayRef(MMO));
9987}
9988
9990 SDVTList VTList,
9991 ArrayRef<SDValue> Ops, EVT MemVT,
9993 assert(!MMOs.empty() && "Must have at least one MMO");
9994 assert(
9995 (Opcode == ISD::INTRINSIC_VOID || Opcode == ISD::INTRINSIC_W_CHAIN ||
9996 Opcode == ISD::PREFETCH ||
9997 (Opcode <= (unsigned)std::numeric_limits<int>::max() &&
9998 Opcode >= ISD::BUILTIN_OP_END && TSI->isTargetMemoryOpcode(Opcode))) &&
9999 "Opcode is not a memory-accessing opcode!");
10000
10002 if (MMOs.size() == 1) {
10003 MemRefs = MMOs[0];
10004 } else {
10005 // Allocate: [size_t count][MMO*][MMO*]...
10006 size_t AllocSize =
10007 sizeof(size_t) + MMOs.size() * sizeof(MachineMemOperand *);
10008 void *Buffer = Allocator.Allocate(AllocSize, alignof(size_t));
10009 size_t *CountPtr = static_cast<size_t *>(Buffer);
10010 *CountPtr = MMOs.size();
10011 MachineMemOperand **Array =
10012 reinterpret_cast<MachineMemOperand **>(CountPtr + 1);
10013 llvm::copy(MMOs, Array);
10014 MemRefs = Array;
10015 }
10016
10017 // Memoize the node unless it returns a glue result.
10019 if (VTList.VTs[VTList.NumVTs-1] != MVT::Glue) {
10021 AddNodeIDNode(ID, Opcode, VTList, Ops);
10022 ID.AddInteger(getSyntheticNodeSubclassData<MemIntrinsicSDNode>(
10023 Opcode, dl.getIROrder(), VTList, MemVT, MemRefs));
10024 ID.AddInteger(MemVT.getRawBits());
10025 for (const MachineMemOperand *MMO : MMOs) {
10026 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10027 ID.AddInteger(MMO->getFlags());
10028 }
10029 void *IP = nullptr;
10030 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10031 cast<MemIntrinsicSDNode>(E)->refineAlignment(MMOs);
10032 return SDValue(E, 0);
10033 }
10034
10035 N = newSDNode<MemIntrinsicSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(),
10036 VTList, MemVT, MemRefs);
10037 createOperands(N, Ops);
10038 CSEMap.InsertNode(N, IP);
10039 } else {
10040 N = newSDNode<MemIntrinsicSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(),
10041 VTList, MemVT, MemRefs);
10042 createOperands(N, Ops);
10043 }
10044 InsertNode(N);
10045 SDValue V(N, 0);
10046 NewSDValueDbgMsg(V, "Creating new node: ", this);
10047 return V;
10048}
10049
10051 SDValue Chain, int FrameIndex) {
10052 const unsigned Opcode = IsStart ? ISD::LIFETIME_START : ISD::LIFETIME_END;
10053 const auto VTs = getVTList(MVT::Other);
10054 SDValue Ops[2] = {
10055 Chain,
10056 getFrameIndex(FrameIndex,
10057 getTargetLoweringInfo().getFrameIndexTy(getDataLayout()),
10058 true)};
10059
10061 AddNodeIDNode(ID, Opcode, VTs, Ops);
10062 ID.AddInteger(FrameIndex);
10063 void *IP = nullptr;
10064 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
10065 return SDValue(E, 0);
10066
10067 LifetimeSDNode *N =
10068 newSDNode<LifetimeSDNode>(Opcode, dl.getIROrder(), dl.getDebugLoc(), VTs);
10069 createOperands(N, Ops);
10070 CSEMap.InsertNode(N, IP);
10071 InsertNode(N);
10072 SDValue V(N, 0);
10073 NewSDValueDbgMsg(V, "Creating new node: ", this);
10074 return V;
10075}
10076
10078 uint64_t Guid, uint64_t Index,
10079 uint32_t Attr) {
10080 const unsigned Opcode = ISD::PSEUDO_PROBE;
10081 const auto VTs = getVTList(MVT::Other);
10082 SDValue Ops[] = {Chain};
10084 AddNodeIDNode(ID, Opcode, VTs, Ops);
10085 ID.AddInteger(Guid);
10086 ID.AddInteger(Index);
10087 void *IP = nullptr;
10088 if (SDNode *E = FindNodeOrInsertPos(ID, Dl, IP))
10089 return SDValue(E, 0);
10090
10091 auto *N = newSDNode<PseudoProbeSDNode>(
10092 Opcode, Dl.getIROrder(), Dl.getDebugLoc(), VTs, Guid, Index, Attr);
10093 createOperands(N, Ops);
10094 CSEMap.InsertNode(N, IP);
10095 InsertNode(N);
10096 SDValue V(N, 0);
10097 NewSDValueDbgMsg(V, "Creating new node: ", this);
10098 return V;
10099}
10100
10101/// InferPointerInfo - If the specified ptr/offset is a frame index, infer a
10102/// MachinePointerInfo record from it. This is particularly useful because the
10103/// code generator has many cases where it doesn't bother passing in a
10104/// MachinePointerInfo to getLoad or getStore when it has "FI+Cst".
10106 SelectionDAG &DAG, SDValue Ptr,
10107 int64_t Offset = 0) {
10108 // If this is FI+Offset, we can model it.
10109 if (const FrameIndexSDNode *FI = dyn_cast<FrameIndexSDNode>(Ptr))
10111 FI->getIndex(), Offset);
10112
10113 // If this is (FI+Offset1)+Offset2, we can model it.
10114 if (Ptr.getOpcode() != ISD::ADD ||
10117 return Info;
10118
10119 int FI = cast<FrameIndexSDNode>(Ptr.getOperand(0))->getIndex();
10121 DAG.getMachineFunction(), FI,
10122 Offset + cast<ConstantSDNode>(Ptr.getOperand(1))->getSExtValue());
10123}
10124
10125/// InferPointerInfo - If the specified ptr/offset is a frame index, infer a
10126/// MachinePointerInfo record from it. This is particularly useful because the
10127/// code generator has many cases where it doesn't bother passing in a
10128/// MachinePointerInfo to getLoad or getStore when it has "FI+Cst".
10130 SelectionDAG &DAG, SDValue Ptr,
10131 SDValue OffsetOp) {
10132 // If the 'Offset' value isn't a constant, we can't handle this.
10134 return InferPointerInfo(Info, DAG, Ptr, OffsetNode->getSExtValue());
10135 if (OffsetOp.isUndef())
10136 return InferPointerInfo(Info, DAG, Ptr);
10137 return Info;
10138}
10139
10141 EVT VT, const SDLoc &dl, SDValue Chain,
10142 SDValue Ptr, SDValue Offset,
10143 MachinePointerInfo PtrInfo, EVT MemVT,
10144 Align Alignment,
10145 MachineMemOperand::Flags MMOFlags,
10146 const AAMDNodes &AAInfo, const MDNode *Ranges) {
10147 assert(Chain.getValueType() == MVT::Other &&
10148 "Invalid chain type");
10149
10150 MMOFlags |= MachineMemOperand::MOLoad;
10151 assert((MMOFlags & MachineMemOperand::MOStore) == 0);
10152 // If we don't have a PtrInfo, infer the trivial frame index case to simplify
10153 // clients.
10154 if (PtrInfo.V.isNull())
10155 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr, Offset);
10156
10157 TypeSize Size = MemVT.getStoreSize();
10159 MachineMemOperand *MMO = MF.getMachineMemOperand(PtrInfo, MMOFlags, Size,
10160 Alignment, AAInfo, Ranges);
10161 return getLoad(AM, ExtType, VT, dl, Chain, Ptr, Offset, MemVT, MMO);
10162}
10163
10165 EVT VT, const SDLoc &dl, SDValue Chain,
10166 SDValue Ptr, SDValue Offset, EVT MemVT,
10167 MachineMemOperand *MMO) {
10168 if (VT == MemVT) {
10169 ExtType = ISD::NON_EXTLOAD;
10170 } else if (ExtType == ISD::NON_EXTLOAD) {
10171 assert(VT == MemVT && "Non-extending load from different memory type!");
10172 } else {
10173 // Extending load.
10174 assert(MemVT.getScalarType().bitsLT(VT.getScalarType()) &&
10175 "Should only be an extending load, not truncating!");
10176 assert(VT.isInteger() == MemVT.isInteger() &&
10177 "Cannot convert from FP to Int or Int -> FP!");
10178 assert(VT.isVector() == MemVT.isVector() &&
10179 "Cannot use an ext load to convert to or from a vector!");
10180 assert((!VT.isVector() ||
10182 "Cannot use an ext load to change the number of vector elements!");
10183 }
10184
10185 assert((!MMO->getRanges() ||
10187 ->getBitWidth() == MemVT.getScalarSizeInBits() &&
10188 MemVT.isInteger())) &&
10189 "Range metadata and load type must match!");
10190
10191 bool Indexed = AM != ISD::UNINDEXED;
10192 assert((Indexed || Offset.isUndef()) && "Unindexed load with an offset!");
10193
10194 SDVTList VTs = Indexed ?
10195 getVTList(VT, Ptr.getValueType(), MVT::Other) : getVTList(VT, MVT::Other);
10196 SDValue Ops[] = { Chain, Ptr, Offset };
10198 AddNodeIDNode(ID, ISD::LOAD, VTs, Ops);
10199 ID.AddInteger(MemVT.getRawBits());
10200 ID.AddInteger(getSyntheticNodeSubclassData<LoadSDNode>(
10201 dl.getIROrder(), VTs, AM, ExtType, MemVT, MMO));
10202 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10203 ID.AddInteger(MMO->getFlags());
10204 void *IP = nullptr;
10205 if (auto *E = cast_or_null<LoadSDNode>(FindNodeOrInsertPos(ID, dl, IP))) {
10206 E->refineAlignment(MMO);
10207 E->refineRanges(MMO);
10208 return SDValue(E, 0);
10209 }
10210 auto *N = newSDNode<LoadSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10211 ExtType, MemVT, MMO);
10212 createOperands(N, Ops);
10213
10214 CSEMap.InsertNode(N, IP);
10215 InsertNode(N);
10216 SDValue V(N, 0);
10217 NewSDValueDbgMsg(V, "Creating new node: ", this);
10218 return V;
10219}
10220
10222 SDValue Ptr, MachinePointerInfo PtrInfo,
10223 MaybeAlign Alignment,
10224 MachineMemOperand::Flags MMOFlags,
10225 const AAMDNodes &AAInfo, const MDNode *Ranges) {
10227 return getLoad(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10228 PtrInfo, VT, Alignment, MMOFlags, AAInfo, Ranges);
10229}
10230
10232 SDValue Ptr, MachineMemOperand *MMO) {
10234 return getLoad(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10235 VT, MMO);
10236}
10237
10239 EVT VT, SDValue Chain, SDValue Ptr,
10240 MachinePointerInfo PtrInfo, EVT MemVT,
10241 MaybeAlign Alignment,
10242 MachineMemOperand::Flags MMOFlags,
10243 const AAMDNodes &AAInfo) {
10245 return getLoad(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef, PtrInfo,
10246 MemVT, Alignment, MMOFlags, AAInfo);
10247}
10248
10250 EVT VT, SDValue Chain, SDValue Ptr, EVT MemVT,
10251 MachineMemOperand *MMO) {
10253 return getLoad(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef,
10254 MemVT, MMO);
10255}
10256
10260 LoadSDNode *LD = cast<LoadSDNode>(OrigLoad);
10261 assert(LD->getOffset().isUndef() && "Load is already a indexed load!");
10262 // Don't propagate the invariant or dereferenceable flags.
10263 auto MMOFlags =
10264 LD->getMemOperand()->getFlags() &
10266 return getLoad(AM, LD->getExtensionType(), OrigLoad.getValueType(), dl,
10267 LD->getChain(), Base, Offset, LD->getPointerInfo(),
10268 LD->getMemoryVT(), LD->getAlign(), MMOFlags, LD->getAAInfo());
10269}
10270
10272 SDValue Ptr, MachinePointerInfo PtrInfo,
10273 Align Alignment,
10274 MachineMemOperand::Flags MMOFlags,
10275 const AAMDNodes &AAInfo) {
10276 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10277
10278 MMOFlags |= MachineMemOperand::MOStore;
10279 assert((MMOFlags & MachineMemOperand::MOLoad) == 0);
10280
10281 if (PtrInfo.V.isNull())
10282 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr);
10283
10286 MachineMemOperand *MMO =
10287 MF.getMachineMemOperand(PtrInfo, MMOFlags, Size, Alignment, AAInfo);
10288 return getStore(Chain, dl, Val, Ptr, MMO);
10289}
10290
10292 SDValue Ptr, MachineMemOperand *MMO) {
10294 return getStore(Chain, dl, Val, Ptr, Undef, Val.getValueType(), MMO,
10296}
10297
10299 SDValue Ptr, SDValue Offset, EVT SVT,
10301 bool IsTruncating) {
10302 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10303 EVT VT = Val.getValueType();
10304 if (VT == SVT) {
10305 IsTruncating = false;
10306 } else if (!IsTruncating) {
10307 assert(VT == SVT && "No-truncating store from different memory type!");
10308 } else {
10310 "Should only be a truncating store, not extending!");
10311 assert(VT.isInteger() == SVT.isInteger() && "Can't do FP-INT conversion!");
10312 assert(VT.isVector() == SVT.isVector() &&
10313 "Cannot use trunc store to convert to or from a vector!");
10314 assert((!VT.isVector() ||
10316 "Cannot use trunc store to change the number of vector elements!");
10317 }
10318
10319 bool Indexed = AM != ISD::UNINDEXED;
10320 assert((Indexed || Offset.isUndef()) && "Unindexed store with an offset!");
10321 SDVTList VTs = Indexed ? getVTList(Ptr.getValueType(), MVT::Other)
10322 : getVTList(MVT::Other);
10323 SDValue Ops[] = {Chain, Val, Ptr, Offset};
10326 ID.AddInteger(SVT.getRawBits());
10327 ID.AddInteger(getSyntheticNodeSubclassData<StoreSDNode>(
10328 dl.getIROrder(), VTs, AM, IsTruncating, SVT, MMO));
10329 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10330 ID.AddInteger(MMO->getFlags());
10331 void *IP = nullptr;
10332 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10333 cast<StoreSDNode>(E)->refineAlignment(MMO);
10334 return SDValue(E, 0);
10335 }
10336 auto *N = newSDNode<StoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10337 IsTruncating, SVT, MMO);
10338 createOperands(N, Ops);
10339
10340 CSEMap.InsertNode(N, IP);
10341 InsertNode(N);
10342 SDValue V(N, 0);
10343 NewSDValueDbgMsg(V, "Creating new node: ", this);
10344 return V;
10345}
10346
10348 SDValue Ptr, MachinePointerInfo PtrInfo,
10349 EVT SVT, Align Alignment,
10350 MachineMemOperand::Flags MMOFlags,
10351 const AAMDNodes &AAInfo) {
10352 assert(Chain.getValueType() == MVT::Other &&
10353 "Invalid chain type");
10354
10355 MMOFlags |= MachineMemOperand::MOStore;
10356 assert((MMOFlags & MachineMemOperand::MOLoad) == 0);
10357
10358 if (PtrInfo.V.isNull())
10359 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr);
10360
10362 MachineMemOperand *MMO = MF.getMachineMemOperand(
10363 PtrInfo, MMOFlags, SVT.getStoreSize(), Alignment, AAInfo);
10364 return getTruncStore(Chain, dl, Val, Ptr, SVT, MMO);
10365}
10366
10368 SDValue Ptr, EVT SVT,
10369 MachineMemOperand *MMO) {
10371 return getStore(Chain, dl, Val, Ptr, Undef, SVT, MMO, ISD::UNINDEXED, true);
10372}
10373
10377 StoreSDNode *ST = cast<StoreSDNode>(OrigStore);
10378 assert(ST->getOffset().isUndef() && "Store is already a indexed store!");
10379 return getStore(ST->getChain(), dl, ST->getValue(), Base, Offset,
10380 ST->getMemoryVT(), ST->getMemOperand(), AM,
10381 ST->isTruncatingStore());
10382}
10383
10385 ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &dl,
10386 SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Mask, SDValue EVL,
10387 MachinePointerInfo PtrInfo, EVT MemVT, Align Alignment,
10388 MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo,
10389 const MDNode *Ranges, bool IsExpanding) {
10390 MMOFlags |= MachineMemOperand::MOLoad;
10391 assert((MMOFlags & MachineMemOperand::MOStore) == 0);
10392 // If we don't have a PtrInfo, infer the trivial frame index case to simplify
10393 // clients.
10394 if (PtrInfo.V.isNull())
10395 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr, Offset);
10396
10397 TypeSize Size = MemVT.getStoreSize();
10399 MachineMemOperand *MMO = MF.getMachineMemOperand(PtrInfo, MMOFlags, Size,
10400 Alignment, AAInfo, Ranges);
10401 return getLoadVP(AM, ExtType, VT, dl, Chain, Ptr, Offset, Mask, EVL, MemVT,
10402 MMO, IsExpanding);
10403}
10404
10406 ISD::LoadExtType ExtType, EVT VT,
10407 const SDLoc &dl, SDValue Chain, SDValue Ptr,
10408 SDValue Offset, SDValue Mask, SDValue EVL,
10409 EVT MemVT, MachineMemOperand *MMO,
10410 bool IsExpanding) {
10411 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10412 assert(Mask.getValueType().getVectorElementCount() ==
10413 VT.getVectorElementCount() &&
10414 "Vector width mismatch between mask and data");
10415
10416 bool Indexed = AM != ISD::UNINDEXED;
10417 assert((Indexed || Offset.isUndef()) && "Unindexed load with an offset!");
10418
10419 SDVTList VTs = Indexed ? getVTList(VT, Ptr.getValueType(), MVT::Other)
10420 : getVTList(VT, MVT::Other);
10421 SDValue Ops[] = {Chain, Ptr, Offset, Mask, EVL};
10423 AddNodeIDNode(ID, ISD::VP_LOAD, VTs, Ops);
10424 ID.AddInteger(MemVT.getRawBits());
10425 ID.AddInteger(getSyntheticNodeSubclassData<VPLoadSDNode>(
10426 dl.getIROrder(), VTs, AM, ExtType, IsExpanding, MemVT, MMO));
10427 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10428 ID.AddInteger(MMO->getFlags());
10429 void *IP = nullptr;
10430 if (auto *E = cast_or_null<VPLoadSDNode>(FindNodeOrInsertPos(ID, dl, IP))) {
10431 E->refineAlignment(MMO);
10432 E->refineRanges(MMO);
10433 return SDValue(E, 0);
10434 }
10435 auto *N = newSDNode<VPLoadSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10436 ExtType, IsExpanding, MemVT, MMO);
10437 createOperands(N, Ops);
10438
10439 CSEMap.InsertNode(N, IP);
10440 InsertNode(N);
10441 SDValue V(N, 0);
10442 NewSDValueDbgMsg(V, "Creating new node: ", this);
10443 return V;
10444}
10445
10447 SDValue Ptr, SDValue Mask, SDValue EVL,
10448 MachinePointerInfo PtrInfo,
10449 MaybeAlign Alignment,
10450 MachineMemOperand::Flags MMOFlags,
10451 const AAMDNodes &AAInfo, const MDNode *Ranges,
10452 bool IsExpanding) {
10454 return getLoadVP(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10455 Mask, EVL, PtrInfo, VT, Alignment, MMOFlags, AAInfo, Ranges,
10456 IsExpanding);
10457}
10458
10460 SDValue Ptr, SDValue Mask, SDValue EVL,
10461 MachineMemOperand *MMO, bool IsExpanding) {
10463 return getLoadVP(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, dl, Chain, Ptr, Undef,
10464 Mask, EVL, VT, MMO, IsExpanding);
10465}
10466
10468 EVT VT, SDValue Chain, SDValue Ptr,
10469 SDValue Mask, SDValue EVL,
10470 MachinePointerInfo PtrInfo, EVT MemVT,
10471 MaybeAlign Alignment,
10472 MachineMemOperand::Flags MMOFlags,
10473 const AAMDNodes &AAInfo, bool IsExpanding) {
10475 return getLoadVP(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef, Mask,
10476 EVL, PtrInfo, MemVT, Alignment, MMOFlags, AAInfo, nullptr,
10477 IsExpanding);
10478}
10479
10481 EVT VT, SDValue Chain, SDValue Ptr,
10482 SDValue Mask, SDValue EVL, EVT MemVT,
10483 MachineMemOperand *MMO, bool IsExpanding) {
10485 return getLoadVP(ISD::UNINDEXED, ExtType, VT, dl, Chain, Ptr, Undef, Mask,
10486 EVL, MemVT, MMO, IsExpanding);
10487}
10488
10492 auto *LD = cast<VPLoadSDNode>(OrigLoad);
10493 assert(LD->getOffset().isUndef() && "Load is already a indexed load!");
10494 // Don't propagate the invariant or dereferenceable flags.
10495 auto MMOFlags =
10496 LD->getMemOperand()->getFlags() &
10498 return getLoadVP(AM, LD->getExtensionType(), OrigLoad.getValueType(), dl,
10499 LD->getChain(), Base, Offset, LD->getMask(),
10500 LD->getVectorLength(), LD->getPointerInfo(),
10501 LD->getMemoryVT(), LD->getAlign(), MMOFlags, LD->getAAInfo(),
10502 nullptr, LD->isExpandingLoad());
10503}
10504
10506 SDValue Ptr, SDValue Offset, SDValue Mask,
10507 SDValue EVL, EVT MemVT, MachineMemOperand *MMO,
10508 ISD::MemIndexedMode AM, bool IsTruncating,
10509 bool IsCompressing) {
10510 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10511 assert(Mask.getValueType().getVectorElementCount() ==
10513 "Vector width mismatch between mask and data");
10514
10515 bool Indexed = AM != ISD::UNINDEXED;
10516 assert((Indexed || Offset.isUndef()) && "Unindexed vp_store with an offset!");
10517 SDVTList VTs = Indexed ? getVTList(Ptr.getValueType(), MVT::Other)
10518 : getVTList(MVT::Other);
10519 SDValue Ops[] = {Chain, Val, Ptr, Offset, Mask, EVL};
10521 AddNodeIDNode(ID, ISD::VP_STORE, VTs, Ops);
10522 ID.AddInteger(MemVT.getRawBits());
10523 ID.AddInteger(getSyntheticNodeSubclassData<VPStoreSDNode>(
10524 dl.getIROrder(), VTs, AM, IsTruncating, IsCompressing, MemVT, MMO));
10525 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10526 ID.AddInteger(MMO->getFlags());
10527 void *IP = nullptr;
10528 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10529 cast<VPStoreSDNode>(E)->refineAlignment(MMO);
10530 return SDValue(E, 0);
10531 }
10532 auto *N = newSDNode<VPStoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10533 IsTruncating, IsCompressing, MemVT, MMO);
10534 createOperands(N, Ops);
10535
10536 CSEMap.InsertNode(N, IP);
10537 InsertNode(N);
10538 SDValue V(N, 0);
10539 NewSDValueDbgMsg(V, "Creating new node: ", this);
10540 return V;
10541}
10542
10544 SDValue Val, SDValue Ptr, SDValue Mask,
10545 SDValue EVL, MachinePointerInfo PtrInfo,
10546 EVT SVT, Align Alignment,
10547 MachineMemOperand::Flags MMOFlags,
10548 const AAMDNodes &AAInfo,
10549 bool IsCompressing) {
10550 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10551
10552 MMOFlags |= MachineMemOperand::MOStore;
10553 assert((MMOFlags & MachineMemOperand::MOLoad) == 0);
10554
10555 if (PtrInfo.V.isNull())
10556 PtrInfo = InferPointerInfo(PtrInfo, *this, Ptr);
10557
10559 MachineMemOperand *MMO = MF.getMachineMemOperand(
10560 PtrInfo, MMOFlags, SVT.getStoreSize(), Alignment, AAInfo);
10561 return getTruncStoreVP(Chain, dl, Val, Ptr, Mask, EVL, SVT, MMO,
10562 IsCompressing);
10563}
10564
10566 SDValue Val, SDValue Ptr, SDValue Mask,
10567 SDValue EVL, EVT SVT,
10568 MachineMemOperand *MMO,
10569 bool IsCompressing) {
10570 EVT VT = Val.getValueType();
10571
10572 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10573 if (VT == SVT)
10574 return getStoreVP(Chain, dl, Val, Ptr, getUNDEF(Ptr.getValueType()), Mask,
10575 EVL, VT, MMO, ISD::UNINDEXED,
10576 /*IsTruncating*/ false, IsCompressing);
10577
10579 "Should only be a truncating store, not extending!");
10580 assert(VT.isInteger() == SVT.isInteger() && "Can't do FP-INT conversion!");
10581 assert(VT.isVector() == SVT.isVector() &&
10582 "Cannot use trunc store to convert to or from a vector!");
10583 assert((!VT.isVector() ||
10585 "Cannot use trunc store to change the number of vector elements!");
10586
10587 SDVTList VTs = getVTList(MVT::Other);
10589 SDValue Ops[] = {Chain, Val, Ptr, Undef, Mask, EVL};
10591 AddNodeIDNode(ID, ISD::VP_STORE, VTs, Ops);
10592 ID.AddInteger(SVT.getRawBits());
10593 ID.AddInteger(getSyntheticNodeSubclassData<VPStoreSDNode>(
10594 dl.getIROrder(), VTs, ISD::UNINDEXED, true, IsCompressing, SVT, MMO));
10595 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10596 ID.AddInteger(MMO->getFlags());
10597 void *IP = nullptr;
10598 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10599 cast<VPStoreSDNode>(E)->refineAlignment(MMO);
10600 return SDValue(E, 0);
10601 }
10602 auto *N =
10603 newSDNode<VPStoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10604 ISD::UNINDEXED, true, IsCompressing, SVT, MMO);
10605 createOperands(N, Ops);
10606
10607 CSEMap.InsertNode(N, IP);
10608 InsertNode(N);
10609 SDValue V(N, 0);
10610 NewSDValueDbgMsg(V, "Creating new node: ", this);
10611 return V;
10612}
10613
10617 auto *ST = cast<VPStoreSDNode>(OrigStore);
10618 assert(ST->getOffset().isUndef() && "Store is already an indexed store!");
10619 SDVTList VTs = getVTList(Base.getValueType(), MVT::Other);
10620 SDValue Ops[] = {ST->getChain(), ST->getValue(), Base,
10621 Offset, ST->getMask(), ST->getVectorLength()};
10623 AddNodeIDNode(ID, ISD::VP_STORE, VTs, Ops);
10624 ID.AddInteger(ST->getMemoryVT().getRawBits());
10625 ID.AddInteger(ST->getRawSubclassData());
10626 ID.AddInteger(ST->getPointerInfo().getAddrSpace());
10627 ID.AddInteger(ST->getMemOperand()->getFlags());
10628 void *IP = nullptr;
10629 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
10630 return SDValue(E, 0);
10631
10632 auto *N = newSDNode<VPStoreSDNode>(
10633 dl.getIROrder(), dl.getDebugLoc(), VTs, AM, ST->isTruncatingStore(),
10634 ST->isCompressingStore(), ST->getMemoryVT(), ST->getMemOperand());
10635 createOperands(N, Ops);
10636
10637 CSEMap.InsertNode(N, IP);
10638 InsertNode(N);
10639 SDValue V(N, 0);
10640 NewSDValueDbgMsg(V, "Creating new node: ", this);
10641 return V;
10642}
10643
10645 ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &DL,
10646 SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Stride, SDValue Mask,
10647 SDValue EVL, EVT MemVT, MachineMemOperand *MMO, bool IsExpanding) {
10648 bool Indexed = AM != ISD::UNINDEXED;
10649 assert((Indexed || Offset.isUndef()) && "Unindexed load with an offset!");
10650
10651 SDValue Ops[] = {Chain, Ptr, Offset, Stride, Mask, EVL};
10652 SDVTList VTs = Indexed ? getVTList(VT, Ptr.getValueType(), MVT::Other)
10653 : getVTList(VT, MVT::Other);
10655 AddNodeIDNode(ID, ISD::EXPERIMENTAL_VP_STRIDED_LOAD, VTs, Ops);
10656 ID.AddInteger(VT.getRawBits());
10657 ID.AddInteger(getSyntheticNodeSubclassData<VPStridedLoadSDNode>(
10658 DL.getIROrder(), VTs, AM, ExtType, IsExpanding, MemVT, MMO));
10659 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10660
10661 void *IP = nullptr;
10662 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
10663 cast<VPStridedLoadSDNode>(E)->refineAlignment(MMO);
10664 return SDValue(E, 0);
10665 }
10666
10667 auto *N =
10668 newSDNode<VPStridedLoadSDNode>(DL.getIROrder(), DL.getDebugLoc(), VTs, AM,
10669 ExtType, IsExpanding, MemVT, MMO);
10670 createOperands(N, Ops);
10671 CSEMap.InsertNode(N, IP);
10672 InsertNode(N);
10673 SDValue V(N, 0);
10674 NewSDValueDbgMsg(V, "Creating new node: ", this);
10675 return V;
10676}
10677
10679 SDValue Ptr, SDValue Stride,
10680 SDValue Mask, SDValue EVL,
10681 MachineMemOperand *MMO,
10682 bool IsExpanding) {
10684 return getStridedLoadVP(ISD::UNINDEXED, ISD::NON_EXTLOAD, VT, DL, Chain, Ptr,
10685 Undef, Stride, Mask, EVL, VT, MMO, IsExpanding);
10686}
10687
10689 ISD::LoadExtType ExtType, const SDLoc &DL, EVT VT, SDValue Chain,
10690 SDValue Ptr, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT,
10691 MachineMemOperand *MMO, bool IsExpanding) {
10693 return getStridedLoadVP(ISD::UNINDEXED, ExtType, VT, DL, Chain, Ptr, Undef,
10694 Stride, Mask, EVL, MemVT, MMO, IsExpanding);
10695}
10696
10698 SDValue Val, SDValue Ptr,
10699 SDValue Offset, SDValue Stride,
10700 SDValue Mask, SDValue EVL, EVT MemVT,
10701 MachineMemOperand *MMO,
10703 bool IsTruncating, bool IsCompressing) {
10704 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10705 bool Indexed = AM != ISD::UNINDEXED;
10706 assert((Indexed || Offset.isUndef()) && "Unindexed vp_store with an offset!");
10707 SDVTList VTs = Indexed ? getVTList(Ptr.getValueType(), MVT::Other)
10708 : getVTList(MVT::Other);
10709 SDValue Ops[] = {Chain, Val, Ptr, Offset, Stride, Mask, EVL};
10711 AddNodeIDNode(ID, ISD::EXPERIMENTAL_VP_STRIDED_STORE, VTs, Ops);
10712 ID.AddInteger(MemVT.getRawBits());
10713 ID.AddInteger(getSyntheticNodeSubclassData<VPStridedStoreSDNode>(
10714 DL.getIROrder(), VTs, AM, IsTruncating, IsCompressing, MemVT, MMO));
10715 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10716 void *IP = nullptr;
10717 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
10718 cast<VPStridedStoreSDNode>(E)->refineAlignment(MMO);
10719 return SDValue(E, 0);
10720 }
10721 auto *N = newSDNode<VPStridedStoreSDNode>(DL.getIROrder(), DL.getDebugLoc(),
10722 VTs, AM, IsTruncating,
10723 IsCompressing, MemVT, MMO);
10724 createOperands(N, Ops);
10725
10726 CSEMap.InsertNode(N, IP);
10727 InsertNode(N);
10728 SDValue V(N, 0);
10729 NewSDValueDbgMsg(V, "Creating new node: ", this);
10730 return V;
10731}
10732
10734 SDValue Val, SDValue Ptr,
10735 SDValue Stride, SDValue Mask,
10736 SDValue EVL, EVT SVT,
10737 MachineMemOperand *MMO,
10738 bool IsCompressing) {
10739 EVT VT = Val.getValueType();
10740
10741 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
10742 if (VT == SVT)
10743 return getStridedStoreVP(Chain, DL, Val, Ptr, getUNDEF(Ptr.getValueType()),
10744 Stride, Mask, EVL, VT, MMO, ISD::UNINDEXED,
10745 /*IsTruncating*/ false, IsCompressing);
10746
10748 "Should only be a truncating store, not extending!");
10749 assert(VT.isInteger() == SVT.isInteger() && "Can't do FP-INT conversion!");
10750 assert(VT.isVector() == SVT.isVector() &&
10751 "Cannot use trunc store to convert to or from a vector!");
10752 assert((!VT.isVector() ||
10754 "Cannot use trunc store to change the number of vector elements!");
10755
10756 SDVTList VTs = getVTList(MVT::Other);
10758 SDValue Ops[] = {Chain, Val, Ptr, Undef, Stride, Mask, EVL};
10760 AddNodeIDNode(ID, ISD::EXPERIMENTAL_VP_STRIDED_STORE, VTs, Ops);
10761 ID.AddInteger(SVT.getRawBits());
10762 ID.AddInteger(getSyntheticNodeSubclassData<VPStridedStoreSDNode>(
10763 DL.getIROrder(), VTs, ISD::UNINDEXED, true, IsCompressing, SVT, MMO));
10764 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10765 void *IP = nullptr;
10766 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
10767 cast<VPStridedStoreSDNode>(E)->refineAlignment(MMO);
10768 return SDValue(E, 0);
10769 }
10770 auto *N = newSDNode<VPStridedStoreSDNode>(DL.getIROrder(), DL.getDebugLoc(),
10771 VTs, ISD::UNINDEXED, true,
10772 IsCompressing, SVT, MMO);
10773 createOperands(N, Ops);
10774
10775 CSEMap.InsertNode(N, IP);
10776 InsertNode(N);
10777 SDValue V(N, 0);
10778 NewSDValueDbgMsg(V, "Creating new node: ", this);
10779 return V;
10780}
10781
10784 ISD::MemIndexType IndexType) {
10785 assert(Ops.size() == 6 && "Incompatible number of operands");
10786
10788 AddNodeIDNode(ID, ISD::VP_GATHER, VTs, Ops);
10789 ID.AddInteger(VT.getRawBits());
10790 ID.AddInteger(getSyntheticNodeSubclassData<VPGatherSDNode>(
10791 dl.getIROrder(), VTs, VT, MMO, IndexType));
10792 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10793 ID.AddInteger(MMO->getFlags());
10794 void *IP = nullptr;
10795 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10796 cast<VPGatherSDNode>(E)->refineAlignment(MMO);
10797 return SDValue(E, 0);
10798 }
10799
10800 auto *N = newSDNode<VPGatherSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10801 VT, MMO, IndexType);
10802 createOperands(N, Ops);
10803
10804 assert(N->getMask().getValueType().getVectorElementCount() ==
10805 N->getValueType(0).getVectorElementCount() &&
10806 "Vector width mismatch between mask and data");
10807 assert(N->getIndex().getValueType().getVectorElementCount().isScalable() ==
10808 N->getValueType(0).getVectorElementCount().isScalable() &&
10809 "Scalable flags of index and data do not match");
10811 N->getIndex().getValueType().getVectorElementCount(),
10812 N->getValueType(0).getVectorElementCount()) &&
10813 "Vector width mismatch between index and data");
10814 assert(isa<ConstantSDNode>(N->getScale()) &&
10815 N->getScale()->getAsAPIntVal().isPowerOf2() &&
10816 "Scale should be a constant power of 2");
10817
10818 CSEMap.InsertNode(N, IP);
10819 InsertNode(N);
10820 SDValue V(N, 0);
10821 NewSDValueDbgMsg(V, "Creating new node: ", this);
10822 return V;
10823}
10824
10827 MachineMemOperand *MMO,
10828 ISD::MemIndexType IndexType) {
10829 assert(Ops.size() == 7 && "Incompatible number of operands");
10830
10832 AddNodeIDNode(ID, ISD::VP_SCATTER, VTs, Ops);
10833 ID.AddInteger(VT.getRawBits());
10834 ID.AddInteger(getSyntheticNodeSubclassData<VPScatterSDNode>(
10835 dl.getIROrder(), VTs, VT, MMO, IndexType));
10836 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10837 ID.AddInteger(MMO->getFlags());
10838 void *IP = nullptr;
10839 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10840 cast<VPScatterSDNode>(E)->refineAlignment(MMO);
10841 return SDValue(E, 0);
10842 }
10843 auto *N = newSDNode<VPScatterSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10844 VT, MMO, IndexType);
10845 createOperands(N, Ops);
10846
10847 assert(N->getMask().getValueType().getVectorElementCount() ==
10848 N->getValue().getValueType().getVectorElementCount() &&
10849 "Vector width mismatch between mask and data");
10850 assert(
10851 N->getIndex().getValueType().getVectorElementCount().isScalable() ==
10852 N->getValue().getValueType().getVectorElementCount().isScalable() &&
10853 "Scalable flags of index and data do not match");
10855 N->getIndex().getValueType().getVectorElementCount(),
10856 N->getValue().getValueType().getVectorElementCount()) &&
10857 "Vector width mismatch between index and data");
10858 assert(isa<ConstantSDNode>(N->getScale()) &&
10859 N->getScale()->getAsAPIntVal().isPowerOf2() &&
10860 "Scale should be a constant power of 2");
10861
10862 CSEMap.InsertNode(N, IP);
10863 InsertNode(N);
10864 SDValue V(N, 0);
10865 NewSDValueDbgMsg(V, "Creating new node: ", this);
10866 return V;
10867}
10868
10871 SDValue PassThru, EVT MemVT,
10872 MachineMemOperand *MMO,
10874 ISD::LoadExtType ExtTy, bool isExpanding) {
10875 bool Indexed = AM != ISD::UNINDEXED;
10876 assert((Indexed || Offset.isUndef()) &&
10877 "Unindexed masked load with an offset!");
10878 SDVTList VTs = Indexed ? getVTList(VT, Base.getValueType(), MVT::Other)
10879 : getVTList(VT, MVT::Other);
10880 SDValue Ops[] = {Chain, Base, Offset, Mask, PassThru};
10883 ID.AddInteger(MemVT.getRawBits());
10884 ID.AddInteger(getSyntheticNodeSubclassData<MaskedLoadSDNode>(
10885 dl.getIROrder(), VTs, AM, ExtTy, isExpanding, MemVT, MMO));
10886 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10887 ID.AddInteger(MMO->getFlags());
10888 void *IP = nullptr;
10889 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10890 cast<MaskedLoadSDNode>(E)->refineAlignment(MMO);
10891 return SDValue(E, 0);
10892 }
10893 auto *N = newSDNode<MaskedLoadSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs,
10894 AM, ExtTy, isExpanding, MemVT, MMO);
10895 createOperands(N, Ops);
10896
10897 CSEMap.InsertNode(N, IP);
10898 InsertNode(N);
10899 SDValue V(N, 0);
10900 NewSDValueDbgMsg(V, "Creating new node: ", this);
10901 return V;
10902}
10903
10908 assert(LD->getOffset().isUndef() && "Masked load is already a indexed load!");
10909 return getMaskedLoad(OrigLoad.getValueType(), dl, LD->getChain(), Base,
10910 Offset, LD->getMask(), LD->getPassThru(),
10911 LD->getMemoryVT(), LD->getMemOperand(), AM,
10912 LD->getExtensionType(), LD->isExpandingLoad());
10913}
10914
10917 SDValue Mask, EVT MemVT,
10918 MachineMemOperand *MMO,
10919 ISD::MemIndexedMode AM, bool IsTruncating,
10920 bool IsCompressing) {
10921 assert(Chain.getValueType() == MVT::Other &&
10922 "Invalid chain type");
10923 bool Indexed = AM != ISD::UNINDEXED;
10924 assert((Indexed || Offset.isUndef()) &&
10925 "Unindexed masked store with an offset!");
10926 SDVTList VTs = Indexed ? getVTList(Base.getValueType(), MVT::Other)
10927 : getVTList(MVT::Other);
10928 SDValue Ops[] = {Chain, Val, Base, Offset, Mask};
10931 ID.AddInteger(MemVT.getRawBits());
10932 ID.AddInteger(getSyntheticNodeSubclassData<MaskedStoreSDNode>(
10933 dl.getIROrder(), VTs, AM, IsTruncating, IsCompressing, MemVT, MMO));
10934 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10935 ID.AddInteger(MMO->getFlags());
10936 void *IP = nullptr;
10937 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10938 cast<MaskedStoreSDNode>(E)->refineAlignment(MMO);
10939 return SDValue(E, 0);
10940 }
10941 auto *N =
10942 newSDNode<MaskedStoreSDNode>(dl.getIROrder(), dl.getDebugLoc(), VTs, AM,
10943 IsTruncating, IsCompressing, MemVT, MMO);
10944 createOperands(N, Ops);
10945
10946 CSEMap.InsertNode(N, IP);
10947 InsertNode(N);
10948 SDValue V(N, 0);
10949 NewSDValueDbgMsg(V, "Creating new node: ", this);
10950 return V;
10951}
10952
10957 assert(ST->getOffset().isUndef() &&
10958 "Masked store is already a indexed store!");
10959 return getMaskedStore(ST->getChain(), dl, ST->getValue(), Base, Offset,
10960 ST->getMask(), ST->getMemoryVT(), ST->getMemOperand(),
10961 AM, ST->isTruncatingStore(), ST->isCompressingStore());
10962}
10963
10966 MachineMemOperand *MMO,
10967 ISD::MemIndexType IndexType,
10968 ISD::LoadExtType ExtTy) {
10969 assert(Ops.size() == 6 && "Incompatible number of operands");
10970
10973 ID.AddInteger(MemVT.getRawBits());
10974 ID.AddInteger(getSyntheticNodeSubclassData<MaskedGatherSDNode>(
10975 dl.getIROrder(), VTs, MemVT, MMO, IndexType, ExtTy));
10976 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
10977 ID.AddInteger(MMO->getFlags());
10978 void *IP = nullptr;
10979 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
10980 cast<MaskedGatherSDNode>(E)->refineAlignment(MMO);
10981 return SDValue(E, 0);
10982 }
10983
10984 auto *N = newSDNode<MaskedGatherSDNode>(dl.getIROrder(), dl.getDebugLoc(),
10985 VTs, MemVT, MMO, IndexType, ExtTy);
10986 createOperands(N, Ops);
10987
10988 assert(N->getPassThru().getValueType() == N->getValueType(0) &&
10989 "Incompatible type of the PassThru value in MaskedGatherSDNode");
10990 assert(N->getMask().getValueType().getVectorElementCount() ==
10991 N->getValueType(0).getVectorElementCount() &&
10992 "Vector width mismatch between mask and data");
10993 assert(N->getIndex().getValueType().getVectorElementCount().isScalable() ==
10994 N->getValueType(0).getVectorElementCount().isScalable() &&
10995 "Scalable flags of index and data do not match");
10997 N->getIndex().getValueType().getVectorElementCount(),
10998 N->getValueType(0).getVectorElementCount()) &&
10999 "Vector width mismatch between index and data");
11000 assert(isa<ConstantSDNode>(N->getScale()) &&
11001 N->getScale()->getAsAPIntVal().isPowerOf2() &&
11002 "Scale should be a constant power of 2");
11003
11004 CSEMap.InsertNode(N, IP);
11005 InsertNode(N);
11006 SDValue V(N, 0);
11007 NewSDValueDbgMsg(V, "Creating new node: ", this);
11008 return V;
11009}
11010
11013 MachineMemOperand *MMO,
11014 ISD::MemIndexType IndexType,
11015 bool IsTrunc) {
11016 assert(Ops.size() == 6 && "Incompatible number of operands");
11017
11020 ID.AddInteger(MemVT.getRawBits());
11021 ID.AddInteger(getSyntheticNodeSubclassData<MaskedScatterSDNode>(
11022 dl.getIROrder(), VTs, MemVT, MMO, IndexType, IsTrunc));
11023 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11024 ID.AddInteger(MMO->getFlags());
11025 void *IP = nullptr;
11026 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
11027 cast<MaskedScatterSDNode>(E)->refineAlignment(MMO);
11028 return SDValue(E, 0);
11029 }
11030
11031 auto *N = newSDNode<MaskedScatterSDNode>(dl.getIROrder(), dl.getDebugLoc(),
11032 VTs, MemVT, MMO, IndexType, IsTrunc);
11033 createOperands(N, Ops);
11034
11035 assert(N->getMask().getValueType().getVectorElementCount() ==
11036 N->getValue().getValueType().getVectorElementCount() &&
11037 "Vector width mismatch between mask and data");
11038 assert(
11039 N->getIndex().getValueType().getVectorElementCount().isScalable() ==
11040 N->getValue().getValueType().getVectorElementCount().isScalable() &&
11041 "Scalable flags of index and data do not match");
11043 N->getIndex().getValueType().getVectorElementCount(),
11044 N->getValue().getValueType().getVectorElementCount()) &&
11045 "Vector width mismatch between index and data");
11046 assert(isa<ConstantSDNode>(N->getScale()) &&
11047 N->getScale()->getAsAPIntVal().isPowerOf2() &&
11048 "Scale should be a constant power of 2");
11049
11050 CSEMap.InsertNode(N, IP);
11051 InsertNode(N);
11052 SDValue V(N, 0);
11053 NewSDValueDbgMsg(V, "Creating new node: ", this);
11054 return V;
11055}
11056
11058 const SDLoc &dl, ArrayRef<SDValue> Ops,
11059 MachineMemOperand *MMO,
11060 ISD::MemIndexType IndexType) {
11061 assert(Ops.size() == 7 && "Incompatible number of operands");
11062
11065 ID.AddInteger(MemVT.getRawBits());
11066 ID.AddInteger(getSyntheticNodeSubclassData<MaskedHistogramSDNode>(
11067 dl.getIROrder(), VTs, MemVT, MMO, IndexType));
11068 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11069 ID.AddInteger(MMO->getFlags());
11070 void *IP = nullptr;
11071 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP)) {
11072 cast<MaskedGatherSDNode>(E)->refineAlignment(MMO);
11073 return SDValue(E, 0);
11074 }
11075
11076 auto *N = newSDNode<MaskedHistogramSDNode>(dl.getIROrder(), dl.getDebugLoc(),
11077 VTs, MemVT, MMO, IndexType);
11078 createOperands(N, Ops);
11079
11080 assert(N->getMask().getValueType().getVectorElementCount() ==
11081 N->getIndex().getValueType().getVectorElementCount() &&
11082 "Vector width mismatch between mask and data");
11083 assert(isa<ConstantSDNode>(N->getScale()) &&
11084 N->getScale()->getAsAPIntVal().isPowerOf2() &&
11085 "Scale should be a constant power of 2");
11086 assert(N->getInc().getValueType().isInteger() && "Non integer update value");
11087
11088 CSEMap.InsertNode(N, IP);
11089 InsertNode(N);
11090 SDValue V(N, 0);
11091 NewSDValueDbgMsg(V, "Creating new node: ", this);
11092 return V;
11093}
11094
11096 SDValue Ptr, SDValue Mask, SDValue EVL,
11097 MachineMemOperand *MMO) {
11098 SDVTList VTs = getVTList(VT, EVL.getValueType(), MVT::Other);
11099 SDValue Ops[] = {Chain, Ptr, Mask, EVL};
11101 AddNodeIDNode(ID, ISD::VP_LOAD_FF, VTs, Ops);
11102 ID.AddInteger(VT.getRawBits());
11103 ID.AddInteger(getSyntheticNodeSubclassData<VPLoadFFSDNode>(DL.getIROrder(),
11104 VTs, VT, MMO));
11105 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11106 ID.AddInteger(MMO->getFlags());
11107 void *IP = nullptr;
11108 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
11109 cast<VPLoadFFSDNode>(E)->refineAlignment(MMO);
11110 return SDValue(E, 0);
11111 }
11112 auto *N = newSDNode<VPLoadFFSDNode>(DL.getIROrder(), DL.getDebugLoc(), VTs,
11113 VT, MMO);
11114 createOperands(N, Ops);
11115
11116 CSEMap.InsertNode(N, IP);
11117 InsertNode(N);
11118 SDValue V(N, 0);
11119 NewSDValueDbgMsg(V, "Creating new node: ", this);
11120 return V;
11121}
11122
11124 EVT MemVT, MachineMemOperand *MMO) {
11125 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
11126 SDVTList VTs = getVTList(MVT::Other);
11127 SDValue Ops[] = {Chain, Ptr};
11130 ID.AddInteger(MemVT.getRawBits());
11131 ID.AddInteger(getSyntheticNodeSubclassData<FPStateAccessSDNode>(
11132 ISD::GET_FPENV_MEM, dl.getIROrder(), VTs, MemVT, MMO));
11133 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11134 ID.AddInteger(MMO->getFlags());
11135 void *IP = nullptr;
11136 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
11137 return SDValue(E, 0);
11138
11139 auto *N = newSDNode<FPStateAccessSDNode>(ISD::GET_FPENV_MEM, dl.getIROrder(),
11140 dl.getDebugLoc(), VTs, MemVT, MMO);
11141 createOperands(N, Ops);
11142
11143 CSEMap.InsertNode(N, IP);
11144 InsertNode(N);
11145 SDValue V(N, 0);
11146 NewSDValueDbgMsg(V, "Creating new node: ", this);
11147 return V;
11148}
11149
11151 EVT MemVT, MachineMemOperand *MMO) {
11152 assert(Chain.getValueType() == MVT::Other && "Invalid chain type");
11153 SDVTList VTs = getVTList(MVT::Other);
11154 SDValue Ops[] = {Chain, Ptr};
11157 ID.AddInteger(MemVT.getRawBits());
11158 ID.AddInteger(getSyntheticNodeSubclassData<FPStateAccessSDNode>(
11159 ISD::SET_FPENV_MEM, dl.getIROrder(), VTs, MemVT, MMO));
11160 ID.AddInteger(MMO->getPointerInfo().getAddrSpace());
11161 ID.AddInteger(MMO->getFlags());
11162 void *IP = nullptr;
11163 if (SDNode *E = FindNodeOrInsertPos(ID, dl, IP))
11164 return SDValue(E, 0);
11165
11166 auto *N = newSDNode<FPStateAccessSDNode>(ISD::SET_FPENV_MEM, dl.getIROrder(),
11167 dl.getDebugLoc(), VTs, MemVT, MMO);
11168 createOperands(N, Ops);
11169
11170 CSEMap.InsertNode(N, IP);
11171 InsertNode(N);
11172 SDValue V(N, 0);
11173 NewSDValueDbgMsg(V, "Creating new node: ", this);
11174 return V;
11175}
11176
11178 // select undef, T, F --> T (if T is a constant), otherwise F
11179 // select, ?, undef, F --> F
11180 // select, ?, T, undef --> T
11181 if (Cond.isUndef())
11182 return isConstantValueOfAnyType(T) ? T : F;
11183 if (T.isUndef())
11185 if (F.isUndef())
11187
11188 // select true, T, F --> T
11189 // select false, T, F --> F
11190 if (auto C = isBoolConstant(Cond))
11191 return *C ? T : F;
11192
11193 // select ?, T, T --> T
11194 if (T == F)
11195 return T;
11196
11197 return SDValue();
11198}
11199
11201 // shift undef, Y --> 0 (can always assume that the undef value is 0)
11202 if (X.isUndef())
11203 return getConstant(0, SDLoc(X.getNode()), X.getValueType());
11204 // shift X, undef --> undef (because it may shift by the bitwidth)
11205 if (Y.isUndef())
11206 return getUNDEF(X.getValueType());
11207
11208 // shift 0, Y --> 0
11209 // shift X, 0 --> X
11211 return X;
11212
11213 // shift X, C >= bitwidth(X) --> undef
11214 // All vector elements must be too big (or undef) to avoid partial undefs.
11215 auto isShiftTooBig = [X](ConstantSDNode *Val) {
11216 return !Val || Val->getAPIntValue().uge(X.getScalarValueSizeInBits());
11217 };
11218 if (ISD::matchUnaryPredicate(Y, isShiftTooBig, true))
11219 return getUNDEF(X.getValueType());
11220
11221 // shift i1/vXi1 X, Y --> X (any non-zero shift amount is undefined).
11222 if (X.getValueType().getScalarType() == MVT::i1)
11223 return X;
11224
11225 return SDValue();
11226}
11227
11229 SDNodeFlags Flags) {
11230 // If this operation has 'nnan' or 'ninf' and at least 1 disallowed operand
11231 // (an undef operand can be chosen to be Nan/Inf), then the result of this
11232 // operation is poison. That result can be relaxed to undef.
11233 ConstantFPSDNode *XC = isConstOrConstSplatFP(X, /* AllowUndefs */ true);
11234 ConstantFPSDNode *YC = isConstOrConstSplatFP(Y, /* AllowUndefs */ true);
11235 bool HasNan = (XC && XC->getValueAPF().isNaN()) ||
11236 (YC && YC->getValueAPF().isNaN());
11237 bool HasInf = (XC && XC->getValueAPF().isInfinity()) ||
11238 (YC && YC->getValueAPF().isInfinity());
11239
11240 if (Flags.hasNoNaNs() && (HasNan || X.isUndef() || Y.isUndef()))
11241 return getUNDEF(X.getValueType());
11242
11243 if (Flags.hasNoInfs() && (HasInf || X.isUndef() || Y.isUndef()))
11244 return getUNDEF(X.getValueType());
11245
11246 if (!YC)
11247 return SDValue();
11248
11249 // X + -0.0 --> X
11250 if (Opcode == ISD::FADD)
11251 if (YC->getValueAPF().isNegZero())
11252 return X;
11253
11254 // X - +0.0 --> X
11255 if (Opcode == ISD::FSUB)
11256 if (YC->getValueAPF().isPosZero())
11257 return X;
11258
11259 // X * 1.0 --> X
11260 // X / 1.0 --> X
11261 if (Opcode == ISD::FMUL || Opcode == ISD::FDIV)
11262 if (YC->getValueAPF().isExactlyValue(1.0))
11263 return X;
11264
11265 // X * 0.0 --> 0.0
11266 if (Opcode == ISD::FMUL && Flags.hasNoNaNs() && Flags.hasNoSignedZeros())
11267 if (YC->getValueAPF().isZero())
11268 return getConstantFP(0.0, SDLoc(Y), Y.getValueType());
11269
11270 return SDValue();
11271}
11272
11274 SDValue Ptr, SDValue SV, unsigned Align) {
11275 SDValue Ops[] = { Chain, Ptr, SV, getTargetConstant(Align, dl, MVT::i32) };
11276 return getNode(ISD::VAARG, dl, getVTList(VT, MVT::Other), Ops);
11277}
11278
11279SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
11281 switch (Ops.size()) {
11282 case 0: return getNode(Opcode, DL, VT);
11283 case 1: return getNode(Opcode, DL, VT, Ops[0].get());
11284 case 2: return getNode(Opcode, DL, VT, Ops[0], Ops[1]);
11285 case 3: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Ops[2]);
11286 default: break;
11287 }
11288
11289 // Copy from an SDUse array into an SDValue array for use with
11290 // the regular getNode logic.
11292 return getNode(Opcode, DL, VT, NewOps);
11293}
11294
11295SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
11297 SDNodeFlags Flags;
11298 if (Inserter)
11299 Flags = Inserter->getFlags();
11300 return getNode(Opcode, DL, VT, Ops, Flags);
11301}
11302
11303SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, EVT VT,
11304 ArrayRef<SDValue> Ops, const SDNodeFlags Flags) {
11305 unsigned NumOps = Ops.size();
11306 switch (NumOps) {
11307 case 0: return getNode(Opcode, DL, VT);
11308 case 1: return getNode(Opcode, DL, VT, Ops[0], Flags);
11309 case 2: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Flags);
11310 case 3: return getNode(Opcode, DL, VT, Ops[0], Ops[1], Ops[2], Flags);
11311 default: break;
11312 }
11313
11314#ifndef NDEBUG
11315 for (const auto &Op : Ops)
11316 assert(Op.getOpcode() != ISD::DELETED_NODE &&
11317 "Operand is DELETED_NODE!");
11318#endif
11319
11320 switch (Opcode) {
11321 default: break;
11322 case ISD::BUILD_VECTOR:
11323 // Attempt to simplify BUILD_VECTOR.
11324 if (SDValue V = FoldBUILD_VECTOR(DL, VT, Ops, *this))
11325 return V;
11326 break;
11328 if (SDValue V = foldCONCAT_VECTORS(DL, VT, Ops, *this))
11329 return V;
11330 break;
11331 case ISD::SELECT_CC:
11332 assert(NumOps == 5 && "SELECT_CC takes 5 operands!");
11333 assert(Ops[0].getValueType() == Ops[1].getValueType() &&
11334 "LHS and RHS of condition must have same type!");
11335 assert(Ops[2].getValueType() == Ops[3].getValueType() &&
11336 "True and False arms of SelectCC must have same type!");
11337 assert(Ops[2].getValueType() == VT &&
11338 "select_cc node must be of same type as true and false value!");
11339 assert((!Ops[0].getValueType().isVector() ||
11340 Ops[0].getValueType().getVectorElementCount() ==
11341 VT.getVectorElementCount()) &&
11342 "Expected select_cc with vector result to have the same sized "
11343 "comparison type!");
11344 break;
11345 case ISD::BR_CC:
11346 assert(NumOps == 5 && "BR_CC takes 5 operands!");
11347 assert(Ops[2].getValueType() == Ops[3].getValueType() &&
11348 "LHS/RHS of comparison should match types!");
11349 break;
11350 case ISD::VP_ADD:
11351 case ISD::VP_SUB:
11352 // If it is VP_ADD/VP_SUB mask operation then turn it to VP_XOR
11353 if (VT.getScalarType() == MVT::i1)
11354 Opcode = ISD::VP_XOR;
11355 break;
11356 case ISD::VP_MUL:
11357 // If it is VP_MUL mask operation then turn it to VP_AND
11358 if (VT.getScalarType() == MVT::i1)
11359 Opcode = ISD::VP_AND;
11360 break;
11361 case ISD::VP_REDUCE_MUL:
11362 // If it is VP_REDUCE_MUL mask operation then turn it to VP_REDUCE_AND
11363 if (VT == MVT::i1)
11364 Opcode = ISD::VP_REDUCE_AND;
11365 break;
11366 case ISD::VP_REDUCE_ADD:
11367 // If it is VP_REDUCE_ADD mask operation then turn it to VP_REDUCE_XOR
11368 if (VT == MVT::i1)
11369 Opcode = ISD::VP_REDUCE_XOR;
11370 break;
11371 case ISD::VP_REDUCE_SMAX:
11372 case ISD::VP_REDUCE_UMIN:
11373 // If it is VP_REDUCE_SMAX/VP_REDUCE_UMIN mask operation then turn it to
11374 // VP_REDUCE_AND.
11375 if (VT == MVT::i1)
11376 Opcode = ISD::VP_REDUCE_AND;
11377 break;
11378 case ISD::VP_REDUCE_SMIN:
11379 case ISD::VP_REDUCE_UMAX:
11380 // If it is VP_REDUCE_SMIN/VP_REDUCE_UMAX mask operation then turn it to
11381 // VP_REDUCE_OR.
11382 if (VT == MVT::i1)
11383 Opcode = ISD::VP_REDUCE_OR;
11384 break;
11385 }
11386
11387 // Memoize nodes.
11388 SDNode *N;
11389 SDVTList VTs = getVTList(VT);
11390
11391 if (VT != MVT::Glue) {
11393 AddNodeIDNode(ID, Opcode, VTs, Ops);
11394 void *IP = nullptr;
11395
11396 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
11397 E->intersectFlagsWith(Flags);
11398 return SDValue(E, 0);
11399 }
11400
11401 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
11402 createOperands(N, Ops);
11403
11404 CSEMap.InsertNode(N, IP);
11405 } else {
11406 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
11407 createOperands(N, Ops);
11408 }
11409
11410 N->setFlags(Flags);
11411 InsertNode(N);
11412 SDValue V(N, 0);
11413 NewSDValueDbgMsg(V, "Creating new node: ", this);
11414 return V;
11415}
11416
11417SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL,
11418 ArrayRef<EVT> ResultTys, ArrayRef<SDValue> Ops) {
11419 SDNodeFlags Flags;
11420 if (Inserter)
11421 Flags = Inserter->getFlags();
11422 return getNode(Opcode, DL, getVTList(ResultTys), Ops, Flags);
11423}
11424
11425SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL,
11427 const SDNodeFlags Flags) {
11428 return getNode(Opcode, DL, getVTList(ResultTys), Ops, Flags);
11429}
11430
11431SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11433 SDNodeFlags Flags;
11434 if (Inserter)
11435 Flags = Inserter->getFlags();
11436 return getNode(Opcode, DL, VTList, Ops, Flags);
11437}
11438
11439SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11440 ArrayRef<SDValue> Ops, const SDNodeFlags Flags) {
11441 if (VTList.NumVTs == 1)
11442 return getNode(Opcode, DL, VTList.VTs[0], Ops, Flags);
11443
11444#ifndef NDEBUG
11445 for (const auto &Op : Ops)
11446 assert(Op.getOpcode() != ISD::DELETED_NODE &&
11447 "Operand is DELETED_NODE!");
11448#endif
11449
11450 switch (Opcode) {
11451 case ISD::SADDO:
11452 case ISD::UADDO:
11453 case ISD::SSUBO:
11454 case ISD::USUBO: {
11455 assert(VTList.NumVTs == 2 && Ops.size() == 2 &&
11456 "Invalid add/sub overflow op!");
11457 assert(VTList.VTs[0].isInteger() && VTList.VTs[1].isInteger() &&
11458 Ops[0].getValueType() == Ops[1].getValueType() &&
11459 Ops[0].getValueType() == VTList.VTs[0] &&
11460 "Binary operator types must match!");
11461 SDValue N1 = Ops[0], N2 = Ops[1];
11462 canonicalizeCommutativeBinop(Opcode, N1, N2);
11463
11464 // (X +- 0) -> X with zero-overflow.
11465 ConstantSDNode *N2CV = isConstOrConstSplat(N2, /*AllowUndefs*/ false,
11466 /*AllowTruncation*/ true);
11467 if (N2CV && N2CV->isZero()) {
11468 SDValue ZeroOverFlow = getConstant(0, DL, VTList.VTs[1]);
11469 return getNode(ISD::MERGE_VALUES, DL, VTList, {N1, ZeroOverFlow}, Flags);
11470 }
11471
11472 if (VTList.VTs[0].getScalarType() == MVT::i1 &&
11473 VTList.VTs[1].getScalarType() == MVT::i1) {
11474 SDValue F1 = getFreeze(N1);
11475 SDValue F2 = getFreeze(N2);
11476 // {vXi1,vXi1} (u/s)addo(vXi1 x, vXi1y) -> {xor(x,y),and(x,y)}
11477 if (Opcode == ISD::UADDO || Opcode == ISD::SADDO)
11478 return getNode(ISD::MERGE_VALUES, DL, VTList,
11479 {getNode(ISD::XOR, DL, VTList.VTs[0], F1, F2),
11480 getNode(ISD::AND, DL, VTList.VTs[1], F1, F2)},
11481 Flags);
11482 // {vXi1,vXi1} (u/s)subo(vXi1 x, vXi1y) -> {xor(x,y),and(~x,y)}
11483 if (Opcode == ISD::USUBO || Opcode == ISD::SSUBO) {
11484 SDValue NotF1 = getNOT(DL, F1, VTList.VTs[0]);
11485 return getNode(ISD::MERGE_VALUES, DL, VTList,
11486 {getNode(ISD::XOR, DL, VTList.VTs[0], F1, F2),
11487 getNode(ISD::AND, DL, VTList.VTs[1], NotF1, F2)},
11488 Flags);
11489 }
11490 }
11491 break;
11492 }
11493 case ISD::SADDO_CARRY:
11494 case ISD::UADDO_CARRY:
11495 case ISD::SSUBO_CARRY:
11496 case ISD::USUBO_CARRY:
11497 assert(VTList.NumVTs == 2 && Ops.size() == 3 &&
11498 "Invalid add/sub overflow op!");
11499 assert(VTList.VTs[0].isInteger() && VTList.VTs[1].isInteger() &&
11500 Ops[0].getValueType() == Ops[1].getValueType() &&
11501 Ops[0].getValueType() == VTList.VTs[0] &&
11502 Ops[2].getValueType() == VTList.VTs[1] &&
11503 "Binary operator types must match!");
11504 break;
11505 case ISD::SMUL_LOHI:
11506 case ISD::UMUL_LOHI: {
11507 assert(VTList.NumVTs == 2 && Ops.size() == 2 && "Invalid mul lo/hi op!");
11508 assert(VTList.VTs[0].isInteger() && VTList.VTs[0] == VTList.VTs[1] &&
11509 VTList.VTs[0] == Ops[0].getValueType() &&
11510 VTList.VTs[0] == Ops[1].getValueType() &&
11511 "Binary operator types must match!");
11512 // Constant fold.
11515 if (LHS && RHS) {
11516 unsigned Width = VTList.VTs[0].getScalarSizeInBits();
11517 unsigned OutWidth = Width * 2;
11518 APInt Val = LHS->getAPIntValue();
11519 APInt Mul = RHS->getAPIntValue();
11520 if (Opcode == ISD::SMUL_LOHI) {
11521 Val = Val.sext(OutWidth);
11522 Mul = Mul.sext(OutWidth);
11523 } else {
11524 Val = Val.zext(OutWidth);
11525 Mul = Mul.zext(OutWidth);
11526 }
11527 Val *= Mul;
11528
11529 SDValue Hi =
11530 getConstant(Val.extractBits(Width, Width), DL, VTList.VTs[0]);
11531 SDValue Lo = getConstant(Val.trunc(Width), DL, VTList.VTs[0]);
11532 return getNode(ISD::MERGE_VALUES, DL, VTList, {Lo, Hi}, Flags);
11533 }
11534 break;
11535 }
11536 case ISD::FFREXP: {
11537 assert(VTList.NumVTs == 2 && Ops.size() == 1 && "Invalid ffrexp op!");
11538 assert(VTList.VTs[0].isFloatingPoint() && VTList.VTs[1].isInteger() &&
11539 VTList.VTs[0] == Ops[0].getValueType() && "frexp type mismatch");
11540
11542 int FrexpExp;
11543 APFloat FrexpMant =
11544 frexp(C->getValueAPF(), FrexpExp, APFloat::rmNearestTiesToEven);
11545 SDValue Result0 = getConstantFP(FrexpMant, DL, VTList.VTs[0]);
11546 SDValue Result1 = getSignedConstant(FrexpMant.isFinite() ? FrexpExp : 0,
11547 DL, VTList.VTs[1]);
11548 return getNode(ISD::MERGE_VALUES, DL, VTList, {Result0, Result1}, Flags);
11549 }
11550
11551 break;
11552 }
11554 assert(VTList.NumVTs == 2 && Ops.size() == 2 &&
11555 "Invalid STRICT_FP_EXTEND!");
11556 assert(VTList.VTs[0].isFloatingPoint() &&
11557 Ops[1].getValueType().isFloatingPoint() && "Invalid FP cast!");
11558 assert(VTList.VTs[0].isVector() == Ops[1].getValueType().isVector() &&
11559 "STRICT_FP_EXTEND result type should be vector iff the operand "
11560 "type is vector!");
11561 assert((!VTList.VTs[0].isVector() ||
11562 VTList.VTs[0].getVectorElementCount() ==
11563 Ops[1].getValueType().getVectorElementCount()) &&
11564 "Vector element count mismatch!");
11565 assert(Ops[1].getValueType().bitsLT(VTList.VTs[0]) &&
11566 "Invalid fpext node, dst <= src!");
11567 break;
11569 assert(VTList.NumVTs == 2 && Ops.size() == 3 && "Invalid STRICT_FP_ROUND!");
11570 assert(VTList.VTs[0].isVector() == Ops[1].getValueType().isVector() &&
11571 "STRICT_FP_ROUND result type should be vector iff the operand "
11572 "type is vector!");
11573 assert((!VTList.VTs[0].isVector() ||
11574 VTList.VTs[0].getVectorElementCount() ==
11575 Ops[1].getValueType().getVectorElementCount()) &&
11576 "Vector element count mismatch!");
11577 assert(VTList.VTs[0].isFloatingPoint() &&
11578 Ops[1].getValueType().isFloatingPoint() &&
11579 VTList.VTs[0].bitsLT(Ops[1].getValueType()) &&
11580 Ops[2].getOpcode() == ISD::TargetConstant &&
11581 (Ops[2]->getAsZExtVal() == 0 || Ops[2]->getAsZExtVal() == 1) &&
11582 "Invalid STRICT_FP_ROUND!");
11583 break;
11584 }
11585
11586 // Memoize the node unless it returns a glue result.
11587 SDNode *N;
11588 if (VTList.VTs[VTList.NumVTs-1] != MVT::Glue) {
11590 AddNodeIDNode(ID, Opcode, VTList, Ops);
11591 void *IP = nullptr;
11592 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
11593 E->intersectFlagsWith(Flags);
11594 return SDValue(E, 0);
11595 }
11596
11597 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTList);
11598 createOperands(N, Ops);
11599 CSEMap.InsertNode(N, IP);
11600 } else {
11601 N = newSDNode<SDNode>(Opcode, DL.getIROrder(), DL.getDebugLoc(), VTList);
11602 createOperands(N, Ops);
11603 }
11604
11605 N->setFlags(Flags);
11606 InsertNode(N);
11607 SDValue V(N, 0);
11608 NewSDValueDbgMsg(V, "Creating new node: ", this);
11609 return V;
11610}
11611
11612SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL,
11613 SDVTList VTList) {
11614 return getNode(Opcode, DL, VTList, ArrayRef<SDValue>());
11615}
11616
11617SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11618 SDValue N1) {
11619 SDValue Ops[] = { N1 };
11620 return getNode(Opcode, DL, VTList, Ops);
11621}
11622
11623SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11624 SDValue N1, SDValue N2) {
11625 SDValue Ops[] = { N1, N2 };
11626 return getNode(Opcode, DL, VTList, Ops);
11627}
11628
11629SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11630 SDValue N1, SDValue N2, SDValue N3) {
11631 SDValue Ops[] = { N1, N2, N3 };
11632 return getNode(Opcode, DL, VTList, Ops);
11633}
11634
11635SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11636 SDValue N1, SDValue N2, SDValue N3, SDValue N4) {
11637 SDValue Ops[] = { N1, N2, N3, N4 };
11638 return getNode(Opcode, DL, VTList, Ops);
11639}
11640
11641SDValue SelectionDAG::getNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
11642 SDValue N1, SDValue N2, SDValue N3, SDValue N4,
11643 SDValue N5) {
11644 SDValue Ops[] = { N1, N2, N3, N4, N5 };
11645 return getNode(Opcode, DL, VTList, Ops);
11646}
11647
11649 if (!VT.isExtended())
11650 return makeVTList(SDNode::getValueTypeList(VT.getSimpleVT()), 1);
11651
11652 return makeVTList(&(*EVTs.insert(VT).first), 1);
11653}
11654
11657 ID.AddInteger(2U);
11658 ID.AddInteger(VT1.getRawBits());
11659 ID.AddInteger(VT2.getRawBits());
11660
11661 void *IP = nullptr;
11662 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11663 if (!Result) {
11664 EVT *Array = Allocator.Allocate<EVT>(2);
11665 Array[0] = VT1;
11666 Array[1] = VT2;
11667 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, 2);
11668 VTListMap.InsertNode(Result, IP);
11669 }
11670 return Result->getSDVTList();
11671}
11672
11675 ID.AddInteger(3U);
11676 ID.AddInteger(VT1.getRawBits());
11677 ID.AddInteger(VT2.getRawBits());
11678 ID.AddInteger(VT3.getRawBits());
11679
11680 void *IP = nullptr;
11681 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11682 if (!Result) {
11683 EVT *Array = Allocator.Allocate<EVT>(3);
11684 Array[0] = VT1;
11685 Array[1] = VT2;
11686 Array[2] = VT3;
11687 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, 3);
11688 VTListMap.InsertNode(Result, IP);
11689 }
11690 return Result->getSDVTList();
11691}
11692
11695 ID.AddInteger(4U);
11696 ID.AddInteger(VT1.getRawBits());
11697 ID.AddInteger(VT2.getRawBits());
11698 ID.AddInteger(VT3.getRawBits());
11699 ID.AddInteger(VT4.getRawBits());
11700
11701 void *IP = nullptr;
11702 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11703 if (!Result) {
11704 EVT *Array = Allocator.Allocate<EVT>(4);
11705 Array[0] = VT1;
11706 Array[1] = VT2;
11707 Array[2] = VT3;
11708 Array[3] = VT4;
11709 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, 4);
11710 VTListMap.InsertNode(Result, IP);
11711 }
11712 return Result->getSDVTList();
11713}
11714
11716 unsigned NumVTs = VTs.size();
11718 ID.AddInteger(NumVTs);
11719 for (unsigned index = 0; index < NumVTs; index++) {
11720 ID.AddInteger(VTs[index].getRawBits());
11721 }
11722
11723 void *IP = nullptr;
11724 SDVTListNode *Result = VTListMap.FindNodeOrInsertPos(ID, IP);
11725 if (!Result) {
11726 EVT *Array = Allocator.Allocate<EVT>(NumVTs);
11727 llvm::copy(VTs, Array);
11728 Result = new (Allocator) SDVTListNode(ID.Intern(Allocator), Array, NumVTs);
11729 VTListMap.InsertNode(Result, IP);
11730 }
11731 return Result->getSDVTList();
11732}
11733
11734
11735/// UpdateNodeOperands - *Mutate* the specified node in-place to have the
11736/// specified operands. If the resultant node already exists in the DAG,
11737/// this does not modify the specified node, instead it returns the node that
11738/// already exists. If the resultant node does not exist in the DAG, the
11739/// input node is returned. As a degenerate case, if you specify the same
11740/// input operands as the node already has, the input node is returned.
11742 assert(N->getNumOperands() == 1 && "Update with wrong number of operands");
11743
11744 // Check to see if there is no change.
11745 if (Op == N->getOperand(0)) return N;
11746
11747 // See if the modified node already exists.
11748 void *InsertPos = nullptr;
11749 if (SDNode *Existing = FindModifiedNodeSlot(N, Op, InsertPos))
11750 return Existing;
11751
11752 // Nope it doesn't. Remove the node from its current place in the maps.
11753 if (InsertPos)
11754 if (!RemoveNodeFromCSEMaps(N))
11755 InsertPos = nullptr;
11756
11757 // Now we update the operands.
11758 N->OperandList[0].set(Op);
11759
11761 // If this gets put into a CSE map, add it.
11762 if (InsertPos) CSEMap.InsertNode(N, InsertPos);
11763 return N;
11764}
11765
11767 assert(N->getNumOperands() == 2 && "Update with wrong number of operands");
11768
11769 // Check to see if there is no change.
11770 if (Op1 == N->getOperand(0) && Op2 == N->getOperand(1))
11771 return N; // No operands changed, just return the input node.
11772
11773 // See if the modified node already exists.
11774 void *InsertPos = nullptr;
11775 if (SDNode *Existing = FindModifiedNodeSlot(N, Op1, Op2, InsertPos))
11776 return Existing;
11777
11778 // Nope it doesn't. Remove the node from its current place in the maps.
11779 if (InsertPos)
11780 if (!RemoveNodeFromCSEMaps(N))
11781 InsertPos = nullptr;
11782
11783 // Now we update the operands.
11784 if (N->OperandList[0] != Op1)
11785 N->OperandList[0].set(Op1);
11786 if (N->OperandList[1] != Op2)
11787 N->OperandList[1].set(Op2);
11788
11790 // If this gets put into a CSE map, add it.
11791 if (InsertPos) CSEMap.InsertNode(N, InsertPos);
11792 return N;
11793}
11794
11797 SDValue Ops[] = { Op1, Op2, Op3 };
11798 return UpdateNodeOperands(N, Ops);
11799}
11800
11803 SDValue Op3, SDValue Op4) {
11804 SDValue Ops[] = { Op1, Op2, Op3, Op4 };
11805 return UpdateNodeOperands(N, Ops);
11806}
11807
11810 SDValue Op3, SDValue Op4, SDValue Op5) {
11811 SDValue Ops[] = { Op1, Op2, Op3, Op4, Op5 };
11812 return UpdateNodeOperands(N, Ops);
11813}
11814
11817 unsigned NumOps = Ops.size();
11818 assert(N->getNumOperands() == NumOps &&
11819 "Update with wrong number of operands");
11820
11821 // If no operands changed just return the input node.
11822 if (std::equal(Ops.begin(), Ops.end(), N->op_begin()))
11823 return N;
11824
11825 // See if the modified node already exists.
11826 void *InsertPos = nullptr;
11827 if (SDNode *Existing = FindModifiedNodeSlot(N, Ops, InsertPos))
11828 return Existing;
11829
11830 // Nope it doesn't. Remove the node from its current place in the maps.
11831 if (InsertPos)
11832 if (!RemoveNodeFromCSEMaps(N))
11833 InsertPos = nullptr;
11834
11835 // Now we update the operands.
11836 for (unsigned i = 0; i != NumOps; ++i)
11837 if (N->OperandList[i] != Ops[i])
11838 N->OperandList[i].set(Ops[i]);
11839
11841 // If this gets put into a CSE map, add it.
11842 if (InsertPos) CSEMap.InsertNode(N, InsertPos);
11843 return N;
11844}
11845
11846/// DropOperands - Release the operands and set this node to have
11847/// zero operands.
11849 // Unlike the code in MorphNodeTo that does this, we don't need to
11850 // watch for dead nodes here.
11851 for (op_iterator I = op_begin(), E = op_end(); I != E; ) {
11852 SDUse &Use = *I++;
11853 Use.set(SDValue());
11854 }
11855}
11856
11858 ArrayRef<MachineMemOperand *> NewMemRefs) {
11859 if (NewMemRefs.empty()) {
11860 N->clearMemRefs();
11861 return;
11862 }
11863
11864 // Check if we can avoid allocating by storing a single reference directly.
11865 if (NewMemRefs.size() == 1) {
11866 N->MemRefs = NewMemRefs[0];
11867 N->NumMemRefs = 1;
11868 return;
11869 }
11870
11871 MachineMemOperand **MemRefsBuffer =
11872 Allocator.template Allocate<MachineMemOperand *>(NewMemRefs.size());
11873 llvm::copy(NewMemRefs, MemRefsBuffer);
11874 N->MemRefs = MemRefsBuffer;
11875 N->NumMemRefs = static_cast<int>(NewMemRefs.size());
11876}
11877
11878/// SelectNodeTo - These are wrappers around MorphNodeTo that accept a
11879/// machine opcode.
11880///
11882 EVT VT) {
11883 SDVTList VTs = getVTList(VT);
11884 return SelectNodeTo(N, MachineOpc, VTs, {});
11885}
11886
11888 EVT VT, SDValue Op1) {
11889 SDVTList VTs = getVTList(VT);
11890 SDValue Ops[] = { Op1 };
11891 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11892}
11893
11895 EVT VT, SDValue Op1,
11896 SDValue Op2) {
11897 SDVTList VTs = getVTList(VT);
11898 SDValue Ops[] = { Op1, Op2 };
11899 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11900}
11901
11903 EVT VT, SDValue Op1,
11904 SDValue Op2, SDValue Op3) {
11905 SDVTList VTs = getVTList(VT);
11906 SDValue Ops[] = { Op1, Op2, Op3 };
11907 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11908}
11909
11912 SDVTList VTs = getVTList(VT);
11913 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11914}
11915
11917 EVT VT1, EVT VT2, ArrayRef<SDValue> Ops) {
11918 SDVTList VTs = getVTList(VT1, VT2);
11919 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11920}
11921
11923 EVT VT1, EVT VT2) {
11924 SDVTList VTs = getVTList(VT1, VT2);
11925 return SelectNodeTo(N, MachineOpc, VTs, {});
11926}
11927
11929 EVT VT1, EVT VT2, EVT VT3,
11931 SDVTList VTs = getVTList(VT1, VT2, VT3);
11932 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11933}
11934
11936 EVT VT1, EVT VT2,
11937 SDValue Op1, SDValue Op2) {
11938 SDVTList VTs = getVTList(VT1, VT2);
11939 SDValue Ops[] = { Op1, Op2 };
11940 return SelectNodeTo(N, MachineOpc, VTs, Ops);
11941}
11942
11945 SDNode *New = MorphNodeTo(N, ~MachineOpc, VTs, Ops);
11946 // Reset the NodeID to -1.
11947 New->setNodeId(-1);
11948 if (New != N) {
11949 ReplaceAllUsesWith(N, New);
11951 }
11952 return New;
11953}
11954
11955/// UpdateSDLocOnMergeSDNode - If the opt level is -O0 then it throws away
11956/// the line number information on the merged node since it is not possible to
11957/// preserve the information that operation is associated with multiple lines.
11958/// This will make the debugger working better at -O0, were there is a higher
11959/// probability having other instructions associated with that line.
11960///
11961/// For IROrder, we keep the smaller of the two
11962SDNode *SelectionDAG::UpdateSDLocOnMergeSDNode(SDNode *N, const SDLoc &OLoc) {
11963 DebugLoc NLoc = N->getDebugLoc();
11964 if (NLoc && OptLevel == CodeGenOptLevel::None && OLoc.getDebugLoc() != NLoc) {
11965 N->setDebugLoc(DebugLoc());
11966 }
11967 unsigned Order = std::min(N->getIROrder(), OLoc.getIROrder());
11968 N->setIROrder(Order);
11969 return N;
11970}
11971
11972/// MorphNodeTo - This *mutates* the specified node to have the specified
11973/// return type, opcode, and operands.
11974///
11975/// Note that MorphNodeTo returns the resultant node. If there is already a
11976/// node of the specified opcode and operands, it returns that node instead of
11977/// the current one. Note that the SDLoc need not be the same.
11978///
11979/// Using MorphNodeTo is faster than creating a new node and swapping it in
11980/// with ReplaceAllUsesWith both because it often avoids allocating a new
11981/// node, and because it doesn't require CSE recalculation for any of
11982/// the node's users.
11983///
11984/// However, note that MorphNodeTo recursively deletes dead nodes from the DAG.
11985/// As a consequence it isn't appropriate to use from within the DAG combiner or
11986/// the legalizer which maintain worklists that would need to be updated when
11987/// deleting things.
11990 // If an identical node already exists, use it.
11991 void *IP = nullptr;
11992 if (VTs.VTs[VTs.NumVTs-1] != MVT::Glue) {
11994 AddNodeIDNode(ID, Opc, VTs, Ops);
11995 if (SDNode *ON = FindNodeOrInsertPos(ID, SDLoc(N), IP))
11996 return UpdateSDLocOnMergeSDNode(ON, SDLoc(N));
11997 }
11998
11999 if (!RemoveNodeFromCSEMaps(N))
12000 IP = nullptr;
12001
12002 // Start the morphing.
12003 N->NodeType = Opc;
12004 N->ValueList = VTs.VTs;
12005 N->NumValues = VTs.NumVTs;
12006
12007 // Clear the operands list, updating used nodes to remove this from their
12008 // use list. Keep track of any operands that become dead as a result.
12009 SmallPtrSet<SDNode*, 16> DeadNodeSet;
12010 for (SDNode::op_iterator I = N->op_begin(), E = N->op_end(); I != E; ) {
12011 SDUse &Use = *I++;
12012 SDNode *Used = Use.getNode();
12013 Use.set(SDValue());
12014 if (Used->use_empty())
12015 DeadNodeSet.insert(Used);
12016 }
12017
12018 // For MachineNode, initialize the memory references information.
12020 MN->clearMemRefs();
12021
12022 // Swap for an appropriately sized array from the recycler.
12023 removeOperands(N);
12024 createOperands(N, Ops);
12025
12026 // Delete any nodes that are still dead after adding the uses for the
12027 // new operands.
12028 if (!DeadNodeSet.empty()) {
12029 SmallVector<SDNode *, 16> DeadNodes;
12030 for (SDNode *N : DeadNodeSet)
12031 if (N->use_empty())
12032 DeadNodes.push_back(N);
12033 RemoveDeadNodes(DeadNodes);
12034 }
12035
12036 if (IP)
12037 CSEMap.InsertNode(N, IP); // Memoize the new node.
12038 return N;
12039}
12040
12042 unsigned OrigOpc = Node->getOpcode();
12043 unsigned NewOpc;
12044 switch (OrigOpc) {
12045 default:
12046 llvm_unreachable("mutateStrictFPToFP called with unexpected opcode!");
12047#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
12048 case ISD::STRICT_##DAGN: NewOpc = ISD::DAGN; break;
12049#define CMP_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
12050 case ISD::STRICT_##DAGN: NewOpc = ISD::SETCC; break;
12051#include "llvm/IR/ConstrainedOps.def"
12052 }
12053
12054 assert(Node->getNumValues() == 2 && "Unexpected number of results!");
12055
12056 // We're taking this node out of the chain, so we need to re-link things.
12057 SDValue InputChain = Node->getOperand(0);
12058 SDValue OutputChain = SDValue(Node, 1);
12059 ReplaceAllUsesOfValueWith(OutputChain, InputChain);
12060
12062 for (unsigned i = 1, e = Node->getNumOperands(); i != e; ++i)
12063 Ops.push_back(Node->getOperand(i));
12064
12065 SDVTList VTs = getVTList(Node->getValueType(0));
12066 SDNode *Res = MorphNodeTo(Node, NewOpc, VTs, Ops);
12067
12068 // MorphNodeTo can operate in two ways: if an existing node with the
12069 // specified operands exists, it can just return it. Otherwise, it
12070 // updates the node in place to have the requested operands.
12071 if (Res == Node) {
12072 // If we updated the node in place, reset the node ID. To the isel,
12073 // this should be just like a newly allocated machine node.
12074 Res->setNodeId(-1);
12075 } else {
12078 }
12079
12080 return Res;
12081}
12082
12083/// getMachineNode - These are used for target selectors to create a new node
12084/// with specified return type(s), MachineInstr opcode, and operands.
12085///
12086/// Note that getMachineNode returns the resultant node. If there is already a
12087/// node of the specified opcode and operands, it returns that node instead of
12088/// the current one.
12090 EVT VT) {
12091 SDVTList VTs = getVTList(VT);
12092 return getMachineNode(Opcode, dl, VTs, {});
12093}
12094
12096 EVT VT, SDValue Op1) {
12097 SDVTList VTs = getVTList(VT);
12098 SDValue Ops[] = { Op1 };
12099 return getMachineNode(Opcode, dl, VTs, Ops);
12100}
12101
12103 EVT VT, SDValue Op1, SDValue Op2) {
12104 SDVTList VTs = getVTList(VT);
12105 SDValue Ops[] = { Op1, Op2 };
12106 return getMachineNode(Opcode, dl, VTs, Ops);
12107}
12108
12110 EVT VT, SDValue Op1, SDValue Op2,
12111 SDValue Op3) {
12112 SDVTList VTs = getVTList(VT);
12113 SDValue Ops[] = { Op1, Op2, Op3 };
12114 return getMachineNode(Opcode, dl, VTs, Ops);
12115}
12116
12119 SDVTList VTs = getVTList(VT);
12120 return getMachineNode(Opcode, dl, VTs, Ops);
12121}
12122
12124 EVT VT1, EVT VT2, SDValue Op1,
12125 SDValue Op2) {
12126 SDVTList VTs = getVTList(VT1, VT2);
12127 SDValue Ops[] = { Op1, Op2 };
12128 return getMachineNode(Opcode, dl, VTs, Ops);
12129}
12130
12132 EVT VT1, EVT VT2, SDValue Op1,
12133 SDValue Op2, SDValue Op3) {
12134 SDVTList VTs = getVTList(VT1, VT2);
12135 SDValue Ops[] = { Op1, Op2, Op3 };
12136 return getMachineNode(Opcode, dl, VTs, Ops);
12137}
12138
12140 EVT VT1, EVT VT2,
12142 SDVTList VTs = getVTList(VT1, VT2);
12143 return getMachineNode(Opcode, dl, VTs, Ops);
12144}
12145
12147 EVT VT1, EVT VT2, EVT VT3,
12148 SDValue Op1, SDValue Op2) {
12149 SDVTList VTs = getVTList(VT1, VT2, VT3);
12150 SDValue Ops[] = { Op1, Op2 };
12151 return getMachineNode(Opcode, dl, VTs, Ops);
12152}
12153
12155 EVT VT1, EVT VT2, EVT VT3,
12156 SDValue Op1, SDValue Op2,
12157 SDValue Op3) {
12158 SDVTList VTs = getVTList(VT1, VT2, VT3);
12159 SDValue Ops[] = { Op1, Op2, Op3 };
12160 return getMachineNode(Opcode, dl, VTs, Ops);
12161}
12162
12164 EVT VT1, EVT VT2, EVT VT3,
12166 SDVTList VTs = getVTList(VT1, VT2, VT3);
12167 return getMachineNode(Opcode, dl, VTs, Ops);
12168}
12169
12171 ArrayRef<EVT> ResultTys,
12173 SDVTList VTs = getVTList(ResultTys);
12174 return getMachineNode(Opcode, dl, VTs, Ops);
12175}
12176
12178 SDVTList VTs,
12180 bool DoCSE = VTs.VTs[VTs.NumVTs-1] != MVT::Glue;
12182 void *IP = nullptr;
12183
12184 if (DoCSE) {
12186 AddNodeIDNode(ID, ~Opcode, VTs, Ops);
12187 IP = nullptr;
12188 if (SDNode *E = FindNodeOrInsertPos(ID, DL, IP)) {
12189 return cast<MachineSDNode>(UpdateSDLocOnMergeSDNode(E, DL));
12190 }
12191 }
12192
12193 // Allocate a new MachineSDNode.
12194 N = newSDNode<MachineSDNode>(~Opcode, DL.getIROrder(), DL.getDebugLoc(), VTs);
12195 createOperands(N, Ops);
12196
12197 if (DoCSE)
12198 CSEMap.InsertNode(N, IP);
12199
12200 InsertNode(N);
12201 NewSDValueDbgMsg(SDValue(N, 0), "Creating new machine node: ", this);
12202 return N;
12203}
12204
12205/// getTargetExtractSubreg - A convenience function for creating
12206/// TargetOpcode::EXTRACT_SUBREG nodes.
12208 SDValue Operand) {
12209 SDValue SRIdxVal = getTargetConstant(SRIdx, DL, MVT::i32);
12210 SDNode *Subreg = getMachineNode(TargetOpcode::EXTRACT_SUBREG, DL,
12211 VT, Operand, SRIdxVal);
12212 return SDValue(Subreg, 0);
12213}
12214
12215/// getTargetInsertSubreg - A convenience function for creating
12216/// TargetOpcode::INSERT_SUBREG nodes.
12218 SDValue Operand, SDValue Subreg) {
12219 SDValue SRIdxVal = getTargetConstant(SRIdx, DL, MVT::i32);
12220 SDNode *Result = getMachineNode(TargetOpcode::INSERT_SUBREG, DL,
12221 VT, Operand, Subreg, SRIdxVal);
12222 return SDValue(Result, 0);
12223}
12224
12225/// getNodeIfExists - Get the specified node if it's already available, or
12226/// else return NULL.
12229 bool AllowCommute) {
12230 SDNodeFlags Flags;
12231 if (Inserter)
12232 Flags = Inserter->getFlags();
12233 return getNodeIfExists(Opcode, VTList, Ops, Flags, AllowCommute);
12234}
12235
12238 const SDNodeFlags Flags,
12239 bool AllowCommute) {
12240 if (VTList.VTs[VTList.NumVTs - 1] == MVT::Glue)
12241 return nullptr;
12242
12243 auto Lookup = [&](ArrayRef<SDValue> LookupOps) -> SDNode * {
12245 AddNodeIDNode(ID, Opcode, VTList, LookupOps);
12246 void *IP = nullptr;
12247 if (SDNode *E = FindNodeOrInsertPos(ID, IP)) {
12248 E->intersectFlagsWith(Flags);
12249 return E;
12250 }
12251 return nullptr;
12252 };
12253
12254 if (SDNode *Existing = Lookup(Ops))
12255 return Existing;
12256
12257 if (AllowCommute && TLI->isCommutativeBinOp(Opcode))
12258 return Lookup({Ops[1], Ops[0]});
12259
12260 return nullptr;
12261}
12262
12263/// doesNodeExist - Check if a node exists without modifying its flags.
12264bool SelectionDAG::doesNodeExist(unsigned Opcode, SDVTList VTList,
12266 if (VTList.VTs[VTList.NumVTs - 1] != MVT::Glue) {
12268 AddNodeIDNode(ID, Opcode, VTList, Ops);
12269 void *IP = nullptr;
12270 if (FindNodeOrInsertPos(ID, SDLoc(), IP))
12271 return true;
12272 }
12273 return false;
12274}
12275
12276/// getDbgValue - Creates a SDDbgValue node.
12277///
12278/// SDNode
12280 SDNode *N, unsigned R, bool IsIndirect,
12281 const DebugLoc &DL, unsigned O) {
12282 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12283 "Expected inlined-at fields to agree");
12284 return new (DbgInfo->getAlloc())
12285 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromNode(N, R),
12286 {}, IsIndirect, DL, O,
12287 /*IsVariadic=*/false);
12288}
12289
12290/// Constant
12292 DIExpression *Expr,
12293 const Value *C,
12294 const DebugLoc &DL, unsigned O) {
12295 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12296 "Expected inlined-at fields to agree");
12297 return new (DbgInfo->getAlloc())
12298 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromConst(C), {},
12299 /*IsIndirect=*/false, DL, O,
12300 /*IsVariadic=*/false);
12301}
12302
12303/// FrameIndex
12305 DIExpression *Expr, unsigned FI,
12306 bool IsIndirect,
12307 const DebugLoc &DL,
12308 unsigned O) {
12309 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12310 "Expected inlined-at fields to agree");
12311 return getFrameIndexDbgValue(Var, Expr, FI, {}, IsIndirect, DL, O);
12312}
12313
12314/// FrameIndex with dependencies
12316 DIExpression *Expr, unsigned FI,
12317 ArrayRef<SDNode *> Dependencies,
12318 bool IsIndirect,
12319 const DebugLoc &DL,
12320 unsigned O) {
12321 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12322 "Expected inlined-at fields to agree");
12323 return new (DbgInfo->getAlloc())
12324 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromFrameIdx(FI),
12325 Dependencies, IsIndirect, DL, O,
12326 /*IsVariadic=*/false);
12327}
12328
12329/// VReg
12331 Register VReg, bool IsIndirect,
12332 const DebugLoc &DL, unsigned O) {
12333 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12334 "Expected inlined-at fields to agree");
12335 return new (DbgInfo->getAlloc())
12336 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, SDDbgOperand::fromVReg(VReg),
12337 {}, IsIndirect, DL, O,
12338 /*IsVariadic=*/false);
12339}
12340
12343 ArrayRef<SDNode *> Dependencies,
12344 bool IsIndirect, const DebugLoc &DL,
12345 unsigned O, bool IsVariadic) {
12346 assert(cast<DILocalVariable>(Var)->isValidLocationForIntrinsic(DL) &&
12347 "Expected inlined-at fields to agree");
12348 return new (DbgInfo->getAlloc())
12349 SDDbgValue(DbgInfo->getAlloc(), Var, Expr, Locs, Dependencies, IsIndirect,
12350 DL, O, IsVariadic);
12351}
12352
12354 unsigned OffsetInBits, unsigned SizeInBits,
12355 bool InvalidateDbg) {
12356 SDNode *FromNode = From.getNode();
12357 SDNode *ToNode = To.getNode();
12358 assert(FromNode && ToNode && "Can't modify dbg values");
12359
12360 // PR35338
12361 // TODO: assert(From != To && "Redundant dbg value transfer");
12362 // TODO: assert(FromNode != ToNode && "Intranode dbg value transfer");
12363 if (From == To || FromNode == ToNode)
12364 return;
12365
12366 if (!FromNode->getHasDebugValue())
12367 return;
12368
12369 SDDbgOperand FromLocOp =
12370 SDDbgOperand::fromNode(From.getNode(), From.getResNo());
12372
12374 for (SDDbgValue *Dbg : GetDbgValues(FromNode)) {
12375 if (Dbg->isInvalidated())
12376 continue;
12377
12378 // TODO: assert(!Dbg->isInvalidated() && "Transfer of invalid dbg value");
12379
12380 // Create a new location ops vector that is equal to the old vector, but
12381 // with each instance of FromLocOp replaced with ToLocOp.
12382 bool Changed = false;
12383 auto NewLocOps = Dbg->copyLocationOps();
12384 std::replace_if(
12385 NewLocOps.begin(), NewLocOps.end(),
12386 [&Changed, FromLocOp](const SDDbgOperand &Op) {
12387 bool Match = Op == FromLocOp;
12388 Changed |= Match;
12389 return Match;
12390 },
12391 ToLocOp);
12392 // Ignore this SDDbgValue if we didn't find a matching location.
12393 if (!Changed)
12394 continue;
12395
12396 DIVariable *Var = Dbg->getVariable();
12397 auto *Expr = Dbg->getExpression();
12398 // If a fragment is requested, update the expression.
12399 if (SizeInBits) {
12400 // When splitting a larger (e.g., sign-extended) value whose
12401 // lower bits are described with an SDDbgValue, do not attempt
12402 // to transfer the SDDbgValue to the upper bits.
12403 if (auto FI = Expr->getFragmentInfo())
12404 if (OffsetInBits + SizeInBits > FI->SizeInBits)
12405 continue;
12406 auto Fragment = DIExpression::createFragmentExpression(Expr, OffsetInBits,
12407 SizeInBits);
12408 if (!Fragment)
12409 continue;
12410 Expr = *Fragment;
12411 }
12412
12413 auto AdditionalDependencies = Dbg->getAdditionalDependencies();
12414 // Clone the SDDbgValue and move it to To.
12415 SDDbgValue *Clone = getDbgValueList(
12416 Var, Expr, NewLocOps, AdditionalDependencies, Dbg->isIndirect(),
12417 Dbg->getDebugLoc(), std::max(ToNode->getIROrder(), Dbg->getOrder()),
12418 Dbg->isVariadic());
12419 ClonedDVs.push_back(Clone);
12420
12421 if (InvalidateDbg) {
12422 // Invalidate value and indicate the SDDbgValue should not be emitted.
12423 Dbg->setIsInvalidated();
12424 Dbg->setIsEmitted();
12425 }
12426 }
12427
12428 for (SDDbgValue *Dbg : ClonedDVs) {
12429 assert(is_contained(Dbg->getSDNodes(), ToNode) &&
12430 "Transferred DbgValues should depend on the new SDNode");
12431 AddDbgValue(Dbg, false);
12432 }
12433}
12434
12436 if (!N.getHasDebugValue())
12437 return;
12438
12439 auto GetLocationOperand = [](SDNode *Node, unsigned ResNo) {
12440 if (auto *FISDN = dyn_cast<FrameIndexSDNode>(Node))
12441 return SDDbgOperand::fromFrameIdx(FISDN->getIndex());
12442 return SDDbgOperand::fromNode(Node, ResNo);
12443 };
12444
12446 for (auto *DV : GetDbgValues(&N)) {
12447 if (DV->isInvalidated())
12448 continue;
12449 switch (N.getOpcode()) {
12450 default:
12451 break;
12452 case ISD::ADD: {
12453 SDValue N0 = N.getOperand(0);
12454 SDValue N1 = N.getOperand(1);
12455 if (!isa<ConstantSDNode>(N0)) {
12456 bool RHSConstant = isa<ConstantSDNode>(N1);
12458 if (RHSConstant)
12459 Offset = N.getConstantOperandVal(1);
12460 // We are not allowed to turn indirect debug values variadic, so
12461 // don't salvage those.
12462 if (!RHSConstant && DV->isIndirect())
12463 continue;
12464
12465 // Rewrite an ADD constant node into a DIExpression. Since we are
12466 // performing arithmetic to compute the variable's *value* in the
12467 // DIExpression, we need to mark the expression with a
12468 // DW_OP_stack_value.
12469 auto *DIExpr = DV->getExpression();
12470 auto NewLocOps = DV->copyLocationOps();
12471 bool Changed = false;
12472 size_t OrigLocOpsSize = NewLocOps.size();
12473 for (size_t i = 0; i < OrigLocOpsSize; ++i) {
12474 // We're not given a ResNo to compare against because the whole
12475 // node is going away. We know that any ISD::ADD only has one
12476 // result, so we can assume any node match is using the result.
12477 if (NewLocOps[i].getKind() != SDDbgOperand::SDNODE ||
12478 NewLocOps[i].getSDNode() != &N)
12479 continue;
12480 NewLocOps[i] = GetLocationOperand(N0.getNode(), N0.getResNo());
12481 if (RHSConstant) {
12484 DIExpr = DIExpression::appendOpsToArg(DIExpr, ExprOps, i, true);
12485 } else {
12486 // Convert to a variadic expression (if not already).
12487 // convertToVariadicExpression() returns a const pointer, so we use
12488 // a temporary const variable here.
12489 const auto *TmpDIExpr =
12493 ExprOps.push_back(NewLocOps.size());
12494 ExprOps.push_back(dwarf::DW_OP_plus);
12495 SDDbgOperand RHS =
12497 NewLocOps.push_back(RHS);
12498 DIExpr = DIExpression::appendOpsToArg(TmpDIExpr, ExprOps, i, true);
12499 }
12500 Changed = true;
12501 }
12502 (void)Changed;
12503 assert(Changed && "Salvage target doesn't use N");
12504
12505 bool IsVariadic =
12506 DV->isVariadic() || OrigLocOpsSize != NewLocOps.size();
12507
12508 auto AdditionalDependencies = DV->getAdditionalDependencies();
12509 SDDbgValue *Clone = getDbgValueList(
12510 DV->getVariable(), DIExpr, NewLocOps, AdditionalDependencies,
12511 DV->isIndirect(), DV->getDebugLoc(), DV->getOrder(), IsVariadic);
12512 ClonedDVs.push_back(Clone);
12513 DV->setIsInvalidated();
12514 DV->setIsEmitted();
12515 LLVM_DEBUG(dbgs() << "SALVAGE: Rewriting";
12516 N0.getNode()->dumprFull(this);
12517 dbgs() << " into " << *DIExpr << '\n');
12518 }
12519 break;
12520 }
12521 case ISD::TRUNCATE: {
12522 SDValue N0 = N.getOperand(0);
12523 TypeSize FromSize = N0.getValueSizeInBits();
12524 TypeSize ToSize = N.getValueSizeInBits(0);
12525
12526 DIExpression *DbgExpression = DV->getExpression();
12527 auto ExtOps = DIExpression::getExtOps(FromSize, ToSize, false);
12528 auto NewLocOps = DV->copyLocationOps();
12529 bool Changed = false;
12530 for (size_t i = 0; i < NewLocOps.size(); ++i) {
12531 if (NewLocOps[i].getKind() != SDDbgOperand::SDNODE ||
12532 NewLocOps[i].getSDNode() != &N)
12533 continue;
12534
12535 NewLocOps[i] = GetLocationOperand(N0.getNode(), N0.getResNo());
12536 DbgExpression = DIExpression::appendOpsToArg(DbgExpression, ExtOps, i);
12537 Changed = true;
12538 }
12539 assert(Changed && "Salvage target doesn't use N");
12540 (void)Changed;
12541
12542 SDDbgValue *Clone =
12543 getDbgValueList(DV->getVariable(), DbgExpression, NewLocOps,
12544 DV->getAdditionalDependencies(), DV->isIndirect(),
12545 DV->getDebugLoc(), DV->getOrder(), DV->isVariadic());
12546
12547 ClonedDVs.push_back(Clone);
12548 DV->setIsInvalidated();
12549 DV->setIsEmitted();
12550 LLVM_DEBUG(dbgs() << "SALVAGE: Rewriting"; N0.getNode()->dumprFull(this);
12551 dbgs() << " into " << *DbgExpression << '\n');
12552 break;
12553 }
12554 }
12555 }
12556
12557 for (SDDbgValue *Dbg : ClonedDVs) {
12558 assert((!Dbg->getSDNodes().empty() ||
12559 llvm::any_of(Dbg->getLocationOps(),
12560 [&](const SDDbgOperand &Op) {
12561 return Op.getKind() == SDDbgOperand::FRAMEIX;
12562 })) &&
12563 "Salvaged DbgValue should depend on a new SDNode");
12564 AddDbgValue(Dbg, false);
12565 }
12566}
12567
12568/// Creates a SDDbgLabel node.
12570 const DebugLoc &DL, unsigned O) {
12571 assert(cast<DILabel>(Label)->isValidLocationForIntrinsic(DL) &&
12572 "Expected inlined-at fields to agree");
12573 return new (DbgInfo->getAlloc()) SDDbgLabel(Label, DL, O);
12574}
12575
12576namespace {
12577
12578/// RAUWUpdateListener - Helper for ReplaceAllUsesWith - When the node
12579/// pointed to by a use iterator is deleted, increment the use iterator
12580/// so that it doesn't dangle.
12581///
12582class RAUWUpdateListener : public SelectionDAG::DAGUpdateListener {
12585
12586 void NodeDeleted(SDNode *N, SDNode *E) override {
12587 // Increment the iterator as needed.
12588 while (UI != UE && N == UI->getUser())
12589 ++UI;
12590 }
12591
12592public:
12593 RAUWUpdateListener(SelectionDAG &d,
12596 : SelectionDAG::DAGUpdateListener(d), UI(ui), UE(ue) {}
12597};
12598
12599} // end anonymous namespace
12600
12601/// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead.
12602/// This can cause recursive merging of nodes in the DAG.
12603///
12604/// This version assumes From has a single result value.
12605///
12607 SDNode *From = FromN.getNode();
12608 assert(From->getNumValues() == 1 && FromN.getResNo() == 0 &&
12609 "Cannot replace with this method!");
12610 assert(From != To.getNode() && "Cannot replace uses of with self");
12611
12612 // Preserve Debug Values
12613 transferDbgValues(FromN, To);
12614 // Preserve extra info.
12615 copyExtraInfo(From, To.getNode());
12616
12617 // Iterate over all the existing uses of From. New uses will be added
12618 // to the beginning of the use list, which we avoid visiting.
12619 // This specifically avoids visiting uses of From that arise while the
12620 // replacement is happening, because any such uses would be the result
12621 // of CSE: If an existing node looks like From after one of its operands
12622 // is replaced by To, we don't want to replace of all its users with To
12623 // too. See PR3018 for more info.
12624 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end();
12625 RAUWUpdateListener Listener(*this, UI, UE);
12626 while (UI != UE) {
12627 SDNode *User = UI->getUser();
12628
12629 // This node is about to morph, remove its old self from the CSE maps.
12630 RemoveNodeFromCSEMaps(User);
12631
12632 // A user can appear in a use list multiple times, and when this
12633 // happens the uses are usually next to each other in the list.
12634 // To help reduce the number of CSE recomputations, process all
12635 // the uses of this user that we can find this way.
12636 do {
12637 SDUse &Use = *UI;
12638 ++UI;
12639 Use.set(To);
12640 if (To->isDivergent() != From->isDivergent())
12642 } while (UI != UE && UI->getUser() == User);
12643 // Now that we have modified User, add it back to the CSE maps. If it
12644 // already exists there, recursively merge the results together.
12645 AddModifiedNodeToCSEMaps(User);
12646 }
12647
12648 // If we just RAUW'd the root, take note.
12649 if (FromN == getRoot())
12650 setRoot(To);
12651}
12652
12653/// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead.
12654/// This can cause recursive merging of nodes in the DAG.
12655///
12656/// This version assumes that for each value of From, there is a
12657/// corresponding value in To in the same position with the same type.
12658///
12660#ifndef NDEBUG
12661 for (unsigned i = 0, e = From->getNumValues(); i != e; ++i)
12662 assert((!From->hasAnyUseOfValue(i) ||
12663 From->getValueType(i) == To->getValueType(i)) &&
12664 "Cannot use this version of ReplaceAllUsesWith!");
12665#endif
12666
12667 // Handle the trivial case.
12668 if (From == To)
12669 return;
12670
12671 // Preserve Debug Info. Only do this if there's a use.
12672 for (unsigned i = 0, e = From->getNumValues(); i != e; ++i)
12673 if (From->hasAnyUseOfValue(i)) {
12674 assert((i < To->getNumValues()) && "Invalid To location");
12675 transferDbgValues(SDValue(From, i), SDValue(To, i));
12676 }
12677 // Preserve extra info.
12678 copyExtraInfo(From, To);
12679
12680 // Iterate over just the existing users of From. See the comments in
12681 // the ReplaceAllUsesWith above.
12682 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end();
12683 RAUWUpdateListener Listener(*this, UI, UE);
12684 while (UI != UE) {
12685 SDNode *User = UI->getUser();
12686
12687 // This node is about to morph, remove its old self from the CSE maps.
12688 RemoveNodeFromCSEMaps(User);
12689
12690 // A user can appear in a use list multiple times, and when this
12691 // happens the uses are usually next to each other in the list.
12692 // To help reduce the number of CSE recomputations, process all
12693 // the uses of this user that we can find this way.
12694 do {
12695 SDUse &Use = *UI;
12696 ++UI;
12697 Use.setNode(To);
12698 if (To->isDivergent() != From->isDivergent())
12700 } while (UI != UE && UI->getUser() == User);
12701
12702 // Now that we have modified User, add it back to the CSE maps. If it
12703 // already exists there, recursively merge the results together.
12704 AddModifiedNodeToCSEMaps(User);
12705 }
12706
12707 // If we just RAUW'd the root, take note.
12708 if (From == getRoot().getNode())
12709 setRoot(SDValue(To, getRoot().getResNo()));
12710}
12711
12712/// ReplaceAllUsesWith - Modify anything using 'From' to use 'To' instead.
12713/// This can cause recursive merging of nodes in the DAG.
12714///
12715/// This version can replace From with any result values. To must match the
12716/// number and types of values returned by From.
12718 if (From->getNumValues() == 1) // Handle the simple case efficiently.
12719 return ReplaceAllUsesWith(SDValue(From, 0), To[0]);
12720
12721 for (unsigned i = 0, e = From->getNumValues(); i != e; ++i) {
12722 // Preserve Debug Info.
12723 transferDbgValues(SDValue(From, i), To[i]);
12724 // Preserve extra info.
12725 copyExtraInfo(From, To[i].getNode());
12726 }
12727
12728 // Iterate over just the existing users of From. See the comments in
12729 // the ReplaceAllUsesWith above.
12730 SDNode::use_iterator UI = From->use_begin(), UE = From->use_end();
12731 RAUWUpdateListener Listener(*this, UI, UE);
12732 while (UI != UE) {
12733 SDNode *User = UI->getUser();
12734
12735 // This node is about to morph, remove its old self from the CSE maps.
12736 RemoveNodeFromCSEMaps(User);
12737
12738 // A user can appear in a use list multiple times, and when this happens the
12739 // uses are usually next to each other in the list. To help reduce the
12740 // number of CSE and divergence recomputations, process all the uses of this
12741 // user that we can find this way.
12742 bool To_IsDivergent = false;
12743 do {
12744 SDUse &Use = *UI;
12745 const SDValue &ToOp = To[Use.getResNo()];
12746 ++UI;
12747 Use.set(ToOp);
12748 if (ToOp.getValueType() != MVT::Other)
12749 To_IsDivergent |= ToOp->isDivergent();
12750 } while (UI != UE && UI->getUser() == User);
12751
12752 if (To_IsDivergent != From->isDivergent())
12754
12755 // Now that we have modified User, add it back to the CSE maps. If it
12756 // already exists there, recursively merge the results together.
12757 AddModifiedNodeToCSEMaps(User);
12758 }
12759
12760 // If we just RAUW'd the root, take note.
12761 if (From == getRoot().getNode())
12762 setRoot(SDValue(To[getRoot().getResNo()]));
12763}
12764
12765/// ReplaceAllUsesOfValueWith - Replace any uses of From with To, leaving
12766/// uses of other values produced by From.getNode() alone. The Deleted
12767/// vector is handled the same way as for ReplaceAllUsesWith.
12769 // Handle the really simple, really trivial case efficiently.
12770 if (From == To) return;
12771
12772 // Handle the simple, trivial, case efficiently.
12773 if (From.getNode()->getNumValues() == 1) {
12774 ReplaceAllUsesWith(From, To);
12775 return;
12776 }
12777
12778 // Preserve Debug Info.
12779 transferDbgValues(From, To);
12780 copyExtraInfo(From.getNode(), To.getNode());
12781
12782 // Iterate over just the existing users of From. See the comments in
12783 // the ReplaceAllUsesWith above.
12784 SDNode::use_iterator UI = From.getNode()->use_begin(),
12785 UE = From.getNode()->use_end();
12786 RAUWUpdateListener Listener(*this, UI, UE);
12787 while (UI != UE) {
12788 SDNode *User = UI->getUser();
12789 bool UserRemovedFromCSEMaps = false;
12790
12791 // A user can appear in a use list multiple times, and when this
12792 // happens the uses are usually next to each other in the list.
12793 // To help reduce the number of CSE recomputations, process all
12794 // the uses of this user that we can find this way.
12795 do {
12796 SDUse &Use = *UI;
12797
12798 // Skip uses of different values from the same node.
12799 if (Use.getResNo() != From.getResNo()) {
12800 ++UI;
12801 continue;
12802 }
12803
12804 // If this node hasn't been modified yet, it's still in the CSE maps,
12805 // so remove its old self from the CSE maps.
12806 if (!UserRemovedFromCSEMaps) {
12807 RemoveNodeFromCSEMaps(User);
12808 UserRemovedFromCSEMaps = true;
12809 }
12810
12811 ++UI;
12812 Use.set(To);
12813 if (To->isDivergent() != From->isDivergent())
12815 } while (UI != UE && UI->getUser() == User);
12816 // We are iterating over all uses of the From node, so if a use
12817 // doesn't use the specific value, no changes are made.
12818 if (!UserRemovedFromCSEMaps)
12819 continue;
12820
12821 // Now that we have modified User, add it back to the CSE maps. If it
12822 // already exists there, recursively merge the results together.
12823 AddModifiedNodeToCSEMaps(User);
12824 }
12825
12826 // If we just RAUW'd the root, take note.
12827 if (From == getRoot())
12828 setRoot(To);
12829}
12830
12831namespace {
12832
12833/// UseMemo - This class is used by SelectionDAG::ReplaceAllUsesOfValuesWith
12834/// to record information about a use.
12835struct UseMemo {
12836 SDNode *User;
12837 unsigned Index;
12838 SDUse *Use;
12839};
12840
12841/// operator< - Sort Memos by User.
12842bool operator<(const UseMemo &L, const UseMemo &R) {
12843 return (intptr_t)L.User < (intptr_t)R.User;
12844}
12845
12846/// RAUOVWUpdateListener - Helper for ReplaceAllUsesOfValuesWith - When the node
12847/// pointed to by a UseMemo is deleted, set the User to nullptr to indicate that
12848/// the node already has been taken care of recursively.
12849class RAUOVWUpdateListener : public SelectionDAG::DAGUpdateListener {
12850 SmallVectorImpl<UseMemo> &Uses;
12851
12852 void NodeDeleted(SDNode *N, SDNode *E) override {
12853 for (UseMemo &Memo : Uses)
12854 if (Memo.User == N)
12855 Memo.User = nullptr;
12856 }
12857
12858public:
12859 RAUOVWUpdateListener(SelectionDAG &d, SmallVectorImpl<UseMemo> &uses)
12860 : SelectionDAG::DAGUpdateListener(d), Uses(uses) {}
12861};
12862
12863} // end anonymous namespace
12864
12865/// Return true if a glue output should propagate divergence information.
12867 switch (Node->getOpcode()) {
12868 case ISD::CopyFromReg:
12869 case ISD::CopyToReg:
12870 return false;
12871 default:
12872 return true;
12873 }
12874
12875 llvm_unreachable("covered opcode switch");
12876}
12877
12879 if (TLI->isSDNodeAlwaysUniform(N)) {
12880 assert(!TLI->isSDNodeSourceOfDivergence(N, FLI, UA) &&
12881 "Conflicting divergence information!");
12882 return false;
12883 }
12884 if (TLI->isSDNodeSourceOfDivergence(N, FLI, UA))
12885 return true;
12886 for (const auto &Op : N->ops()) {
12887 EVT VT = Op.getValueType();
12888
12889 // Skip Chain. It does not carry divergence.
12890 if (VT != MVT::Other && Op.getNode()->isDivergent() &&
12891 (VT != MVT::Glue || gluePropagatesDivergence(Op.getNode())))
12892 return true;
12893 }
12894 return false;
12895}
12896
12898 SmallVector<SDNode *, 16> Worklist(1, N);
12899 do {
12900 N = Worklist.pop_back_val();
12901 bool IsDivergent = calculateDivergence(N);
12902 if (N->SDNodeBits.IsDivergent != IsDivergent) {
12903 N->SDNodeBits.IsDivergent = IsDivergent;
12904 llvm::append_range(Worklist, N->users());
12905 }
12906 } while (!Worklist.empty());
12907}
12908
12909void SelectionDAG::CreateTopologicalOrder(std::vector<SDNode *> &Order) {
12911 Order.reserve(AllNodes.size());
12912 for (auto &N : allnodes()) {
12913 unsigned NOps = N.getNumOperands();
12914 Degree[&N] = NOps;
12915 if (0 == NOps)
12916 Order.push_back(&N);
12917 }
12918 for (size_t I = 0; I != Order.size(); ++I) {
12919 SDNode *N = Order[I];
12920 for (auto *U : N->users()) {
12921 unsigned &UnsortedOps = Degree[U];
12922 if (0 == --UnsortedOps)
12923 Order.push_back(U);
12924 }
12925 }
12926}
12927
12928#if !defined(NDEBUG) && LLVM_ENABLE_ABI_BREAKING_CHECKS
12929void SelectionDAG::VerifyDAGDivergence() {
12930 std::vector<SDNode *> TopoOrder;
12931 CreateTopologicalOrder(TopoOrder);
12932 for (auto *N : TopoOrder) {
12933 assert(calculateDivergence(N) == N->isDivergent() &&
12934 "Divergence bit inconsistency detected");
12935 }
12936}
12937#endif
12938
12939/// ReplaceAllUsesOfValuesWith - Replace any uses of From with To, leaving
12940/// uses of other values produced by From.getNode() alone. The same value
12941/// may appear in both the From and To list. The Deleted vector is
12942/// handled the same way as for ReplaceAllUsesWith.
12944 const SDValue *To,
12945 unsigned Num){
12946 // Handle the simple, trivial case efficiently.
12947 if (Num == 1)
12948 return ReplaceAllUsesOfValueWith(*From, *To);
12949
12950 transferDbgValues(*From, *To);
12951 copyExtraInfo(From->getNode(), To->getNode());
12952
12953 // Read up all the uses and make records of them. This helps
12954 // processing new uses that are introduced during the
12955 // replacement process.
12957 for (unsigned i = 0; i != Num; ++i) {
12958 unsigned FromResNo = From[i].getResNo();
12959 SDNode *FromNode = From[i].getNode();
12960 for (SDUse &Use : FromNode->uses()) {
12961 if (Use.getResNo() == FromResNo) {
12962 UseMemo Memo = {Use.getUser(), i, &Use};
12963 Uses.push_back(Memo);
12964 }
12965 }
12966 }
12967
12968 // Sort the uses, so that all the uses from a given User are together.
12970 RAUOVWUpdateListener Listener(*this, Uses);
12971
12972 for (unsigned UseIndex = 0, UseIndexEnd = Uses.size();
12973 UseIndex != UseIndexEnd; ) {
12974 // We know that this user uses some value of From. If it is the right
12975 // value, update it.
12976 SDNode *User = Uses[UseIndex].User;
12977 // If the node has been deleted by recursive CSE updates when updating
12978 // another node, then just skip this entry.
12979 if (User == nullptr) {
12980 ++UseIndex;
12981 continue;
12982 }
12983
12984 // This node is about to morph, remove its old self from the CSE maps.
12985 RemoveNodeFromCSEMaps(User);
12986
12987 // The Uses array is sorted, so all the uses for a given User
12988 // are next to each other in the list.
12989 // To help reduce the number of CSE recomputations, process all
12990 // the uses of this user that we can find this way.
12991 do {
12992 unsigned i = Uses[UseIndex].Index;
12993 SDUse &Use = *Uses[UseIndex].Use;
12994 ++UseIndex;
12995
12996 Use.set(To[i]);
12997 } while (UseIndex != UseIndexEnd && Uses[UseIndex].User == User);
12998
12999 // Now that we have modified User, add it back to the CSE maps. If it
13000 // already exists there, recursively merge the results together.
13001 AddModifiedNodeToCSEMaps(User);
13002 }
13003}
13004
13005/// AssignTopologicalOrder - Assign a unique node id for each node in the DAG
13006/// based on their topological order. It returns the maximum id and a vector
13007/// of the SDNodes* in assigned order by reference.
13009 unsigned DAGSize = 0;
13010
13011 // SortedPos tracks the progress of the algorithm. Nodes before it are
13012 // sorted, nodes after it are unsorted. When the algorithm completes
13013 // it is at the end of the list.
13014 allnodes_iterator SortedPos = allnodes_begin();
13015
13016 // Visit all the nodes. Move nodes with no operands to the front of
13017 // the list immediately. Annotate nodes that do have operands with their
13018 // operand count. Before we do this, the Node Id fields of the nodes
13019 // may contain arbitrary values. After, the Node Id fields for nodes
13020 // before SortedPos will contain the topological sort index, and the
13021 // Node Id fields for nodes At SortedPos and after will contain the
13022 // count of outstanding operands.
13024 checkForCycles(&N, this);
13025 unsigned Degree = N.getNumOperands();
13026 if (Degree == 0) {
13027 // A node with no uses, add it to the result array immediately.
13028 N.setNodeId(DAGSize++);
13029 allnodes_iterator Q(&N);
13030 if (Q != SortedPos)
13031 SortedPos = AllNodes.insert(SortedPos, AllNodes.remove(Q));
13032 assert(SortedPos != AllNodes.end() && "Overran node list");
13033 ++SortedPos;
13034 } else {
13035 // Temporarily use the Node Id as scratch space for the degree count.
13036 N.setNodeId(Degree);
13037 }
13038 }
13039
13040 // Visit all the nodes. As we iterate, move nodes into sorted order,
13041 // such that by the time the end is reached all nodes will be sorted.
13042 for (SDNode &Node : allnodes()) {
13043 SDNode *N = &Node;
13044 checkForCycles(N, this);
13045 // N is in sorted position, so all its uses have one less operand
13046 // that needs to be sorted.
13047 for (SDNode *P : N->users()) {
13048 unsigned Degree = P->getNodeId();
13049 assert(Degree != 0 && "Invalid node degree");
13050 --Degree;
13051 if (Degree == 0) {
13052 // All of P's operands are sorted, so P may sorted now.
13053 P->setNodeId(DAGSize++);
13054 if (P->getIterator() != SortedPos)
13055 SortedPos = AllNodes.insert(SortedPos, AllNodes.remove(P));
13056 assert(SortedPos != AllNodes.end() && "Overran node list");
13057 ++SortedPos;
13058 } else {
13059 // Update P's outstanding operand count.
13060 P->setNodeId(Degree);
13061 }
13062 }
13063 if (Node.getIterator() == SortedPos) {
13064#ifndef NDEBUG
13066 SDNode *S = &*++I;
13067 dbgs() << "Overran sorted position:\n";
13068 S->dumprFull(this); dbgs() << "\n";
13069 dbgs() << "Checking if this is due to cycles\n";
13070 checkForCycles(this, true);
13071#endif
13072 llvm_unreachable(nullptr);
13073 }
13074 }
13075
13076 assert(SortedPos == AllNodes.end() &&
13077 "Topological sort incomplete!");
13078 assert(AllNodes.front().getOpcode() == ISD::EntryToken &&
13079 "First node in topological sort is not the entry token!");
13080 assert(AllNodes.front().getNodeId() == 0 &&
13081 "First node in topological sort has non-zero id!");
13082 assert(AllNodes.front().getNumOperands() == 0 &&
13083 "First node in topological sort has operands!");
13084 assert(AllNodes.back().getNodeId() == (int)DAGSize-1 &&
13085 "Last node in topologic sort has unexpected id!");
13086 assert(AllNodes.back().use_empty() &&
13087 "Last node in topologic sort has users!");
13088 assert(DAGSize == allnodes_size() && "Node count mismatch!");
13089 return DAGSize;
13090}
13091
13093 SmallVectorImpl<const SDNode *> &SortedNodes) const {
13094 SortedNodes.clear();
13095 // Node -> remaining number of outstanding operands.
13096 DenseMap<const SDNode *, unsigned> RemainingOperands;
13097
13098 // Put nodes without any operands into SortedNodes first.
13099 for (const SDNode &N : allnodes()) {
13100 checkForCycles(&N, this);
13101 unsigned NumOperands = N.getNumOperands();
13102 if (NumOperands == 0)
13103 SortedNodes.push_back(&N);
13104 else
13105 // Record their total number of outstanding operands.
13106 RemainingOperands[&N] = NumOperands;
13107 }
13108
13109 // A node is pushed into SortedNodes when all of its operands (predecessors in
13110 // the graph) are also in SortedNodes.
13111 for (unsigned i = 0U; i < SortedNodes.size(); ++i) {
13112 const SDNode *N = SortedNodes[i];
13113 for (const SDNode *U : N->users()) {
13114 // HandleSDNode is never part of a DAG and therefore has no entry in
13115 // RemainingOperands.
13116 if (U->getOpcode() == ISD::HANDLENODE)
13117 continue;
13118 unsigned &NumRemOperands = RemainingOperands[U];
13119 assert(NumRemOperands && "Invalid number of remaining operands");
13120 --NumRemOperands;
13121 if (!NumRemOperands)
13122 SortedNodes.push_back(U);
13123 }
13124 }
13125
13126 assert(SortedNodes.size() == AllNodes.size() && "Node count mismatch");
13127 assert(SortedNodes.front()->getOpcode() == ISD::EntryToken &&
13128 "First node in topological sort is not the entry token");
13129 assert(SortedNodes.front()->getNumOperands() == 0 &&
13130 "First node in topological sort has operands");
13131}
13132
13133/// AddDbgValue - Add a dbg_value SDNode. If SD is non-null that means the
13134/// value is produced by SD.
13135void SelectionDAG::AddDbgValue(SDDbgValue *DB, bool isParameter) {
13136 for (SDNode *SD : DB->getSDNodes()) {
13137 if (!SD)
13138 continue;
13139 assert(DbgInfo->getSDDbgValues(SD).empty() || SD->getHasDebugValue());
13140 SD->setHasDebugValue(true);
13141 }
13142 DbgInfo->add(DB, isParameter);
13143}
13144
13145void SelectionDAG::AddDbgLabel(SDDbgLabel *DB) { DbgInfo->add(DB); }
13146
13148 SDValue NewMemOpChain) {
13149 assert(isa<MemSDNode>(NewMemOpChain) && "Expected a memop node");
13150 assert(NewMemOpChain.getValueType() == MVT::Other && "Expected a token VT");
13151 // The new memory operation must have the same position as the old load in
13152 // terms of memory dependency. Create a TokenFactor for the old load and new
13153 // memory operation and update uses of the old load's output chain to use that
13154 // TokenFactor.
13155 if (OldChain == NewMemOpChain || OldChain.use_empty())
13156 return NewMemOpChain;
13157
13158 SDValue TokenFactor = getNode(ISD::TokenFactor, SDLoc(OldChain), MVT::Other,
13159 OldChain, NewMemOpChain);
13160 ReplaceAllUsesOfValueWith(OldChain, TokenFactor);
13161 UpdateNodeOperands(TokenFactor.getNode(), OldChain, NewMemOpChain);
13162 return TokenFactor;
13163}
13164
13166 SDValue NewMemOp) {
13167 assert(isa<MemSDNode>(NewMemOp.getNode()) && "Expected a memop node");
13168 SDValue OldChain = SDValue(OldLoad, 1);
13169 SDValue NewMemOpChain = NewMemOp.getValue(1);
13170 return makeEquivalentMemoryOrdering(OldChain, NewMemOpChain);
13171}
13172
13174 Function **OutFunction) {
13175 assert(isa<ExternalSymbolSDNode>(Op) && "Node should be an ExternalSymbol");
13176
13177 auto *Symbol = cast<ExternalSymbolSDNode>(Op)->getSymbol();
13178 auto *Module = MF->getFunction().getParent();
13179 auto *Function = Module->getFunction(Symbol);
13180
13181 if (OutFunction != nullptr)
13182 *OutFunction = Function;
13183
13184 if (Function != nullptr) {
13185 auto PtrTy = TLI->getPointerTy(getDataLayout(), Function->getAddressSpace());
13186 return getGlobalAddress(Function, SDLoc(Op), PtrTy);
13187 }
13188
13189 std::string ErrorStr;
13190 raw_string_ostream ErrorFormatter(ErrorStr);
13191 ErrorFormatter << "Undefined external symbol ";
13192 ErrorFormatter << '"' << Symbol << '"';
13193 report_fatal_error(Twine(ErrorStr));
13194}
13195
13196//===----------------------------------------------------------------------===//
13197// SDNode Class
13198//===----------------------------------------------------------------------===//
13199
13202 return Const != nullptr && Const->isZero();
13203}
13204
13206 return V.isUndef() || isNullConstant(V);
13207}
13208
13211 return Const != nullptr && Const->isZero() && !Const->isNegative();
13212}
13213
13216 return Const != nullptr && Const->isAllOnes();
13217}
13218
13221 return Const != nullptr && Const->isOne();
13222}
13223
13226 return Const != nullptr && Const->isMinSignedValue();
13227}
13228
13229bool llvm::isNeutralConstant(unsigned Opcode, SDNodeFlags Flags, SDValue V,
13230 unsigned OperandNo) {
13231 // NOTE: The cases should match with IR's ConstantExpr::getBinOpIdentity().
13232 // TODO: Target-specific opcodes could be added.
13233 if (auto *ConstV = isConstOrConstSplat(V, /*AllowUndefs*/ false,
13234 /*AllowTruncation*/ true)) {
13235 APInt Const = ConstV->getAPIntValue().trunc(V.getScalarValueSizeInBits());
13236 switch (Opcode) {
13237 case ISD::ADD:
13238 case ISD::OR:
13239 case ISD::XOR:
13240 case ISD::UMAX:
13241 return Const.isZero();
13242 case ISD::MUL:
13243 return Const.isOne();
13244 case ISD::AND:
13245 case ISD::UMIN:
13246 return Const.isAllOnes();
13247 case ISD::SMAX:
13248 return Const.isMinSignedValue();
13249 case ISD::SMIN:
13250 return Const.isMaxSignedValue();
13251 case ISD::SUB:
13252 case ISD::SHL:
13253 case ISD::SRA:
13254 case ISD::SRL:
13255 return OperandNo == 1 && Const.isZero();
13256 case ISD::UDIV:
13257 case ISD::SDIV:
13258 return OperandNo == 1 && Const.isOne();
13259 }
13260 } else if (auto *ConstFP = isConstOrConstSplatFP(V)) {
13261 switch (Opcode) {
13262 case ISD::FADD:
13263 return ConstFP->isZero() &&
13264 (Flags.hasNoSignedZeros() || ConstFP->isNegative());
13265 case ISD::FSUB:
13266 return OperandNo == 1 && ConstFP->isZero() &&
13267 (Flags.hasNoSignedZeros() || !ConstFP->isNegative());
13268 case ISD::FMUL:
13269 return ConstFP->isExactlyValue(1.0);
13270 case ISD::FDIV:
13271 return OperandNo == 1 && ConstFP->isExactlyValue(1.0);
13272 case ISD::FMINNUM:
13273 case ISD::FMAXNUM: {
13274 // Neutral element for fminnum is NaN, Inf or FLT_MAX, depending on FMF.
13275 EVT VT = V.getValueType();
13276 const fltSemantics &Semantics = VT.getFltSemantics();
13277 APFloat NeutralAF = !Flags.hasNoNaNs()
13278 ? APFloat::getQNaN(Semantics)
13279 : !Flags.hasNoInfs()
13280 ? APFloat::getInf(Semantics)
13281 : APFloat::getLargest(Semantics);
13282 if (Opcode == ISD::FMAXNUM)
13283 NeutralAF.changeSign();
13284
13285 return ConstFP->isExactlyValue(NeutralAF);
13286 }
13287 }
13288 }
13289 return false;
13290}
13291
13293 while (V.getOpcode() == ISD::BITCAST)
13294 V = V.getOperand(0);
13295 return V;
13296}
13297
13299 while (V.getOpcode() == ISD::BITCAST && V.getOperand(0).hasOneUse())
13300 V = V.getOperand(0);
13301 return V;
13302}
13303
13305 while (V.getOpcode() == ISD::EXTRACT_SUBVECTOR)
13306 V = V.getOperand(0);
13307 return V;
13308}
13309
13311 while (V.getOpcode() == ISD::INSERT_VECTOR_ELT) {
13312 SDValue InVec = V.getOperand(0);
13313 SDValue EltNo = V.getOperand(2);
13314 EVT VT = InVec.getValueType();
13315 auto *IndexC = dyn_cast<ConstantSDNode>(EltNo);
13316 if (IndexC && VT.isFixedLengthVector() &&
13317 IndexC->getAPIntValue().ult(VT.getVectorNumElements()) &&
13318 !DemandedElts[IndexC->getZExtValue()]) {
13319 V = InVec;
13320 continue;
13321 }
13322 break;
13323 }
13324 return V;
13325}
13326
13328 while (V.getOpcode() == ISD::TRUNCATE)
13329 V = V.getOperand(0);
13330 return V;
13331}
13332
13333bool llvm::isBitwiseNot(SDValue V, bool AllowUndefs) {
13334 if (V.getOpcode() != ISD::XOR)
13335 return false;
13336 V = peekThroughBitcasts(V.getOperand(1));
13337 unsigned NumBits = V.getScalarValueSizeInBits();
13338 ConstantSDNode *C =
13339 isConstOrConstSplat(V, AllowUndefs, /*AllowTruncation*/ true);
13340 return C && (C->getAPIntValue().countr_one() >= NumBits);
13341}
13342
13344 bool AllowTruncation) {
13345 EVT VT = N.getValueType();
13346 APInt DemandedElts = VT.isFixedLengthVector()
13348 : APInt(1, 1);
13349 return isConstOrConstSplat(N, DemandedElts, AllowUndefs, AllowTruncation);
13350}
13351
13353 bool AllowUndefs,
13354 bool AllowTruncation) {
13356 return CN;
13357
13358 // SplatVectors can truncate their operands. Ignore that case here unless
13359 // AllowTruncation is set.
13360 if (N->getOpcode() == ISD::SPLAT_VECTOR) {
13361 EVT VecEltVT = N->getValueType(0).getVectorElementType();
13362 if (auto *CN = dyn_cast<ConstantSDNode>(N->getOperand(0))) {
13363 EVT CVT = CN->getValueType(0);
13364 assert(CVT.bitsGE(VecEltVT) && "Illegal splat_vector element extension");
13365 if (AllowTruncation || CVT == VecEltVT)
13366 return CN;
13367 }
13368 }
13369
13371 BitVector UndefElements;
13372 ConstantSDNode *CN = BV->getConstantSplatNode(DemandedElts, &UndefElements);
13373
13374 // BuildVectors can truncate their operands. Ignore that case here unless
13375 // AllowTruncation is set.
13376 // TODO: Look into whether we should allow UndefElements in non-DemandedElts
13377 if (CN && (UndefElements.none() || AllowUndefs)) {
13378 EVT CVT = CN->getValueType(0);
13379 EVT NSVT = N.getValueType().getScalarType();
13380 assert(CVT.bitsGE(NSVT) && "Illegal build vector element extension");
13381 if (AllowTruncation || (CVT == NSVT))
13382 return CN;
13383 }
13384 }
13385
13386 return nullptr;
13387}
13388
13390 EVT VT = N.getValueType();
13391 APInt DemandedElts = VT.isFixedLengthVector()
13393 : APInt(1, 1);
13394 return isConstOrConstSplatFP(N, DemandedElts, AllowUndefs);
13395}
13396
13398 const APInt &DemandedElts,
13399 bool AllowUndefs) {
13401 return CN;
13402
13404 BitVector UndefElements;
13405 ConstantFPSDNode *CN =
13406 BV->getConstantFPSplatNode(DemandedElts, &UndefElements);
13407 // TODO: Look into whether we should allow UndefElements in non-DemandedElts
13408 if (CN && (UndefElements.none() || AllowUndefs))
13409 return CN;
13410 }
13411
13412 if (N.getOpcode() == ISD::SPLAT_VECTOR)
13413 if (ConstantFPSDNode *CN = dyn_cast<ConstantFPSDNode>(N.getOperand(0)))
13414 return CN;
13415
13416 return nullptr;
13417}
13418
13419bool llvm::isNullOrNullSplat(SDValue N, bool AllowUndefs) {
13420 // TODO: may want to use peekThroughBitcast() here.
13421 ConstantSDNode *C =
13422 isConstOrConstSplat(N, AllowUndefs, /*AllowTruncation=*/true);
13423 return C && C->isZero();
13424}
13425
13426bool llvm::isOneOrOneSplat(SDValue N, bool AllowUndefs) {
13427 ConstantSDNode *C =
13428 isConstOrConstSplat(N, AllowUndefs, /*AllowTruncation*/ true);
13429 return C && C->isOne();
13430}
13431
13432bool llvm::isOneOrOneSplatFP(SDValue N, bool AllowUndefs) {
13433 ConstantFPSDNode *C = isConstOrConstSplatFP(N, AllowUndefs);
13434 return C && C->isExactlyValue(1.0);
13435}
13436
13437bool llvm::isAllOnesOrAllOnesSplat(SDValue N, bool AllowUndefs) {
13439 unsigned BitWidth = N.getScalarValueSizeInBits();
13440 ConstantSDNode *C = isConstOrConstSplat(N, AllowUndefs);
13441 return C && C->isAllOnes() && C->getValueSizeInBits(0) == BitWidth;
13442}
13443
13444bool llvm::isOnesOrOnesSplat(SDValue N, bool AllowUndefs) {
13445 ConstantSDNode *C = isConstOrConstSplat(N, AllowUndefs);
13446 return C && APInt::isSameValue(C->getAPIntValue(),
13447 APInt(C->getAPIntValue().getBitWidth(), 1));
13448}
13449
13450bool llvm::isZeroOrZeroSplat(SDValue N, bool AllowUndefs) {
13452 ConstantSDNode *C = isConstOrConstSplat(N, AllowUndefs, true);
13453 return C && C->isZero();
13454}
13455
13456bool llvm::isZeroOrZeroSplatFP(SDValue N, bool AllowUndefs) {
13457 ConstantFPSDNode *C = isConstOrConstSplatFP(N, AllowUndefs);
13458 return C && C->isZero();
13459}
13460
13464
13466 unsigned Opc, unsigned Order, const DebugLoc &dl, SDVTList VTs, EVT memvt,
13468 : SDNode(Opc, Order, dl, VTs), MemoryVT(memvt), MemRefs(memrefs) {
13469 bool IsVolatile = false;
13470 bool IsNonTemporal = false;
13471 bool IsDereferenceable = true;
13472 bool IsInvariant = true;
13473 for (const MachineMemOperand *MMO : memoperands()) {
13474 IsVolatile |= MMO->isVolatile();
13475 IsNonTemporal |= MMO->isNonTemporal();
13476 IsDereferenceable &= MMO->isDereferenceable();
13477 IsInvariant &= MMO->isInvariant();
13478 }
13479 MemSDNodeBits.IsVolatile = IsVolatile;
13480 MemSDNodeBits.IsNonTemporal = IsNonTemporal;
13481 MemSDNodeBits.IsDereferenceable = IsDereferenceable;
13482 MemSDNodeBits.IsInvariant = IsInvariant;
13483
13484 // For the single-MMO case, we check here that the size of the memory operand
13485 // fits within the size of the MMO. This is because the MMO might indicate
13486 // only a possible address range instead of specifying the affected memory
13487 // addresses precisely.
13490 getMemOperand()->getSize().getValue())) &&
13491 "Size mismatch!");
13492}
13493
13494/// Profile - Gather unique data for the node.
13495///
13497 AddNodeIDNode(ID, this);
13498}
13499
13500namespace {
13501
13502 struct EVTArray {
13503 std::vector<EVT> VTs;
13504
13505 EVTArray() {
13506 VTs.reserve(MVT::VALUETYPE_SIZE);
13507 for (unsigned i = 0; i < MVT::VALUETYPE_SIZE; ++i)
13508 VTs.push_back(MVT((MVT::SimpleValueType)i));
13509 }
13510 };
13511
13512} // end anonymous namespace
13513
13514/// getValueTypeList - Return a pointer to the specified value type.
13515///
13516const EVT *SDNode::getValueTypeList(MVT VT) {
13517 static EVTArray SimpleVTArray;
13518
13519 assert(VT < MVT::VALUETYPE_SIZE && "Value type out of range!");
13520 return &SimpleVTArray.VTs[VT.SimpleTy];
13521}
13522
13523/// hasAnyUseOfValue - Return true if there are any use of the indicated
13524/// value. This method ignores uses of other values defined by this operation.
13525bool SDNode::hasAnyUseOfValue(unsigned Value) const {
13526 assert(Value < getNumValues() && "Bad value!");
13527
13528 for (SDUse &U : uses())
13529 if (U.getResNo() == Value)
13530 return true;
13531
13532 return false;
13533}
13534
13535/// isOnlyUserOf - Return true if this node is the only use of N.
13536bool SDNode::isOnlyUserOf(const SDNode *N) const {
13537 bool Seen = false;
13538 for (const SDNode *User : N->users()) {
13539 if (User == this)
13540 Seen = true;
13541 else
13542 return false;
13543 }
13544
13545 return Seen;
13546}
13547
13548/// Return true if the only users of N are contained in Nodes.
13550 bool Seen = false;
13551 for (const SDNode *User : N->users()) {
13552 if (llvm::is_contained(Nodes, User))
13553 Seen = true;
13554 else
13555 return false;
13556 }
13557
13558 return Seen;
13559}
13560
13561/// Return true if the referenced return value is an operand of N.
13562bool SDValue::isOperandOf(const SDNode *N) const {
13563 return is_contained(N->op_values(), *this);
13564}
13565
13566bool SDNode::isOperandOf(const SDNode *N) const {
13567 return any_of(N->op_values(),
13568 [this](SDValue Op) { return this == Op.getNode(); });
13569}
13570
13571/// reachesChainWithoutSideEffects - Return true if this operand (which must
13572/// be a chain) reaches the specified operand without crossing any
13573/// side-effecting instructions on any chain path. In practice, this looks
13574/// through token factors and non-volatile loads. In order to remain efficient,
13575/// this only looks a couple of nodes in, it does not do an exhaustive search.
13576///
13577/// Note that we only need to examine chains when we're searching for
13578/// side-effects; SelectionDAG requires that all side-effects are represented
13579/// by chains, even if another operand would force a specific ordering. This
13580/// constraint is necessary to allow transformations like splitting loads.
13582 unsigned Depth) const {
13583 if (*this == Dest) return true;
13584
13585 // Don't search too deeply, we just want to be able to see through
13586 // TokenFactor's etc.
13587 if (Depth == 0) return false;
13588
13589 // If this is a token factor, all inputs to the TF happen in parallel.
13590 if (getOpcode() == ISD::TokenFactor) {
13591 // First, try a shallow search.
13592 if (is_contained((*this)->ops(), Dest)) {
13593 // We found the chain we want as an operand of this TokenFactor.
13594 // Essentially, we reach the chain without side-effects if we could
13595 // serialize the TokenFactor into a simple chain of operations with
13596 // Dest as the last operation. This is automatically true if the
13597 // chain has one use: there are no other ordering constraints.
13598 // If the chain has more than one use, we give up: some other
13599 // use of Dest might force a side-effect between Dest and the current
13600 // node.
13601 if (Dest.hasOneUse())
13602 return true;
13603 }
13604 // Next, try a deep search: check whether every operand of the TokenFactor
13605 // reaches Dest.
13606 return llvm::all_of((*this)->ops(), [=](SDValue Op) {
13607 return Op.reachesChainWithoutSideEffects(Dest, Depth - 1);
13608 });
13609 }
13610
13611 // Loads don't have side effects, look through them.
13612 if (LoadSDNode *Ld = dyn_cast<LoadSDNode>(*this)) {
13613 if (Ld->isUnordered())
13614 return Ld->getChain().reachesChainWithoutSideEffects(Dest, Depth-1);
13615 }
13616 return false;
13617}
13618
13619bool SDNode::hasPredecessor(const SDNode *N) const {
13622 Worklist.push_back(this);
13623 return hasPredecessorHelper(N, Visited, Worklist);
13624}
13625
13627 this->Flags &= Flags;
13628}
13629
13630SDValue
13632 ArrayRef<ISD::NodeType> CandidateBinOps,
13633 bool AllowPartials) {
13634 // The pattern must end in an extract from index 0.
13635 if (Extract->getOpcode() != ISD::EXTRACT_VECTOR_ELT ||
13636 !isNullConstant(Extract->getOperand(1)))
13637 return SDValue();
13638
13639 // Match against one of the candidate binary ops.
13640 SDValue Op = Extract->getOperand(0);
13641 if (llvm::none_of(CandidateBinOps, [Op](ISD::NodeType BinOp) {
13642 return Op.getOpcode() == unsigned(BinOp);
13643 }))
13644 return SDValue();
13645
13646 // Floating-point reductions may require relaxed constraints on the final step
13647 // of the reduction because they may reorder intermediate operations.
13648 unsigned CandidateBinOp = Op.getOpcode();
13649 if (Op.getValueType().isFloatingPoint()) {
13650 SDNodeFlags Flags = Op->getFlags();
13651 switch (CandidateBinOp) {
13652 case ISD::FADD:
13653 if (!Flags.hasNoSignedZeros() || !Flags.hasAllowReassociation())
13654 return SDValue();
13655 break;
13656 default:
13657 llvm_unreachable("Unhandled FP opcode for binop reduction");
13658 }
13659 }
13660
13661 // Matching failed - attempt to see if we did enough stages that a partial
13662 // reduction from a subvector is possible.
13663 auto PartialReduction = [&](SDValue Op, unsigned NumSubElts) {
13664 if (!AllowPartials || !Op)
13665 return SDValue();
13666 EVT OpVT = Op.getValueType();
13667 EVT OpSVT = OpVT.getScalarType();
13668 EVT SubVT = EVT::getVectorVT(*getContext(), OpSVT, NumSubElts);
13669 if (!TLI->isExtractSubvectorCheap(SubVT, OpVT, 0))
13670 return SDValue();
13671 BinOp = (ISD::NodeType)CandidateBinOp;
13672 return getExtractSubvector(SDLoc(Op), SubVT, Op, 0);
13673 };
13674
13675 // At each stage, we're looking for something that looks like:
13676 // %s = shufflevector <8 x i32> %op, <8 x i32> undef,
13677 // <8 x i32> <i32 2, i32 3, i32 undef, i32 undef,
13678 // i32 undef, i32 undef, i32 undef, i32 undef>
13679 // %a = binop <8 x i32> %op, %s
13680 // Where the mask changes according to the stage. E.g. for a 3-stage pyramid,
13681 // we expect something like:
13682 // <4,5,6,7,u,u,u,u>
13683 // <2,3,u,u,u,u,u,u>
13684 // <1,u,u,u,u,u,u,u>
13685 // While a partial reduction match would be:
13686 // <2,3,u,u,u,u,u,u>
13687 // <1,u,u,u,u,u,u,u>
13688 unsigned Stages = Log2_32(Op.getValueType().getVectorNumElements());
13689 SDValue PrevOp;
13690 for (unsigned i = 0; i < Stages; ++i) {
13691 unsigned MaskEnd = (1 << i);
13692
13693 if (Op.getOpcode() != CandidateBinOp)
13694 return PartialReduction(PrevOp, MaskEnd);
13695
13696 SDValue Op0 = Op.getOperand(0);
13697 SDValue Op1 = Op.getOperand(1);
13698
13700 if (Shuffle) {
13701 Op = Op1;
13702 } else {
13703 Shuffle = dyn_cast<ShuffleVectorSDNode>(Op1);
13704 Op = Op0;
13705 }
13706
13707 // The first operand of the shuffle should be the same as the other operand
13708 // of the binop.
13709 if (!Shuffle || Shuffle->getOperand(0) != Op)
13710 return PartialReduction(PrevOp, MaskEnd);
13711
13712 // Verify the shuffle has the expected (at this stage of the pyramid) mask.
13713 for (int Index = 0; Index < (int)MaskEnd; ++Index)
13714 if (Shuffle->getMaskElt(Index) != (int)(MaskEnd + Index))
13715 return PartialReduction(PrevOp, MaskEnd);
13716
13717 PrevOp = Op;
13718 }
13719
13720 // Handle subvector reductions, which tend to appear after the shuffle
13721 // reduction stages.
13722 while (Op.getOpcode() == CandidateBinOp) {
13723 unsigned NumElts = Op.getValueType().getVectorNumElements();
13724 SDValue Op0 = Op.getOperand(0);
13725 SDValue Op1 = Op.getOperand(1);
13726 if (Op0.getOpcode() != ISD::EXTRACT_SUBVECTOR ||
13728 Op0.getOperand(0) != Op1.getOperand(0))
13729 break;
13730 SDValue Src = Op0.getOperand(0);
13731 unsigned NumSrcElts = Src.getValueType().getVectorNumElements();
13732 if (NumSrcElts != (2 * NumElts))
13733 break;
13734 if (!(Op0.getConstantOperandAPInt(1) == 0 &&
13735 Op1.getConstantOperandAPInt(1) == NumElts) &&
13736 !(Op1.getConstantOperandAPInt(1) == 0 &&
13737 Op0.getConstantOperandAPInt(1) == NumElts))
13738 break;
13739 Op = Src;
13740 }
13741
13742 BinOp = (ISD::NodeType)CandidateBinOp;
13743 return Op;
13744}
13745
13747 EVT VT = N->getValueType(0);
13748 EVT EltVT = VT.getVectorElementType();
13749 unsigned NE = VT.getVectorNumElements();
13750
13751 SDLoc dl(N);
13752
13753 // If ResNE is 0, fully unroll the vector op.
13754 if (ResNE == 0)
13755 ResNE = NE;
13756 else if (NE > ResNE)
13757 NE = ResNE;
13758
13759 if (N->getNumValues() == 2) {
13760 SmallVector<SDValue, 8> Scalars0, Scalars1;
13761 SmallVector<SDValue, 4> Operands(N->getNumOperands());
13762 EVT VT1 = N->getValueType(1);
13763 EVT EltVT1 = VT1.getVectorElementType();
13764
13765 unsigned i;
13766 for (i = 0; i != NE; ++i) {
13767 for (unsigned j = 0, e = N->getNumOperands(); j != e; ++j) {
13768 SDValue Operand = N->getOperand(j);
13769 EVT OperandVT = Operand.getValueType();
13770
13771 // A vector operand; extract a single element.
13772 EVT OperandEltVT = OperandVT.getVectorElementType();
13773 Operands[j] = getExtractVectorElt(dl, OperandEltVT, Operand, i);
13774 }
13775
13776 SDValue EltOp = getNode(N->getOpcode(), dl, {EltVT, EltVT1}, Operands);
13777 Scalars0.push_back(EltOp);
13778 Scalars1.push_back(EltOp.getValue(1));
13779 }
13780
13781 for (; i < ResNE; ++i) {
13782 Scalars0.push_back(getUNDEF(EltVT));
13783 Scalars1.push_back(getUNDEF(EltVT1));
13784 }
13785
13786 EVT VecVT = EVT::getVectorVT(*getContext(), EltVT, ResNE);
13787 EVT VecVT1 = EVT::getVectorVT(*getContext(), EltVT1, ResNE);
13788 SDValue Vec0 = getBuildVector(VecVT, dl, Scalars0);
13789 SDValue Vec1 = getBuildVector(VecVT1, dl, Scalars1);
13790 return getMergeValues({Vec0, Vec1}, dl);
13791 }
13792
13793 assert(N->getNumValues() == 1 &&
13794 "Can't unroll a vector with multiple results!");
13795
13797 SmallVector<SDValue, 4> Operands(N->getNumOperands());
13798
13799 unsigned i;
13800 for (i= 0; i != NE; ++i) {
13801 for (unsigned j = 0, e = N->getNumOperands(); j != e; ++j) {
13802 SDValue Operand = N->getOperand(j);
13803 EVT OperandVT = Operand.getValueType();
13804 if (OperandVT.isVector()) {
13805 // A vector operand; extract a single element.
13806 EVT OperandEltVT = OperandVT.getVectorElementType();
13807 Operands[j] = getExtractVectorElt(dl, OperandEltVT, Operand, i);
13808 } else {
13809 // A scalar operand; just use it as is.
13810 Operands[j] = Operand;
13811 }
13812 }
13813
13814 switch (N->getOpcode()) {
13815 default: {
13816 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT, Operands,
13817 N->getFlags()));
13818 break;
13819 }
13820 case ISD::VSELECT:
13821 Scalars.push_back(getNode(ISD::SELECT, dl, EltVT, Operands));
13822 break;
13823 case ISD::SHL:
13824 case ISD::SRA:
13825 case ISD::SRL:
13826 case ISD::ROTL:
13827 case ISD::ROTR:
13828 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT, Operands[0],
13829 getShiftAmountOperand(Operands[0].getValueType(),
13830 Operands[1])));
13831 break;
13833 EVT ExtVT = cast<VTSDNode>(Operands[1])->getVT().getVectorElementType();
13834 Scalars.push_back(getNode(N->getOpcode(), dl, EltVT,
13835 Operands[0],
13836 getValueType(ExtVT)));
13837 break;
13838 }
13839 case ISD::ADDRSPACECAST: {
13840 const auto *ASC = cast<AddrSpaceCastSDNode>(N);
13841 Scalars.push_back(getAddrSpaceCast(dl, EltVT, Operands[0],
13842 ASC->getSrcAddressSpace(),
13843 ASC->getDestAddressSpace()));
13844 break;
13845 }
13846 }
13847 }
13848
13849 for (; i < ResNE; ++i)
13850 Scalars.push_back(getUNDEF(EltVT));
13851
13852 EVT VecVT = EVT::getVectorVT(*getContext(), EltVT, ResNE);
13853 return getBuildVector(VecVT, dl, Scalars);
13854}
13855
13856std::pair<SDValue, SDValue> SelectionDAG::UnrollVectorOverflowOp(
13857 SDNode *N, unsigned ResNE) {
13858 unsigned Opcode = N->getOpcode();
13859 assert((Opcode == ISD::UADDO || Opcode == ISD::SADDO ||
13860 Opcode == ISD::USUBO || Opcode == ISD::SSUBO ||
13861 Opcode == ISD::UMULO || Opcode == ISD::SMULO) &&
13862 "Expected an overflow opcode");
13863
13864 EVT ResVT = N->getValueType(0);
13865 EVT OvVT = N->getValueType(1);
13866 EVT ResEltVT = ResVT.getVectorElementType();
13867 EVT OvEltVT = OvVT.getVectorElementType();
13868 SDLoc dl(N);
13869
13870 // If ResNE is 0, fully unroll the vector op.
13871 unsigned NE = ResVT.getVectorNumElements();
13872 if (ResNE == 0)
13873 ResNE = NE;
13874 else if (NE > ResNE)
13875 NE = ResNE;
13876
13877 SmallVector<SDValue, 8> LHSScalars;
13878 SmallVector<SDValue, 8> RHSScalars;
13879 ExtractVectorElements(N->getOperand(0), LHSScalars, 0, NE);
13880 ExtractVectorElements(N->getOperand(1), RHSScalars, 0, NE);
13881
13882 EVT SVT = TLI->getSetCCResultType(getDataLayout(), *getContext(), ResEltVT);
13883 SDVTList VTs = getVTList(ResEltVT, SVT);
13884 SmallVector<SDValue, 8> ResScalars;
13885 SmallVector<SDValue, 8> OvScalars;
13886 for (unsigned i = 0; i < NE; ++i) {
13887 SDValue Res = getNode(Opcode, dl, VTs, LHSScalars[i], RHSScalars[i]);
13888 SDValue Ov =
13889 getSelect(dl, OvEltVT, Res.getValue(1),
13890 getBoolConstant(true, dl, OvEltVT, ResVT),
13891 getConstant(0, dl, OvEltVT));
13892
13893 ResScalars.push_back(Res);
13894 OvScalars.push_back(Ov);
13895 }
13896
13897 ResScalars.append(ResNE - NE, getUNDEF(ResEltVT));
13898 OvScalars.append(ResNE - NE, getUNDEF(OvEltVT));
13899
13900 EVT NewResVT = EVT::getVectorVT(*getContext(), ResEltVT, ResNE);
13901 EVT NewOvVT = EVT::getVectorVT(*getContext(), OvEltVT, ResNE);
13902 return std::make_pair(getBuildVector(NewResVT, dl, ResScalars),
13903 getBuildVector(NewOvVT, dl, OvScalars));
13904}
13905
13908 unsigned Bytes,
13909 int Dist) const {
13910 if (LD->isVolatile() || Base->isVolatile())
13911 return false;
13912 // TODO: probably too restrictive for atomics, revisit
13913 if (!LD->isSimple())
13914 return false;
13915 if (LD->isIndexed() || Base->isIndexed())
13916 return false;
13917 if (LD->getChain() != Base->getChain())
13918 return false;
13919 EVT VT = LD->getMemoryVT();
13920 if (VT.getSizeInBits() / 8 != Bytes)
13921 return false;
13922
13923 auto BaseLocDecomp = BaseIndexOffset::match(Base, *this);
13924 auto LocDecomp = BaseIndexOffset::match(LD, *this);
13925
13926 int64_t Offset = 0;
13927 if (BaseLocDecomp.equalBaseIndex(LocDecomp, *this, Offset))
13928 return (Dist * (int64_t)Bytes == Offset);
13929 return false;
13930}
13931
13932/// InferPtrAlignment - Infer alignment of a load / store address. Return
13933/// std::nullopt if it cannot be inferred.
13935 // If this is a GlobalAddress + cst, return the alignment.
13936 const GlobalValue *GV = nullptr;
13937 int64_t GVOffset = 0;
13938 if (TLI->isGAPlusOffset(Ptr.getNode(), GV, GVOffset)) {
13939 unsigned PtrWidth = getDataLayout().getPointerTypeSizeInBits(GV->getType());
13940 KnownBits Known(PtrWidth);
13942 unsigned AlignBits = Known.countMinTrailingZeros();
13943 if (AlignBits)
13944 return commonAlignment(Align(1ull << std::min(31U, AlignBits)), GVOffset);
13945 }
13946
13947 // If this is a direct reference to a stack slot, use information about the
13948 // stack slot's alignment.
13949 int FrameIdx = INT_MIN;
13950 int64_t FrameOffset = 0;
13952 FrameIdx = FI->getIndex();
13953 } else if (isBaseWithConstantOffset(Ptr) &&
13955 // Handle FI+Cst
13956 FrameIdx = cast<FrameIndexSDNode>(Ptr.getOperand(0))->getIndex();
13957 FrameOffset = Ptr.getConstantOperandVal(1);
13958 }
13959
13960 if (FrameIdx != INT_MIN) {
13962 return commonAlignment(MFI.getObjectAlign(FrameIdx), FrameOffset);
13963 }
13964
13965 return std::nullopt;
13966}
13967
13968/// Split the scalar node with EXTRACT_ELEMENT using the provided
13969/// VTs and return the low/high part.
13970std::pair<SDValue, SDValue> SelectionDAG::SplitScalar(const SDValue &N,
13971 const SDLoc &DL,
13972 const EVT &LoVT,
13973 const EVT &HiVT) {
13974 assert(!LoVT.isVector() && !HiVT.isVector() && !N.getValueType().isVector() &&
13975 "Split node must be a scalar type");
13976 SDValue Lo =
13978 SDValue Hi =
13980 return std::make_pair(Lo, Hi);
13981}
13982
13983/// GetSplitDestVTs - Compute the VTs needed for the low/hi parts of a type
13984/// which is split (or expanded) into two not necessarily identical pieces.
13985std::pair<EVT, EVT> SelectionDAG::GetSplitDestVTs(const EVT &VT) const {
13986 // Currently all types are split in half.
13987 EVT LoVT, HiVT;
13988 if (!VT.isVector())
13989 LoVT = HiVT = TLI->getTypeToTransformTo(*getContext(), VT);
13990 else
13991 LoVT = HiVT = VT.getHalfNumVectorElementsVT(*getContext());
13992
13993 return std::make_pair(LoVT, HiVT);
13994}
13995
13996/// GetDependentSplitDestVTs - Compute the VTs needed for the low/hi parts of a
13997/// type, dependent on an enveloping VT that has been split into two identical
13998/// pieces. Sets the HiIsEmpty flag when hi type has zero storage size.
13999std::pair<EVT, EVT>
14001 bool *HiIsEmpty) const {
14002 EVT EltTp = VT.getVectorElementType();
14003 // Examples:
14004 // custom VL=8 with enveloping VL=8/8 yields 8/0 (hi empty)
14005 // custom VL=9 with enveloping VL=8/8 yields 8/1
14006 // custom VL=10 with enveloping VL=8/8 yields 8/2
14007 // etc.
14008 ElementCount VTNumElts = VT.getVectorElementCount();
14009 ElementCount EnvNumElts = EnvVT.getVectorElementCount();
14010 assert(VTNumElts.isScalable() == EnvNumElts.isScalable() &&
14011 "Mixing fixed width and scalable vectors when enveloping a type");
14012 EVT LoVT, HiVT;
14013 if (VTNumElts.getKnownMinValue() > EnvNumElts.getKnownMinValue()) {
14014 LoVT = EVT::getVectorVT(*getContext(), EltTp, EnvNumElts);
14015 HiVT = EVT::getVectorVT(*getContext(), EltTp, VTNumElts - EnvNumElts);
14016 *HiIsEmpty = false;
14017 } else {
14018 // Flag that hi type has zero storage size, but return split envelop type
14019 // (this would be easier if vector types with zero elements were allowed).
14020 LoVT = EVT::getVectorVT(*getContext(), EltTp, VTNumElts);
14021 HiVT = EVT::getVectorVT(*getContext(), EltTp, EnvNumElts);
14022 *HiIsEmpty = true;
14023 }
14024 return std::make_pair(LoVT, HiVT);
14025}
14026
14027/// SplitVector - Split the vector with EXTRACT_SUBVECTOR and return the
14028/// low/high part.
14029std::pair<SDValue, SDValue>
14030SelectionDAG::SplitVector(const SDValue &N, const SDLoc &DL, const EVT &LoVT,
14031 const EVT &HiVT) {
14032 assert(LoVT.isScalableVector() == HiVT.isScalableVector() &&
14033 LoVT.isScalableVector() == N.getValueType().isScalableVector() &&
14034 "Splitting vector with an invalid mixture of fixed and scalable "
14035 "vector types");
14037 N.getValueType().getVectorMinNumElements() &&
14038 "More vector elements requested than available!");
14039 SDValue Lo, Hi;
14040 Lo = getExtractSubvector(DL, LoVT, N, 0);
14041 // For scalable vectors it is safe to use LoVT.getVectorMinNumElements()
14042 // (rather than having to use ElementCount), because EXTRACT_SUBVECTOR scales
14043 // IDX with the runtime scaling factor of the result vector type. For
14044 // fixed-width result vectors, that runtime scaling factor is 1.
14047 return std::make_pair(Lo, Hi);
14048}
14049
14050std::pair<SDValue, SDValue> SelectionDAG::SplitEVL(SDValue N, EVT VecVT,
14051 const SDLoc &DL) {
14052 // Split the vector length parameter.
14053 // %evl -> umin(%evl, %halfnumelts) and usubsat(%evl - %halfnumelts).
14054 EVT VT = N.getValueType();
14056 "Expecting the mask to be an evenly-sized vector");
14057 SDValue HalfNumElts = getElementCount(
14059 SDValue Lo = getNode(ISD::UMIN, DL, VT, N, HalfNumElts);
14060 SDValue Hi = getNode(ISD::USUBSAT, DL, VT, N, HalfNumElts);
14061 return std::make_pair(Lo, Hi);
14062}
14063
14064/// Widen the vector up to the next power of two using INSERT_SUBVECTOR.
14066 EVT VT = N.getValueType();
14069 return getInsertSubvector(DL, getUNDEF(WideVT), N, 0);
14070}
14071
14074 unsigned Start, unsigned Count,
14075 EVT EltVT) {
14076 EVT VT = Op.getValueType();
14077 if (Count == 0)
14079 if (EltVT == EVT())
14080 EltVT = VT.getVectorElementType();
14081 SDLoc SL(Op);
14082 for (unsigned i = Start, e = Start + Count; i != e; ++i) {
14083 Args.push_back(getExtractVectorElt(SL, EltVT, Op, i));
14084 }
14085}
14086
14087// getAddressSpace - Return the address space this GlobalAddress belongs to.
14089 return getGlobal()->getType()->getAddressSpace();
14090}
14091
14094 return Val.MachineCPVal->getType();
14095 return Val.ConstVal->getType();
14096}
14097
14098bool BuildVectorSDNode::isConstantSplat(APInt &SplatValue, APInt &SplatUndef,
14099 unsigned &SplatBitSize,
14100 bool &HasAnyUndefs,
14101 unsigned MinSplatBits,
14102 bool IsBigEndian) const {
14103 EVT VT = getValueType(0);
14104 assert(VT.isVector() && "Expected a vector type");
14105 unsigned VecWidth = VT.getSizeInBits();
14106 if (MinSplatBits > VecWidth)
14107 return false;
14108
14109 // FIXME: The widths are based on this node's type, but build vectors can
14110 // truncate their operands.
14111 SplatValue = APInt(VecWidth, 0);
14112 SplatUndef = APInt(VecWidth, 0);
14113
14114 // Get the bits. Bits with undefined values (when the corresponding element
14115 // of the vector is an ISD::UNDEF value) are set in SplatUndef and cleared
14116 // in SplatValue. If any of the values are not constant, give up and return
14117 // false.
14118 unsigned int NumOps = getNumOperands();
14119 assert(NumOps > 0 && "isConstantSplat has 0-size build vector");
14120 unsigned EltWidth = VT.getScalarSizeInBits();
14121
14122 for (unsigned j = 0; j < NumOps; ++j) {
14123 unsigned i = IsBigEndian ? NumOps - 1 - j : j;
14124 SDValue OpVal = getOperand(i);
14125 unsigned BitPos = j * EltWidth;
14126
14127 if (OpVal.isUndef())
14128 SplatUndef.setBits(BitPos, BitPos + EltWidth);
14129 else if (auto *CN = dyn_cast<ConstantSDNode>(OpVal))
14130 SplatValue.insertBits(CN->getAPIntValue().zextOrTrunc(EltWidth), BitPos);
14131 else if (auto *CN = dyn_cast<ConstantFPSDNode>(OpVal))
14132 SplatValue.insertBits(CN->getValueAPF().bitcastToAPInt(), BitPos);
14133 else
14134 return false;
14135 }
14136
14137 // The build_vector is all constants or undefs. Find the smallest element
14138 // size that splats the vector.
14139 HasAnyUndefs = (SplatUndef != 0);
14140
14141 // FIXME: This does not work for vectors with elements less than 8 bits.
14142 while (VecWidth > 8) {
14143 // If we can't split in half, stop here.
14144 if (VecWidth & 1)
14145 break;
14146
14147 unsigned HalfSize = VecWidth / 2;
14148 APInt HighValue = SplatValue.extractBits(HalfSize, HalfSize);
14149 APInt LowValue = SplatValue.extractBits(HalfSize, 0);
14150 APInt HighUndef = SplatUndef.extractBits(HalfSize, HalfSize);
14151 APInt LowUndef = SplatUndef.extractBits(HalfSize, 0);
14152
14153 // If the two halves do not match (ignoring undef bits), stop here.
14154 if ((HighValue & ~LowUndef) != (LowValue & ~HighUndef) ||
14155 MinSplatBits > HalfSize)
14156 break;
14157
14158 SplatValue = HighValue | LowValue;
14159 SplatUndef = HighUndef & LowUndef;
14160
14161 VecWidth = HalfSize;
14162 }
14163
14164 // FIXME: The loop above only tries to split in halves. But if the input
14165 // vector for example is <3 x i16> it wouldn't be able to detect a
14166 // SplatBitSize of 16. No idea if that is a design flaw currently limiting
14167 // optimizations. I guess that back in the days when this helper was created
14168 // vectors normally was power-of-2 sized.
14169
14170 SplatBitSize = VecWidth;
14171 return true;
14172}
14173
14175 BitVector *UndefElements) const {
14176 unsigned NumOps = getNumOperands();
14177 if (UndefElements) {
14178 UndefElements->clear();
14179 UndefElements->resize(NumOps);
14180 }
14181 assert(NumOps == DemandedElts.getBitWidth() && "Unexpected vector size");
14182 if (!DemandedElts)
14183 return SDValue();
14184 SDValue Splatted;
14185 for (unsigned i = 0; i != NumOps; ++i) {
14186 if (!DemandedElts[i])
14187 continue;
14188 SDValue Op = getOperand(i);
14189 if (Op.isUndef()) {
14190 if (UndefElements)
14191 (*UndefElements)[i] = true;
14192 } else if (!Splatted) {
14193 Splatted = Op;
14194 } else if (Splatted != Op) {
14195 return SDValue();
14196 }
14197 }
14198
14199 if (!Splatted) {
14200 unsigned FirstDemandedIdx = DemandedElts.countr_zero();
14201 assert(getOperand(FirstDemandedIdx).isUndef() &&
14202 "Can only have a splat without a constant for all undefs.");
14203 return getOperand(FirstDemandedIdx);
14204 }
14205
14206 return Splatted;
14207}
14208
14210 APInt DemandedElts = APInt::getAllOnes(getNumOperands());
14211 return getSplatValue(DemandedElts, UndefElements);
14212}
14213
14215 SmallVectorImpl<SDValue> &Sequence,
14216 BitVector *UndefElements) const {
14217 unsigned NumOps = getNumOperands();
14218 Sequence.clear();
14219 if (UndefElements) {
14220 UndefElements->clear();
14221 UndefElements->resize(NumOps);
14222 }
14223 assert(NumOps == DemandedElts.getBitWidth() && "Unexpected vector size");
14224 if (!DemandedElts || NumOps < 2 || !isPowerOf2_32(NumOps))
14225 return false;
14226
14227 // Set the undefs even if we don't find a sequence (like getSplatValue).
14228 if (UndefElements)
14229 for (unsigned I = 0; I != NumOps; ++I)
14230 if (DemandedElts[I] && getOperand(I).isUndef())
14231 (*UndefElements)[I] = true;
14232
14233 // Iteratively widen the sequence length looking for repetitions.
14234 for (unsigned SeqLen = 1; SeqLen < NumOps; SeqLen *= 2) {
14235 Sequence.append(SeqLen, SDValue());
14236 for (unsigned I = 0; I != NumOps; ++I) {
14237 if (!DemandedElts[I])
14238 continue;
14239 SDValue &SeqOp = Sequence[I % SeqLen];
14241 if (Op.isUndef()) {
14242 if (!SeqOp)
14243 SeqOp = Op;
14244 continue;
14245 }
14246 if (SeqOp && !SeqOp.isUndef() && SeqOp != Op) {
14247 Sequence.clear();
14248 break;
14249 }
14250 SeqOp = Op;
14251 }
14252 if (!Sequence.empty())
14253 return true;
14254 }
14255
14256 assert(Sequence.empty() && "Failed to empty non-repeating sequence pattern");
14257 return false;
14258}
14259
14261 BitVector *UndefElements) const {
14262 APInt DemandedElts = APInt::getAllOnes(getNumOperands());
14263 return getRepeatedSequence(DemandedElts, Sequence, UndefElements);
14264}
14265
14268 BitVector *UndefElements) const {
14270 getSplatValue(DemandedElts, UndefElements));
14271}
14272
14275 return dyn_cast_or_null<ConstantSDNode>(getSplatValue(UndefElements));
14276}
14277
14280 BitVector *UndefElements) const {
14282 getSplatValue(DemandedElts, UndefElements));
14283}
14284
14289
14290int32_t
14292 uint32_t BitWidth) const {
14293 if (ConstantFPSDNode *CN =
14295 bool IsExact;
14296 APSInt IntVal(BitWidth);
14297 const APFloat &APF = CN->getValueAPF();
14298 if (APF.convertToInteger(IntVal, APFloat::rmTowardZero, &IsExact) !=
14299 APFloat::opOK ||
14300 !IsExact)
14301 return -1;
14302
14303 return IntVal.exactLogBase2();
14304 }
14305 return -1;
14306}
14307
14309 bool IsLittleEndian, unsigned DstEltSizeInBits,
14310 SmallVectorImpl<APInt> &RawBitElements, BitVector &UndefElements) const {
14311 // Early-out if this contains anything but Undef/Constant/ConstantFP.
14312 if (!isConstant())
14313 return false;
14314
14315 unsigned NumSrcOps = getNumOperands();
14316 unsigned SrcEltSizeInBits = getValueType(0).getScalarSizeInBits();
14317 assert(((NumSrcOps * SrcEltSizeInBits) % DstEltSizeInBits) == 0 &&
14318 "Invalid bitcast scale");
14319
14320 // Extract raw src bits.
14321 SmallVector<APInt> SrcBitElements(NumSrcOps,
14322 APInt::getZero(SrcEltSizeInBits));
14323 BitVector SrcUndeElements(NumSrcOps, false);
14324
14325 for (unsigned I = 0; I != NumSrcOps; ++I) {
14327 if (Op.isUndef()) {
14328 SrcUndeElements.set(I);
14329 continue;
14330 }
14331 auto *CInt = dyn_cast<ConstantSDNode>(Op);
14332 auto *CFP = dyn_cast<ConstantFPSDNode>(Op);
14333 assert((CInt || CFP) && "Unknown constant");
14334 SrcBitElements[I] = CInt ? CInt->getAPIntValue().trunc(SrcEltSizeInBits)
14335 : CFP->getValueAPF().bitcastToAPInt();
14336 }
14337
14338 // Recast to dst width.
14339 recastRawBits(IsLittleEndian, DstEltSizeInBits, RawBitElements,
14340 SrcBitElements, UndefElements, SrcUndeElements);
14341 return true;
14342}
14343
14344void BuildVectorSDNode::recastRawBits(bool IsLittleEndian,
14345 unsigned DstEltSizeInBits,
14346 SmallVectorImpl<APInt> &DstBitElements,
14347 ArrayRef<APInt> SrcBitElements,
14348 BitVector &DstUndefElements,
14349 const BitVector &SrcUndefElements) {
14350 unsigned NumSrcOps = SrcBitElements.size();
14351 unsigned SrcEltSizeInBits = SrcBitElements[0].getBitWidth();
14352 assert(((NumSrcOps * SrcEltSizeInBits) % DstEltSizeInBits) == 0 &&
14353 "Invalid bitcast scale");
14354 assert(NumSrcOps == SrcUndefElements.size() &&
14355 "Vector size mismatch");
14356
14357 unsigned NumDstOps = (NumSrcOps * SrcEltSizeInBits) / DstEltSizeInBits;
14358 DstUndefElements.clear();
14359 DstUndefElements.resize(NumDstOps, false);
14360 DstBitElements.assign(NumDstOps, APInt::getZero(DstEltSizeInBits));
14361
14362 // Concatenate src elements constant bits together into dst element.
14363 if (SrcEltSizeInBits <= DstEltSizeInBits) {
14364 unsigned Scale = DstEltSizeInBits / SrcEltSizeInBits;
14365 for (unsigned I = 0; I != NumDstOps; ++I) {
14366 DstUndefElements.set(I);
14367 APInt &DstBits = DstBitElements[I];
14368 for (unsigned J = 0; J != Scale; ++J) {
14369 unsigned Idx = (I * Scale) + (IsLittleEndian ? J : (Scale - J - 1));
14370 if (SrcUndefElements[Idx])
14371 continue;
14372 DstUndefElements.reset(I);
14373 const APInt &SrcBits = SrcBitElements[Idx];
14374 assert(SrcBits.getBitWidth() == SrcEltSizeInBits &&
14375 "Illegal constant bitwidths");
14376 DstBits.insertBits(SrcBits, J * SrcEltSizeInBits);
14377 }
14378 }
14379 return;
14380 }
14381
14382 // Split src element constant bits into dst elements.
14383 unsigned Scale = SrcEltSizeInBits / DstEltSizeInBits;
14384 for (unsigned I = 0; I != NumSrcOps; ++I) {
14385 if (SrcUndefElements[I]) {
14386 DstUndefElements.set(I * Scale, (I + 1) * Scale);
14387 continue;
14388 }
14389 const APInt &SrcBits = SrcBitElements[I];
14390 for (unsigned J = 0; J != Scale; ++J) {
14391 unsigned Idx = (I * Scale) + (IsLittleEndian ? J : (Scale - J - 1));
14392 APInt &DstBits = DstBitElements[Idx];
14393 DstBits = SrcBits.extractBits(DstEltSizeInBits, J * DstEltSizeInBits);
14394 }
14395 }
14396}
14397
14399 for (const SDValue &Op : op_values()) {
14400 unsigned Opc = Op.getOpcode();
14401 if (!Op.isUndef() && Opc != ISD::Constant && Opc != ISD::ConstantFP)
14402 return false;
14403 }
14404 return true;
14405}
14406
14407std::optional<std::pair<APInt, APInt>>
14409 unsigned NumOps = getNumOperands();
14410 if (NumOps < 2)
14411 return std::nullopt;
14412
14413 unsigned EltSize = getValueType(0).getScalarSizeInBits();
14414 APInt Start, Stride;
14415 int FirstIdx = -1, SecondIdx = -1;
14416
14417 // Find the first two non-undef constant elements to determine Start and
14418 // Stride, then verify all remaining elements match the sequence.
14419 for (unsigned I = 0; I < NumOps; ++I) {
14421 if (Op->isUndef())
14422 continue;
14423 if (!isa<ConstantSDNode>(Op))
14424 return std::nullopt;
14425
14426 APInt Val = getConstantOperandAPInt(I).trunc(EltSize);
14427 if (FirstIdx < 0) {
14428 FirstIdx = I;
14429 Start = Val;
14430 } else if (SecondIdx < 0) {
14431 SecondIdx = I;
14432 // Compute stride using modular arithmetic. Simple division would handle
14433 // common strides (1, 2, -1, etc.), but modular inverse maximizes matches.
14434 // Example: <0, poison, poison, 0xFF> has stride 0x55 since 3*0x55 = 0xFF
14435 // Note that modular arithmetic is agnostic to signed/unsigned.
14436 unsigned IdxDiff = I - FirstIdx;
14437 APInt ValDiff = Val - Start;
14438
14439 // Step 1: Factor out common powers of 2 from IdxDiff and ValDiff.
14440 unsigned CommonPow2Bits = llvm::countr_zero(IdxDiff);
14441 if (ValDiff.countr_zero() < CommonPow2Bits)
14442 return std::nullopt; // ValDiff not divisible by 2^CommonPow2Bits
14443 IdxDiff >>= CommonPow2Bits;
14444 ValDiff.lshrInPlace(CommonPow2Bits);
14445
14446 // Step 2: IdxDiff is now odd, so its inverse mod 2^EltSize exists.
14447 // TODO: There are 2^CommonPow2Bits valid strides; currently we only try
14448 // one, but we could try all candidates to handle more cases.
14449 Stride = ValDiff * APInt(EltSize, IdxDiff).multiplicativeInverse();
14450 if (Stride.isZero())
14451 return std::nullopt;
14452
14453 // Step 3: Adjust Start based on the first defined element's index.
14454 Start -= Stride * FirstIdx;
14455 } else {
14456 // Verify this element matches the sequence.
14457 if (Val != Start + Stride * I)
14458 return std::nullopt;
14459 }
14460 }
14461
14462 // Need at least two defined elements.
14463 if (SecondIdx < 0)
14464 return std::nullopt;
14465
14466 return std::make_pair(Start, Stride);
14467}
14468
14470 // Find the first non-undef value in the shuffle mask.
14471 unsigned i, e;
14472 for (i = 0, e = Mask.size(); i != e && Mask[i] < 0; ++i)
14473 /* search */;
14474
14475 // If all elements are undefined, this shuffle can be considered a splat
14476 // (although it should eventually get simplified away completely).
14477 if (i == e)
14478 return true;
14479
14480 // Make sure all remaining elements are either undef or the same as the first
14481 // non-undef value.
14482 for (int Idx = Mask[i]; i != e; ++i)
14483 if (Mask[i] >= 0 && Mask[i] != Idx)
14484 return false;
14485 return true;
14486}
14487
14488// Returns true if it is a constant integer BuildVector or constant integer,
14489// possibly hidden by a bitcast.
14491 SDValue N, bool AllowOpaques) const {
14493
14494 if (auto *C = dyn_cast<ConstantSDNode>(N))
14495 return AllowOpaques || !C->isOpaque();
14496
14498 return true;
14499
14500 // Treat a GlobalAddress supporting constant offset folding as a
14501 // constant integer.
14502 if (auto *GA = dyn_cast<GlobalAddressSDNode>(N))
14503 if (GA->getOpcode() == ISD::GlobalAddress &&
14504 TLI->isOffsetFoldingLegal(GA))
14505 return true;
14506
14507 if ((N.getOpcode() == ISD::SPLAT_VECTOR) &&
14508 isa<ConstantSDNode>(N.getOperand(0)))
14509 return true;
14510 return false;
14511}
14512
14513// Returns true if it is a constant float BuildVector or constant float.
14516 return true;
14517
14519 return true;
14520
14521 if ((N.getOpcode() == ISD::SPLAT_VECTOR) &&
14522 isa<ConstantFPSDNode>(N.getOperand(0)))
14523 return true;
14524
14525 return false;
14526}
14527
14528std::optional<bool> SelectionDAG::isBoolConstant(SDValue N) const {
14529 ConstantSDNode *Const =
14530 isConstOrConstSplat(N, false, /*AllowTruncation=*/true);
14531 if (!Const)
14532 return std::nullopt;
14533
14534 EVT VT = N->getValueType(0);
14535 const APInt CVal = Const->getAPIntValue().trunc(VT.getScalarSizeInBits());
14536 switch (TLI->getBooleanContents(N.getValueType())) {
14538 if (CVal.isOne())
14539 return true;
14540 if (CVal.isZero())
14541 return false;
14542 return std::nullopt;
14544 if (CVal.isAllOnes())
14545 return true;
14546 if (CVal.isZero())
14547 return false;
14548 return std::nullopt;
14550 return CVal[0];
14551 }
14552 llvm_unreachable("Unknown BooleanContent enum");
14553}
14554
14555void SelectionDAG::createOperands(SDNode *Node, ArrayRef<SDValue> Vals) {
14556 assert(!Node->OperandList && "Node already has operands");
14558 "too many operands to fit into SDNode");
14559 SDUse *Ops = OperandRecycler.allocate(
14560 ArrayRecycler<SDUse>::Capacity::get(Vals.size()), OperandAllocator);
14561
14562 bool IsDivergent = false;
14563 for (unsigned I = 0; I != Vals.size(); ++I) {
14564 Ops[I].setUser(Node);
14565 Ops[I].setInitial(Vals[I]);
14566 EVT VT = Ops[I].getValueType();
14567
14568 // Skip Chain. It does not carry divergence.
14569 if (VT != MVT::Other &&
14570 (VT != MVT::Glue || gluePropagatesDivergence(Ops[I].getNode())) &&
14571 Ops[I].getNode()->isDivergent()) {
14572 IsDivergent = true;
14573 }
14574 }
14575 Node->NumOperands = Vals.size();
14576 Node->OperandList = Ops;
14577 if (!TLI->isSDNodeAlwaysUniform(Node)) {
14578 IsDivergent |= TLI->isSDNodeSourceOfDivergence(Node, FLI, UA);
14579 Node->SDNodeBits.IsDivergent = IsDivergent;
14580 }
14581 checkForCycles(Node);
14582}
14583
14586 size_t Limit = SDNode::getMaxNumOperands();
14587 while (Vals.size() > Limit) {
14588 unsigned SliceIdx = Vals.size() - Limit;
14589 auto ExtractedTFs = ArrayRef<SDValue>(Vals).slice(SliceIdx, Limit);
14590 SDValue NewTF = getNode(ISD::TokenFactor, DL, MVT::Other, ExtractedTFs);
14591 Vals.erase(Vals.begin() + SliceIdx, Vals.end());
14592 Vals.emplace_back(NewTF);
14593 }
14594 return getNode(ISD::TokenFactor, DL, MVT::Other, Vals);
14595}
14596
14598 EVT VT, SDNodeFlags Flags) {
14599 switch (Opcode) {
14600 default:
14601 return SDValue();
14602 case ISD::ADD:
14603 case ISD::OR:
14604 case ISD::XOR:
14605 case ISD::UMAX:
14606 return getConstant(0, DL, VT);
14607 case ISD::MUL:
14608 return getConstant(1, DL, VT);
14609 case ISD::AND:
14610 case ISD::UMIN:
14611 return getAllOnesConstant(DL, VT);
14612 case ISD::SMAX:
14614 case ISD::SMIN:
14616 case ISD::FADD:
14617 // If flags allow, prefer positive zero since it's generally cheaper
14618 // to materialize on most targets.
14619 return getConstantFP(Flags.hasNoSignedZeros() ? 0.0 : -0.0, DL, VT);
14620 case ISD::FMUL:
14621 return getConstantFP(1.0, DL, VT);
14622 case ISD::FMINNUM:
14623 case ISD::FMAXNUM: {
14624 // Neutral element for fminnum is NaN, Inf or FLT_MAX, depending on FMF.
14625 const fltSemantics &Semantics = VT.getFltSemantics();
14626 APFloat NeutralAF = !Flags.hasNoNaNs() ? APFloat::getQNaN(Semantics) :
14627 !Flags.hasNoInfs() ? APFloat::getInf(Semantics) :
14628 APFloat::getLargest(Semantics);
14629 if (Opcode == ISD::FMAXNUM)
14630 NeutralAF.changeSign();
14631
14632 return getConstantFP(NeutralAF, DL, VT);
14633 }
14634 case ISD::FMINIMUM:
14635 case ISD::FMAXIMUM: {
14636 // Neutral element for fminimum is Inf or FLT_MAX, depending on FMF.
14637 const fltSemantics &Semantics = VT.getFltSemantics();
14638 APFloat NeutralAF = !Flags.hasNoInfs() ? APFloat::getInf(Semantics)
14639 : APFloat::getLargest(Semantics);
14640 if (Opcode == ISD::FMAXIMUM)
14641 NeutralAF.changeSign();
14642
14643 return getConstantFP(NeutralAF, DL, VT);
14644 }
14645
14646 }
14647}
14648
14649/// Helper used to make a call to a library function that has one argument of
14650/// pointer type.
14651///
14652/// Such functions include 'fegetmode', 'fesetenv' and some others, which are
14653/// used to get or set floating-point state. They have one argument of pointer
14654/// type, which points to the memory region containing bits of the
14655/// floating-point state. The value returned by such function is ignored in the
14656/// created call.
14657///
14658/// \param LibFunc Reference to library function (value of RTLIB::Libcall).
14659/// \param Ptr Pointer used to save/load state.
14660/// \param InChain Ingoing token chain.
14661/// \returns Outgoing chain token.
14663 SDValue InChain,
14664 const SDLoc &DLoc) {
14665 assert(InChain.getValueType() == MVT::Other && "Expected token chain");
14667 Args.emplace_back(Ptr, Ptr.getValueType().getTypeForEVT(*getContext()));
14668 RTLIB::LibcallImpl LibcallImpl =
14669 Libcalls->getLibcallImpl(static_cast<RTLIB::Libcall>(LibFunc));
14670 if (LibcallImpl == RTLIB::Unsupported)
14671 reportFatalUsageError("emitting call to unsupported libcall");
14672
14673 SDValue Callee =
14674 getExternalSymbol(LibcallImpl, TLI->getPointerTy(getDataLayout()));
14676 CLI.setDebugLoc(DLoc).setChain(InChain).setLibCallee(
14677 Libcalls->getLibcallImplCallingConv(LibcallImpl),
14678 Type::getVoidTy(*getContext()), Callee, std::move(Args));
14679 return TLI->LowerCallTo(CLI).second;
14680}
14681
14683 assert(From && To && "Invalid SDNode; empty source SDValue?");
14684 auto I = SDEI.find(From);
14685 if (I == SDEI.end())
14686 return;
14687
14688 // Use of operator[] on the DenseMap may cause an insertion, which invalidates
14689 // the iterator, hence the need to make a copy to prevent a use-after-free.
14690 NodeExtraInfo NEI = I->second;
14691 if (LLVM_LIKELY(!NEI.PCSections)) {
14692 // No deep copy required for the types of extra info set.
14693 //
14694 // FIXME: Investigate if other types of extra info also need deep copy. This
14695 // depends on the types of nodes they can be attached to: if some extra info
14696 // is only ever attached to nodes where a replacement To node is always the
14697 // node where later use and propagation of the extra info has the intended
14698 // semantics, no deep copy is required.
14699 SDEI[To] = std::move(NEI);
14700 return;
14701 }
14702
14703 const SDNode *EntrySDN = getEntryNode().getNode();
14704
14705 // We need to copy NodeExtraInfo to all _new_ nodes that are being introduced
14706 // through the replacement of From with To. Otherwise, replacements of a node
14707 // (From) with more complex nodes (To and its operands) may result in lost
14708 // extra info where the root node (To) is insignificant in further propagating
14709 // and using extra info when further lowering to MIR.
14710 //
14711 // In the first step pre-populate the visited set with the nodes reachable
14712 // from the old From node. This avoids copying NodeExtraInfo to parts of the
14713 // DAG that is not new and should be left untouched.
14714 SmallVector<const SDNode *> Leafs{From}; // Leafs reachable with VisitFrom.
14715 DenseSet<const SDNode *> FromReach; // The set of nodes reachable from From.
14716 auto VisitFrom = [&](auto &&Self, const SDNode *N, int MaxDepth) {
14717 if (MaxDepth == 0) {
14718 // Remember this node in case we need to increase MaxDepth and continue
14719 // populating FromReach from this node.
14720 Leafs.emplace_back(N);
14721 return;
14722 }
14723 if (!FromReach.insert(N).second)
14724 return;
14725 for (const SDValue &Op : N->op_values())
14726 Self(Self, Op.getNode(), MaxDepth - 1);
14727 };
14728
14729 // Copy extra info to To and all its transitive operands (that are new).
14731 auto DeepCopyTo = [&](auto &&Self, const SDNode *N) {
14732 if (FromReach.contains(N))
14733 return true;
14734 if (!Visited.insert(N).second)
14735 return true;
14736 if (EntrySDN == N)
14737 return false;
14738 for (const SDValue &Op : N->op_values()) {
14739 if (N == To && Op.getNode() == EntrySDN) {
14740 // Special case: New node's operand is the entry node; just need to
14741 // copy extra info to new node.
14742 break;
14743 }
14744 if (!Self(Self, Op.getNode()))
14745 return false;
14746 }
14747 // Copy only if entry node was not reached.
14748 SDEI[N] = std::move(NEI);
14749 return true;
14750 };
14751
14752 // We first try with a lower MaxDepth, assuming that the path to common
14753 // operands between From and To is relatively short. This significantly
14754 // improves performance in the common case. The initial MaxDepth is big
14755 // enough to avoid retry in the common case; the last MaxDepth is large
14756 // enough to avoid having to use the fallback below (and protects from
14757 // potential stack exhaustion from recursion).
14758 for (int PrevDepth = 0, MaxDepth = 16; MaxDepth <= 1024;
14759 PrevDepth = MaxDepth, MaxDepth *= 2, Visited.clear()) {
14760 // StartFrom is the previous (or initial) set of leafs reachable at the
14761 // previous maximum depth.
14763 std::swap(StartFrom, Leafs);
14764 for (const SDNode *N : StartFrom)
14765 VisitFrom(VisitFrom, N, MaxDepth - PrevDepth);
14766 if (LLVM_LIKELY(DeepCopyTo(DeepCopyTo, To)))
14767 return;
14768 // This should happen very rarely (reached the entry node).
14769 LLVM_DEBUG(dbgs() << __func__ << ": MaxDepth=" << MaxDepth << " too low\n");
14770 assert(!Leafs.empty());
14771 }
14772
14773 // This should not happen - but if it did, that means the subgraph reachable
14774 // from From has depth greater or equal to maximum MaxDepth, and VisitFrom()
14775 // could not visit all reachable common operands. Consequently, we were able
14776 // to reach the entry node.
14777 errs() << "warning: incomplete propagation of SelectionDAG::NodeExtraInfo\n";
14778 assert(false && "From subgraph too complex - increase max. MaxDepth?");
14779 // Best-effort fallback if assertions disabled.
14780 SDEI[To] = std::move(NEI);
14781}
14782
14783#ifndef NDEBUG
14784static void checkForCyclesHelper(const SDNode *N,
14787 const llvm::SelectionDAG *DAG) {
14788 // If this node has already been checked, don't check it again.
14789 if (Checked.count(N))
14790 return;
14791
14792 // If a node has already been visited on this depth-first walk, reject it as
14793 // a cycle.
14794 if (!Visited.insert(N).second) {
14795 errs() << "Detected cycle in SelectionDAG\n";
14796 dbgs() << "Offending node:\n";
14797 N->dumprFull(DAG); dbgs() << "\n";
14798 abort();
14799 }
14800
14801 for (const SDValue &Op : N->op_values())
14802 checkForCyclesHelper(Op.getNode(), Visited, Checked, DAG);
14803
14804 Checked.insert(N);
14805 Visited.erase(N);
14806}
14807#endif
14808
14810 const llvm::SelectionDAG *DAG,
14811 bool force) {
14812#ifndef NDEBUG
14813 bool check = force;
14814#ifdef EXPENSIVE_CHECKS
14815 check = true;
14816#endif // EXPENSIVE_CHECKS
14817 if (check) {
14818 assert(N && "Checking nonexistent SDNode");
14821 checkForCyclesHelper(N, visited, checked, DAG);
14822 }
14823#endif // !NDEBUG
14824}
14825
14826void llvm::checkForCycles(const llvm::SelectionDAG *DAG, bool force) {
14827 checkForCycles(DAG->getRoot().getNode(), DAG, force);
14828}
return SDValue()
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static bool isConstant(const MachineInstr &MI)
constexpr LLT S1
This file declares a class to represent arbitrary precision floating point values and provide a varie...
This file implements a class to represent arbitrary precision integral constant values and operations...
This file implements the APSInt class, which is a simple class that represents an arbitrary sized int...
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
This file implements the BitVector class.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static std::optional< bool > isBigEndian(const SmallDenseMap< int64_t, int64_t, 8 > &MemOffset2Idx, int64_t LowestIdx)
Given a map from byte offsets in memory to indices in a load/store, determine if that map corresponds...
#define __asan_unpoison_memory_region(p, size)
Definition Compiler.h:592
#define LLVM_LIKELY(EXPR)
Definition Compiler.h:335
This file contains the declarations for the subclasses of Constant, which represent the different fla...
This file defines the DenseSet and SmallDenseSet classes.
This file contains constants used for implementing Dwarf debug support.
This file defines a hash set that can be used to remove duplication of nodes in a graph.
iv users
Definition IVUsers.cpp:48
std::pair< Instruction::BinaryOps, Value * > OffsetOp
Find all possible pairs (BinOp, RHS) that BinOp V, RHS can be simplified.
const size_t AbstractManglingParser< Derived, Alloc >::NumOps
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
static LVOptions Options
Definition LVOptions.cpp:25
static Register getMemsetValue(Register Val, LLT Ty, MachineIRBuilder &MIB)
static bool shouldLowerMemFuncForSize(const MachineFunction &MF)
static bool isZero(Value *V, const DataLayout &DL, DominatorTree *DT, AssumptionCache *AC)
Definition Lint.cpp:539
static Align getPrefTypeAlign(EVT VT, SelectionDAG &DAG)
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
#define G(x, y, z)
Definition MD5.cpp:55
This file declares the MachineConstantPool class which is an abstract constant pool to keep track of ...
Register const TargetRegisterInfo * TRI
This file provides utility analysis objects describing memory locations.
This file contains the declarations for metadata subclasses.
#define T
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
#define P(N)
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > & Cond
Remove Loads Into Fake Uses
static bool isValid(const char C)
Returns true if C is a valid mangled character: <0-9a-zA-Z_>.
Contains matchers for matching SelectionDAG nodes and values.
static Type * getValueType(Value *V)
Returns the type of the given value/instruction V.
This file contains some templates that are useful if you are working with the STL at all.
static uint64_t umul_ov(uint64_t i, uint64_t j, bool &Overflow)
static SDValue getMemcpyLoadsAndStores(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src, uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo, BatchAAResults *BatchAA)
static SDValue getFixedOrScalableQuantity(SelectionDAG &DAG, const SDLoc &DL, EVT VT, Ty Quantity)
static std::pair< SDValue, SDValue > getRuntimeCallSDValueHelper(SDValue Chain, const SDLoc &dl, TargetLowering::ArgListTy &&Args, const CallInst *CI, RTLIB::Libcall Call, SelectionDAG *DAG, const TargetLowering *TLI)
static SDValue getMemsetStores(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src, uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline, MachinePointerInfo DstPtrInfo, const AAMDNodes &AAInfo)
Lower the call to 'memset' intrinsic function into a series of store operations.
static std::optional< APInt > FoldValueWithUndef(unsigned Opcode, const APInt &C1, bool IsUndef1, const APInt &C2, bool IsUndef2)
static SDValue FoldSTEP_VECTOR(const SDLoc &DL, EVT VT, SDValue Step, SelectionDAG &DAG)
static void AddNodeIDNode(FoldingSetNodeID &ID, unsigned OpC, SDVTList VTList, ArrayRef< SDValue > OpList)
static SDValue getMemsetStringVal(EVT VT, const SDLoc &dl, SelectionDAG &DAG, const TargetLowering &TLI, const ConstantDataArraySlice &Slice)
getMemsetStringVal - Similar to getMemsetValue.
static cl::opt< bool > EnableMemCpyDAGOpt("enable-memcpy-dag-opt", cl::Hidden, cl::init(true), cl::desc("Gang up loads and stores generated by inlining of memcpy"))
static bool haveNoCommonBitsSetCommutative(SDValue A, SDValue B)
static void AddNodeIDValueTypes(FoldingSetNodeID &ID, SDVTList VTList)
AddNodeIDValueTypes - Value type lists are intern'd so we can represent them solely with their pointe...
static void commuteShuffle(SDValue &N1, SDValue &N2, MutableArrayRef< int > M)
Swaps the values of N1 and N2.
static bool isMemSrcFromConstant(SDValue Src, ConstantDataArraySlice &Slice)
Returns true if memcpy source is constant data.
static SDValue getMemmoveLoadsAndStores(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src, uint64_t Size, Align Alignment, bool isVol, bool AlwaysInline, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo)
static void AddNodeIDOpcode(FoldingSetNodeID &ID, unsigned OpC)
AddNodeIDOpcode - Add the node opcode to the NodeID data.
static ISD::CondCode getSetCCInverseImpl(ISD::CondCode Op, bool isIntegerLike)
static bool doNotCSE(SDNode *N)
doNotCSE - Return true if CSE should not be performed for this node.
static cl::opt< int > MaxLdStGlue("ldstmemcpy-glue-max", cl::desc("Number limit for gluing ld/st of memcpy."), cl::Hidden, cl::init(0))
static void AddNodeIDOperands(FoldingSetNodeID &ID, ArrayRef< SDValue > Ops)
AddNodeIDOperands - Various routines for adding operands to the NodeID data.
static SDValue foldCONCAT_VECTORS(const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops, SelectionDAG &DAG)
Try to simplify vector concatenation to an input value, undef, or build vector.
static MachinePointerInfo InferPointerInfo(const MachinePointerInfo &Info, SelectionDAG &DAG, SDValue Ptr, int64_t Offset=0)
InferPointerInfo - If the specified ptr/offset is a frame index, infer a MachinePointerInfo record fr...
static bool isInTailCallPositionWrapper(const CallInst *CI, const SelectionDAG *SelDAG, bool AllowReturnsFirstArg)
static void AddNodeIDCustom(FoldingSetNodeID &ID, const SDNode *N)
If this is an SDNode with special info, add this info to the NodeID data.
static bool gluePropagatesDivergence(const SDNode *Node)
Return true if a glue output should propagate divergence information.
static void NewSDValueDbgMsg(SDValue V, StringRef Msg, SelectionDAG *G)
static SDVTList makeVTList(const EVT *VTs, unsigned NumVTs)
makeVTList - Return an instance of the SDVTList struct initialized with the specified members.
static void checkForCyclesHelper(const SDNode *N, SmallPtrSetImpl< const SDNode * > &Visited, SmallPtrSetImpl< const SDNode * > &Checked, const llvm::SelectionDAG *DAG)
static void chainLoadsAndStoresForMemcpy(SelectionDAG &DAG, const SDLoc &dl, SmallVector< SDValue, 32 > &OutChains, unsigned From, unsigned To, SmallVector< SDValue, 16 > &OutLoadChains, SmallVector< SDValue, 16 > &OutStoreChains)
static int isSignedOp(ISD::CondCode Opcode)
For an integer comparison, return 1 if the comparison is a signed operation and 2 if the result is an...
static std::optional< APInt > FoldValue(unsigned Opcode, const APInt &C1, const APInt &C2)
static SDValue FoldBUILD_VECTOR(const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops, SelectionDAG &DAG)
static void checkAddrSpaceIsValidForLibcall(const TargetLowering *TLI, unsigned AS)
static cl::opt< unsigned > MaxSteps("has-predecessor-max-steps", cl::Hidden, cl::init(8192), cl::desc("DAG combiner limit number of steps when searching DAG " "for predecessor nodes"))
This file defines the SmallPtrSet class.
This file defines the SmallVector class.
#define LLVM_DEBUG(...)
Definition Debug.h:114
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
static SymbolRef::Type getType(const Symbol *Sym)
Definition TapiFile.cpp:39
This file describes how to lower LLVM code to machine code.
static void removeOperands(MachineInstr &MI, unsigned i)
static std::optional< unsigned > getOpcode(ArrayRef< VPValue * > Values)
Returns the opcode of Values or ~0 if they do not all agree.
Definition VPlanSLP.cpp:247
static OverflowResult mapOverflowResult(ConstantRange::OverflowResult OR)
Convert ConstantRange OverflowResult into ValueTracking OverflowResult.
static int Lookup(ArrayRef< TableEntry > Table, unsigned Opcode)
static unsigned getSize(unsigned Kind)
static const fltSemantics & IEEEsingle()
Definition APFloat.h:296
cmpResult
IEEE-754R 5.11: Floating Point Comparison Relations.
Definition APFloat.h:334
static constexpr roundingMode rmTowardZero
Definition APFloat.h:348
static const fltSemantics & BFloat()
Definition APFloat.h:295
static const fltSemantics & IEEEquad()
Definition APFloat.h:298
static const fltSemantics & IEEEdouble()
Definition APFloat.h:297
static constexpr roundingMode rmTowardNegative
Definition APFloat.h:347
static constexpr roundingMode rmNearestTiesToEven
Definition APFloat.h:344
static constexpr roundingMode rmTowardPositive
Definition APFloat.h:346
static const fltSemantics & IEEEhalf()
Definition APFloat.h:294
opStatus
IEEE-754R 7: Default exception handling.
Definition APFloat.h:360
static APFloat getQNaN(const fltSemantics &Sem, bool Negative=false, const APInt *payload=nullptr)
Factory for QNaN values.
Definition APFloat.h:1171
opStatus divide(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1259
void copySign(const APFloat &RHS)
Definition APFloat.h:1353
LLVM_ABI opStatus convert(const fltSemantics &ToSemantics, roundingMode RM, bool *losesInfo)
Definition APFloat.cpp:5975
opStatus subtract(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1241
bool isExactlyValue(double V) const
We don't rely on operator== working on double values, as it returns true for things that are clearly ...
Definition APFloat.h:1495
opStatus add(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1232
bool isFinite() const
Definition APFloat.h:1517
opStatus convertFromAPInt(const APInt &Input, bool IsSigned, roundingMode RM)
Definition APFloat.h:1398
opStatus multiply(const APFloat &RHS, roundingMode RM)
Definition APFloat.h:1250
bool isSignaling() const
Definition APFloat.h:1514
opStatus fusedMultiplyAdd(const APFloat &Multiplicand, const APFloat &Addend, roundingMode RM)
Definition APFloat.h:1286
bool isZero() const
Definition APFloat.h:1508
static APFloat getLargest(const fltSemantics &Sem, bool Negative=false)
Returns the largest finite number in the given semantics.
Definition APFloat.h:1189
opStatus convertToInteger(MutableArrayRef< integerPart > Input, unsigned int Width, bool IsSigned, roundingMode RM, bool *IsExact) const
Definition APFloat.h:1383
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
Definition APFloat.h:1149
opStatus mod(const APFloat &RHS)
Definition APFloat.h:1277
bool isPosZero() const
Definition APFloat.h:1523
bool isNegZero() const
Definition APFloat.h:1524
void changeSign()
Definition APFloat.h:1348
static APFloat getNaN(const fltSemantics &Sem, bool Negative=false, uint64_t payload=0)
Factory for NaN values.
Definition APFloat.h:1160
Class for arbitrary precision integers.
Definition APInt.h:78
LLVM_ABI APInt umul_ov(const APInt &RHS, bool &Overflow) const
Definition APInt.cpp:1982
LLVM_ABI APInt usub_sat(const APInt &RHS) const
Definition APInt.cpp:2066
LLVM_ABI APInt udiv(const APInt &RHS) const
Unsigned division operation.
Definition APInt.cpp:1584
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
Definition APInt.h:235
void clearBit(unsigned BitPosition)
Set a given bit to 0.
Definition APInt.h:1421
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
Definition APInt.cpp:1023
static APInt getSignMask(unsigned BitWidth)
Get the SignMask for a specific bit width.
Definition APInt.h:230
uint64_t getZExtValue() const
Get zero extended value.
Definition APInt.h:1555
void setHighBits(unsigned hiBits)
Set the top hiBits bits.
Definition APInt.h:1406
unsigned popcount() const
Count the number of bits set.
Definition APInt.h:1685
void setBitsFrom(unsigned loBit)
Set the top bits starting from loBit.
Definition APInt.h:1400
LLVM_ABI APInt getHiBits(unsigned numBits) const
Compute an APInt containing numBits highbits from this APInt.
Definition APInt.cpp:639
LLVM_ABI APInt zextOrTrunc(unsigned width) const
Zero extend or truncate to width.
Definition APInt.cpp:1044
unsigned getActiveBits() const
Compute the number of active bits in the value.
Definition APInt.h:1527
LLVM_ABI APInt trunc(unsigned width) const
Truncate to new width.
Definition APInt.cpp:936
void setBit(unsigned BitPosition)
Set the given bit to 1 whose position is given as "bitPosition".
Definition APInt.h:1345
APInt abs() const
Get the absolute value.
Definition APInt.h:1810
LLVM_ABI APInt sadd_sat(const APInt &RHS) const
Definition APInt.cpp:2037
bool isAllOnes() const
Determine if all bits are set. This is true for zero-width values.
Definition APInt.h:372
bool ugt(const APInt &RHS) const
Unsigned greater than comparison.
Definition APInt.h:1189
static APInt getBitsSet(unsigned numBits, unsigned loBit, unsigned hiBit)
Get a value with a block of bits set.
Definition APInt.h:259
bool isZero() const
Determine if this value is zero, i.e. all bits are clear.
Definition APInt.h:381
LLVM_ABI APInt urem(const APInt &RHS) const
Unsigned remainder operation.
Definition APInt.cpp:1677
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition APInt.h:1503
bool ult(const APInt &RHS) const
Unsigned less than comparison.
Definition APInt.h:1118
static APInt getSignedMaxValue(unsigned numBits)
Gets maximum signed value of APInt for a specific bit width.
Definition APInt.h:210
bool isNegative() const
Determine sign of this APInt.
Definition APInt.h:330
LLVM_ABI APInt sdiv(const APInt &RHS) const
Signed division function for APInt.
Definition APInt.cpp:1655
void clearAllBits()
Set every bit to 0.
Definition APInt.h:1411
LLVM_ABI APInt rotr(unsigned rotateAmt) const
Rotate right by rotateAmt.
Definition APInt.cpp:1165
LLVM_ABI APInt reverseBits() const
Definition APInt.cpp:768
void ashrInPlace(unsigned ShiftAmt)
Arithmetic right-shift this APInt by ShiftAmt in place.
Definition APInt.h:841
bool sle(const APInt &RHS) const
Signed less or equal comparison.
Definition APInt.h:1173
unsigned countr_zero() const
Count the number of trailing zero bits.
Definition APInt.h:1654
unsigned getNumSignBits() const
Computes the number of leading bits of this APInt that are equal to its sign bit.
Definition APInt.h:1643
unsigned countl_zero() const
The APInt version of std::countl_zero.
Definition APInt.h:1613
static LLVM_ABI APInt getSplat(unsigned NewLen, const APInt &V)
Return a value containing V broadcasted over NewLen bits.
Definition APInt.cpp:651
static APInt getSignedMinValue(unsigned numBits)
Gets minimum signed value of APInt for a specific bit width.
Definition APInt.h:220
LLVM_ABI APInt sshl_sat(const APInt &RHS) const
Definition APInt.cpp:2097
LLVM_ABI APInt ushl_sat(const APInt &RHS) const
Definition APInt.cpp:2111
LLVM_ABI APInt sextOrTrunc(unsigned width) const
Sign extend or truncate to width.
Definition APInt.cpp:1052
static bool isSameValue(const APInt &I1, const APInt &I2, bool SignedCompare=false)
Determine if two APInts have the same value, after zero-extending or sign-extending (if SignedCompare...
Definition APInt.h:555
LLVM_ABI APInt rotl(unsigned rotateAmt) const
Rotate left by rotateAmt.
Definition APInt.cpp:1152
LLVM_ABI void insertBits(const APInt &SubBits, unsigned bitPosition)
Insert the bits from a smaller APInt starting at bitPosition.
Definition APInt.cpp:397
void clearLowBits(unsigned loBits)
Set bottom loBits bits to 0.
Definition APInt.h:1450
unsigned logBase2() const
Definition APInt.h:1776
LLVM_ABI APInt uadd_sat(const APInt &RHS) const
Definition APInt.cpp:2047
APInt ashr(unsigned ShiftAmt) const
Arithmetic right-shift function.
Definition APInt.h:834
LLVM_ABI APInt multiplicativeInverse() const
Definition APInt.cpp:1285
LLVM_ABI APInt srem(const APInt &RHS) const
Function for signed remainder operation.
Definition APInt.cpp:1747
bool isNonNegative() const
Determine if this APInt Value is non-negative (>= 0)
Definition APInt.h:335
bool ule(const APInt &RHS) const
Unsigned less or equal comparison.
Definition APInt.h:1157
LLVM_ABI APInt sext(unsigned width) const
Sign extend to a new width.
Definition APInt.cpp:996
void setBits(unsigned loBit, unsigned hiBit)
Set the bits from loBit (inclusive) to hiBit (exclusive) to 1.
Definition APInt.h:1382
APInt shl(unsigned shiftAmt) const
Left-shift function.
Definition APInt.h:880
LLVM_ABI APInt byteSwap() const
Definition APInt.cpp:746
bool isSubsetOf(const APInt &RHS) const
This operation checks that all bits set in this APInt are also set in RHS.
Definition APInt.h:1264
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
Definition APInt.h:441
static APInt getLowBitsSet(unsigned numBits, unsigned loBitsSet)
Constructs an APInt value that has the bottom loBitsSet bits set.
Definition APInt.h:307
void clearBits(unsigned LoBit, unsigned HiBit)
Clear the bits from LoBit (inclusive) to HiBit (exclusive) to 0.
Definition APInt.h:1432
static APInt getZero(unsigned numBits)
Get the '0' value for the specified bit-width.
Definition APInt.h:201
void setLowBits(unsigned loBits)
Set the bottom loBits bits.
Definition APInt.h:1403
LLVM_ABI APInt extractBits(unsigned numBits, unsigned bitPosition) const
Return an APInt with the extracted bits [bitPosition,bitPosition+numBits).
Definition APInt.cpp:482
bool sge(const APInt &RHS) const
Signed greater or equal comparison.
Definition APInt.h:1244
bool isOne() const
Determine if this is a value of 1.
Definition APInt.h:390
static APInt getBitsSetFrom(unsigned numBits, unsigned loBit)
Constructs an APInt value that has a contiguous range of bits set.
Definition APInt.h:287
static APInt getOneBitSet(unsigned numBits, unsigned BitNo)
Return an APInt with exactly one bit set in the result.
Definition APInt.h:240
void lshrInPlace(unsigned ShiftAmt)
Logical right-shift this APInt by ShiftAmt in place.
Definition APInt.h:865
APInt lshr(unsigned shiftAmt) const
Logical right-shift function.
Definition APInt.h:858
bool uge(const APInt &RHS) const
Unsigned greater or equal comparison.
Definition APInt.h:1228
LLVM_ABI APInt ssub_sat(const APInt &RHS) const
Definition APInt.cpp:2056
An arbitrary precision integer that knows its signedness.
Definition APSInt.h:24
unsigned getSrcAddressSpace() const
unsigned getDestAddressSpace() const
static Capacity get(size_t N)
Get the capacity of an array that can hold at least N elements.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
size_t size() const
size - Get the array size.
Definition ArrayRef.h:142
bool empty() const
empty - Check if the array is empty.
Definition ArrayRef.h:137
This is an SDNode representing atomic operations.
static LLVM_ABI BaseIndexOffset match(const SDNode *N, const SelectionDAG &DAG)
Parses tree in N for base, index, offset addresses.
This class is a wrapper over an AAResults, and it is intended to be used only when there are no IR ch...
bool pointsToConstantMemory(const MemoryLocation &Loc, bool OrLocal=false)
BitVector & reset()
Definition BitVector.h:411
void resize(unsigned N, bool t=false)
resize - Grow or shrink the bitvector.
Definition BitVector.h:360
void clear()
clear - Removes all bits from the bitvector.
Definition BitVector.h:354
BitVector & set()
Definition BitVector.h:370
bool none() const
none - Returns true if none of the bits are set.
Definition BitVector.h:207
size_type size() const
size - Returns the number of bits in this bitvector.
Definition BitVector.h:178
const BlockAddress * getBlockAddress() const
The address of a basic block.
Definition Constants.h:904
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
A "pseudo-class" with methods for operating on BUILD_VECTORs.
LLVM_ABI bool getConstantRawBits(bool IsLittleEndian, unsigned DstEltSizeInBits, SmallVectorImpl< APInt > &RawBitElements, BitVector &UndefElements) const
Extract the raw bit data from a build vector of Undef, Constant or ConstantFP node elements.
static LLVM_ABI void recastRawBits(bool IsLittleEndian, unsigned DstEltSizeInBits, SmallVectorImpl< APInt > &DstBitElements, ArrayRef< APInt > SrcBitElements, BitVector &DstUndefElements, const BitVector &SrcUndefElements)
Recast bit data SrcBitElements to DstEltSizeInBits wide elements.
LLVM_ABI bool getRepeatedSequence(const APInt &DemandedElts, SmallVectorImpl< SDValue > &Sequence, BitVector *UndefElements=nullptr) const
Find the shortest repeating sequence of values in the build vector.
LLVM_ABI ConstantFPSDNode * getConstantFPSplatNode(const APInt &DemandedElts, BitVector *UndefElements=nullptr) const
Returns the demanded splatted constant FP or null if this is not a constant FP splat.
LLVM_ABI SDValue getSplatValue(const APInt &DemandedElts, BitVector *UndefElements=nullptr) const
Returns the demanded splatted value or a null value if this is not a splat.
LLVM_ABI bool isConstantSplat(APInt &SplatValue, APInt &SplatUndef, unsigned &SplatBitSize, bool &HasAnyUndefs, unsigned MinSplatBits=0, bool isBigEndian=false) const
Check if this is a constant splat, and if so, find the smallest element size that splats the vector.
LLVM_ABI ConstantSDNode * getConstantSplatNode(const APInt &DemandedElts, BitVector *UndefElements=nullptr) const
Returns the demanded splatted constant or null if this is not a constant splat.
LLVM_ABI int32_t getConstantFPSplatPow2ToLog2Int(BitVector *UndefElements, uint32_t BitWidth) const
If this is a constant FP splat and the splatted constant FP is an exact power or 2,...
LLVM_ABI std::optional< std::pair< APInt, APInt > > isArithmeticSequence() const
If this BuildVector is constant and represents an arithmetic sequence "<a, a+n, a+2n,...
LLVM_ABI bool isConstant() const
This class represents a function call, abstracting a target machine's calling convention.
bool isTailCall() const
static LLVM_ABI bool isValueValidForType(EVT VT, const APFloat &Val)
const APFloat & getValueAPF() const
bool isExactlyValue(double V) const
We don't rely on operator== working on double values, as it returns true for things that are clearly ...
ConstantFP - Floating Point Values [float, double].
Definition Constants.h:282
const APFloat & getValue() const
Definition Constants.h:326
This is the shared class of boolean and integer constants.
Definition Constants.h:87
unsigned getBitWidth() const
getBitWidth - Return the scalar bitwidth of this constant.
Definition Constants.h:162
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:159
MachineConstantPoolValue * getMachineCPVal() const
const Constant * getConstVal() const
LLVM_ABI Type * getType() const
This class represents a range of values.
LLVM_ABI ConstantRange multiply(const ConstantRange &Other) const
Return a new range representing the possible values resulting from a multiplication of a value in thi...
const APInt * getSingleElement() const
If this set contains a single element, return it, otherwise return null.
static LLVM_ABI ConstantRange fromKnownBits(const KnownBits &Known, bool IsSigned)
Initialize a range based on a known bits constraint.
LLVM_ABI OverflowResult unsignedSubMayOverflow(const ConstantRange &Other) const
Return whether unsigned sub of the two ranges always/never overflows.
LLVM_ABI OverflowResult unsignedAddMayOverflow(const ConstantRange &Other) const
Return whether unsigned add of the two ranges always/never overflows.
LLVM_ABI KnownBits toKnownBits() const
Return known bits for values in this range.
LLVM_ABI ConstantRange zeroExtend(uint32_t BitWidth) const
Return a new range in the specified integer type, which must be strictly larger than the current type...
LLVM_ABI APInt getSignedMin() const
Return the smallest signed value contained in the ConstantRange.
LLVM_ABI OverflowResult unsignedMulMayOverflow(const ConstantRange &Other) const
Return whether unsigned mul of the two ranges always/never overflows.
LLVM_ABI ConstantRange signExtend(uint32_t BitWidth) const
Return a new range in the specified integer type, which must be strictly larger than the current type...
LLVM_ABI bool contains(const APInt &Val) const
Return true if the specified value is in the set.
LLVM_ABI APInt getUnsignedMax() const
Return the largest unsigned value contained in the ConstantRange.
LLVM_ABI APInt getSignedMax() const
Return the largest signed value contained in the ConstantRange.
OverflowResult
Represents whether an operation on the given constant range is known to always or never overflow.
@ AlwaysOverflowsHigh
Always overflows in the direction of signed/unsigned max value.
@ AlwaysOverflowsLow
Always overflows in the direction of signed/unsigned min value.
@ MayOverflow
May or may not overflow.
uint32_t getBitWidth() const
Get the bit width of this ConstantRange.
LLVM_ABI OverflowResult signedSubMayOverflow(const ConstantRange &Other) const
Return whether signed sub of the two ranges always/never overflows.
uint64_t getZExtValue() const
const APInt & getAPIntValue() const
This is an important base class in LLVM.
Definition Constant.h:43
LLVM_ABI Constant * getSplatValue(bool AllowPoison=false) const
If all elements of the vector constant have the same value, return that value.
LLVM_ABI Constant * getAggregateElement(unsigned Elt) const
For aggregates (struct/array/vector) return the constant that corresponds to the specified element if...
DWARF expression.
static LLVM_ABI ExtOps getExtOps(unsigned FromSize, unsigned ToSize, bool Signed)
Returns the ops for a zero- or sign-extension in a DIExpression.
static LLVM_ABI void appendOffset(SmallVectorImpl< uint64_t > &Ops, int64_t Offset)
Append Ops with operations to apply the Offset.
static LLVM_ABI DIExpression * appendOpsToArg(const DIExpression *Expr, ArrayRef< uint64_t > Ops, unsigned ArgNo, bool StackValue=false)
Create a copy of Expr by appending the given list of Ops to each instance of the operand DW_OP_LLVM_a...
static LLVM_ABI const DIExpression * convertToVariadicExpression(const DIExpression *Expr)
If Expr is a non-variadic expression (i.e.
static LLVM_ABI std::optional< DIExpression * > createFragmentExpression(const DIExpression *Expr, unsigned OffsetInBits, unsigned SizeInBits)
Create a DIExpression to describe one part of an aggregate variable that is fragmented across multipl...
Base class for variables.
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
bool isLittleEndian() const
Layout endianness...
Definition DataLayout.h:214
LLVM_ABI IntegerType * getIntPtrType(LLVMContext &C, unsigned AddressSpace=0) const
Returns an integer type with size at least as big as that of a pointer in the given address space.
LLVM_ABI Align getABITypeAlign(Type *Ty) const
Returns the minimum ABI-required alignment for the specified type.
LLVM_ABI unsigned getPointerTypeSizeInBits(Type *) const
The pointer representation size in bits for this type.
LLVM_ABI Align getPrefTypeAlign(Type *Ty) const
Returns the preferred stack/global alignment for the specified type.
A debug info location.
Definition DebugLoc.h:123
Implements a dense probed hash-table based set.
Definition DenseSet.h:279
const char * getSymbol() const
FoldingSetNodeID - This class is used to gather all the unique data bits of a node.
Definition FoldingSet.h:209
Data structure describing the variable locations in a function.
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Definition Function.h:711
AttributeList getAttributes() const
Return the attribute list for this Function.
Definition Function.h:354
LLVM_ABI unsigned getAddressSpace() const
const GlobalValue * getGlobal() const
bool isThreadLocal() const
If the value is "Thread Local", its value isn't shared by the threads.
unsigned getAddressSpace() const
Module * getParent()
Get the module that this global value is contained inside of...
PointerType * getType() const
Global values are always pointers.
This class is used to form a handle around another node that is persistent and is updated across invo...
const SDValue & getValue() const
static LLVM_ABI bool compare(const APInt &LHS, const APInt &RHS, ICmpInst::Predicate Pred)
Return result of LHS Pred RHS comparison.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
Tracks which library functions to use for a particular subtarget.
LLVM_ABI CallingConv::ID getLibcallImplCallingConv(RTLIB::LibcallImpl Call) const
Get the CallingConv that should be used for the specified libcall.
LLVM_ABI RTLIB::LibcallImpl getLibcallImpl(RTLIB::Libcall Call) const
Return the lowering's selection of implementation call for Call.
This SDNode is used for LIFETIME_START/LIFETIME_END values.
This class is used to represent ISD::LOAD nodes.
static LocationSize precise(uint64_t Value)
MCSymbol - Instances of this class represent a symbol name in the MC file, and MCSymbols are created ...
Definition MCSymbol.h:42
Metadata node.
Definition Metadata.h:1080
const MDOperand & getOperand(unsigned I) const
Definition Metadata.h:1444
Machine Value Type.
SimpleValueType SimpleTy
static MVT getIntegerVT(unsigned BitWidth)
Abstract base class for all machine specific constantpool value subclasses.
virtual void addSelectionDAGCSEId(FoldingSetNodeID &ID)=0
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
void setObjectAlignment(int ObjectIdx, Align Alignment)
setObjectAlignment - Change the alignment of the specified stack object.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
Function & getFunction()
Return the LLVM function that this machine code represents.
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
A description of a memory reference used in the backend.
const MDNode * getRanges() const
Return the range tag for the memory reference.
Flags
Flags values. These may be or'd together.
@ MOVolatile
The memory access is volatile.
@ MODereferenceable
The memory access is dereferenceable (i.e., doesn't trap).
@ MOLoad
The memory access reads data.
@ MOInvariant
The memory access always returns the same value (or traps).
@ MOStore
The memory access writes data.
const MachinePointerInfo & getPointerInfo() const
Flags getFlags() const
Return the raw flags of the source value,.
This class contains meta information specific to a module.
An SDNode that represents everything that will be needed to construct a MachineInstr.
This class is used to represent an MGATHER node.
This class is used to represent an MLOAD node.
This class is used to represent an MSCATTER node.
This class is used to represent an MSTORE node.
This SDNode is used for target intrinsics that touch memory and need an associated MachineMemOperand.
size_t getNumMemOperands() const
Return the number of memory operands.
LLVM_ABI MemSDNode(unsigned Opc, unsigned Order, const DebugLoc &dl, SDVTList VTs, EVT memvt, PointerUnion< MachineMemOperand *, MachineMemOperand ** > memrefs)
Constructor that supports single or multiple MMOs.
PointerUnion< MachineMemOperand *, MachineMemOperand ** > MemRefs
Memory reference information.
MachineMemOperand * getMemOperand() const
Return the unique MachineMemOperand object describing the memory reference performed by operation.
const MachinePointerInfo & getPointerInfo() const
ArrayRef< MachineMemOperand * > memoperands() const
Return the memory operands for this node.
unsigned getRawSubclassData() const
Return the SubclassData value, without HasDebugValue.
EVT getMemoryVT() const
Return the type of the in-memory value.
Representation for a specific memory location.
A Module instance is used to store all the information related to an LLVM module.
Definition Module.h:67
Function * getFunction(StringRef Name) const
Look up the specified function in the module symbol table.
Definition Module.cpp:235
MutableArrayRef - Represent a mutable reference to an array (0 or more elements consecutively in memo...
Definition ArrayRef.h:298
The optimization diagnostic interface.
Pass interface - Implemented by all 'passes'.
Definition Pass.h:99
Class to represent pointers.
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
unsigned getAddressSpace() const
Return the address space of the Pointer type.
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
A discriminated union of two or more pointer types, with the discriminator in the low bit of the poin...
bool isNull() const
Test if the pointer held in the union is null, regardless of which type it is.
Analysis providing profile information.
void Deallocate(SubClass *E)
Deallocate - Release storage for the pointed-to object.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
Keeps track of dbg_value information through SDISel.
LLVM_ABI void add(SDDbgValue *V, bool isParameter)
LLVM_ABI void erase(const SDNode *Node)
Invalidate all DbgValues attached to the node and remove it from the Node-to-DbgValues map.
Holds the information from a dbg_label node through SDISel.
Holds the information for a single machine location through SDISel; either an SDNode,...
static SDDbgOperand fromNode(SDNode *Node, unsigned ResNo)
static SDDbgOperand fromFrameIdx(unsigned FrameIdx)
static SDDbgOperand fromVReg(Register VReg)
static SDDbgOperand fromConst(const Value *Const)
@ SDNODE
Value is the result of an expression.
Holds the information from a dbg_value node through SDISel.
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
const DebugLoc & getDebugLoc() const
unsigned getIROrder() const
This class provides iterator support for SDUse operands that use a specific SDNode.
Represents one node in the SelectionDAG.
ArrayRef< SDUse > ops() const
const APInt & getAsAPIntVal() const
Helper method returns the APInt value of a ConstantSDNode.
LLVM_ABI void dumprFull(const SelectionDAG *G=nullptr) const
printrFull to dbgs().
unsigned getOpcode() const
Return the SelectionDAG opcode value for this node.
bool isDivergent() const
LLVM_ABI bool isOnlyUserOf(const SDNode *N) const
Return true if this node is the only use of N.
iterator_range< value_op_iterator > op_values() const
unsigned getIROrder() const
Return the node ordering.
static constexpr size_t getMaxNumOperands()
Return the maximum number of operands that a SDNode can hold.
iterator_range< use_iterator > uses()
MemSDNodeBitfields MemSDNodeBits
LLVM_ABI void Profile(FoldingSetNodeID &ID) const
Gather unique data for the node.
bool getHasDebugValue() const
SDNodeFlags getFlags() const
void setNodeId(int Id)
Set unique node id.
LLVM_ABI void intersectFlagsWith(const SDNodeFlags Flags)
Clear any flags in this node that aren't also set in Flags.
static bool hasPredecessorHelper(const SDNode *N, SmallPtrSetImpl< const SDNode * > &Visited, SmallVectorImpl< const SDNode * > &Worklist, unsigned int MaxSteps=0, bool TopologicalPrune=false)
Returns true if N is a predecessor of any node in Worklist.
uint64_t getAsZExtVal() const
Helper method returns the zero-extended integer value of a ConstantSDNode.
bool use_empty() const
Return true if there are no uses of this node.
unsigned getNumValues() const
Return the number of values defined/returned by this operator.
unsigned getNumOperands() const
Return the number of values used by this operation.
const SDValue & getOperand(unsigned Num) const
static LLVM_ABI bool areOnlyUsersOf(ArrayRef< const SDNode * > Nodes, const SDNode *N)
Return true if all the users of N are contained in Nodes.
use_iterator use_begin() const
Provide iteration support to walk over all uses of an SDNode.
LLVM_ABI bool isOperandOf(const SDNode *N) const
Return true if this node is an operand of N.
const APInt & getConstantOperandAPInt(unsigned Num) const
Helper method returns the APInt of a ConstantSDNode operand.
std::optional< APInt > bitcastToAPInt() const
LLVM_ABI bool hasPredecessor(const SDNode *N) const
Return true if N is a predecessor of this node.
LLVM_ABI bool hasAnyUseOfValue(unsigned Value) const
Return true if there are any use of the indicated value.
EVT getValueType(unsigned ResNo) const
Return the type of a specified result.
bool isUndef() const
Returns true if the node type is UNDEF or POISON.
op_iterator op_end() const
op_iterator op_begin() const
static use_iterator use_end()
LLVM_ABI void DropOperands()
Release the operands and set this node to have zero operands.
SDNode(unsigned Opc, unsigned Order, DebugLoc dl, SDVTList VTs)
Create an SDNode.
Represents a use of a SDNode.
SDNode * getUser()
This returns the SDNode that contains this Use.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
bool isUndef() const
SDNode * getNode() const
get the SDNode which holds the desired result
bool hasOneUse() const
Return true if there is exactly one node using value ResNo of Node.
LLVM_ABI bool isOperandOf(const SDNode *N) const
Return true if the referenced return value is an operand of N.
SDValue()=default
LLVM_ABI bool reachesChainWithoutSideEffects(SDValue Dest, unsigned Depth=2) const
Return true if this operand (which must be a chain) reaches the specified operand without crossing an...
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
TypeSize getValueSizeInBits() const
Returns the size of the value in bits.
const SDValue & getOperand(unsigned i) const
bool use_empty() const
Return true if there are no nodes using value ResNo of Node.
const APInt & getConstantOperandAPInt(unsigned i) const
uint64_t getScalarValueSizeInBits() const
unsigned getResNo() const
get the index which selects a specific result in the SDNode
uint64_t getConstantOperandVal(unsigned i) const
unsigned getOpcode() const
virtual void verifyTargetNode(const SelectionDAG &DAG, const SDNode *N) const
Checks that the given target-specific node is valid. Aborts if it is not.
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
LLVM_ABI SDValue getElementCount(const SDLoc &DL, EVT VT, ElementCount EC)
LLVM_ABI Align getReducedAlign(EVT VT, bool UseABI)
In most cases this function returns the ABI alignment for a given type, except for illegal vector typ...
LLVM_ABI SDValue getVPZeroExtendInReg(SDValue Op, SDValue Mask, SDValue EVL, const SDLoc &DL, EVT VT)
Return the expression required to zero extend the Op value assuming it was the smaller SrcTy value.
LLVM_ABI SDValue getShiftAmountOperand(EVT LHSTy, SDValue Op)
Return the specified value casted to the target's desired shift amount type.
LLVM_ABI SDValue getExtLoad(ISD::LoadExtType ExtType, const SDLoc &dl, EVT VT, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, EVT MemVT, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI std::pair< SDValue, SDValue > getMemccpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue C, SDValue Size, const CallInst *CI)
Lower a memccpy operation into a target library call and return the resulting chain and call result a...
LLVM_ABI SDValue getExtLoadVP(ISD::LoadExtType ExtType, const SDLoc &dl, EVT VT, SDValue Chain, SDValue Ptr, SDValue Mask, SDValue EVL, MachinePointerInfo PtrInfo, EVT MemVT, MaybeAlign Alignment, MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo, bool IsExpanding=false)
SDValue getExtractVectorElt(const SDLoc &DL, EVT VT, SDValue Vec, unsigned Idx)
Extract element at Idx from Vec.
LLVM_ABI SDValue getSplatSourceVector(SDValue V, int &SplatIndex)
If V is a splatted value, return the source vector and its splat index.
LLVM_ABI SDValue getLabelNode(unsigned Opcode, const SDLoc &dl, SDValue Root, MCSymbol *Label)
LLVM_ABI OverflowKind computeOverflowForUnsignedSub(SDValue N0, SDValue N1) const
Determine if the result of the unsigned sub of 2 nodes can overflow.
LLVM_ABI unsigned ComputeMaxSignificantBits(SDValue Op, unsigned Depth=0) const
Get the upper bound on bit size for this Value Op as a signed integer.
const SDValue & getRoot() const
Return the root tag of the SelectionDAG.
LLVM_ABI std::pair< SDValue, SDValue > getStrlen(SDValue Chain, const SDLoc &dl, SDValue Src, const CallInst *CI)
Lower a strlen operation into a target library call and return the resulting chain and call result as...
LLVM_ABI SDValue getMaskedGather(SDVTList VTs, EVT MemVT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType, ISD::LoadExtType ExtTy)
LLVM_ABI SDValue getAddrSpaceCast(const SDLoc &dl, EVT VT, SDValue Ptr, unsigned SrcAS, unsigned DestAS)
Return an AddrSpaceCastSDNode.
LLVM_ABI SDValue FoldSetCC(EVT VT, SDValue N1, SDValue N2, ISD::CondCode Cond, const SDLoc &dl, SDNodeFlags Flags={})
Constant fold a setcc to true or false.
bool isKnownNeverSNaN(SDValue Op, const APInt &DemandedElts, unsigned Depth=0) const
LLVM_ABI std::optional< bool > isBoolConstant(SDValue N) const
Check if a value \op N is a constant using the target's BooleanContent for its type.
LLVM_ABI SDValue getStackArgumentTokenFactor(SDValue Chain)
Compute a TokenFactor to force all the incoming stack arguments to be loaded from the stack.
const TargetSubtargetInfo & getSubtarget() const
LLVM_ABI SDValue getMergeValues(ArrayRef< SDValue > Ops, const SDLoc &dl)
Create a MERGE_VALUES node from the given operands.
LLVM_ABI SDVTList getVTList(EVT VT)
Return an SDVTList that represents the list of values specified.
LLVM_ABI SDValue getShiftAmountConstant(uint64_t Val, EVT VT, const SDLoc &DL)
LLVM_ABI void updateDivergence(SDNode *N)
LLVM_ABI SDValue getSplatValue(SDValue V, bool LegalTypes=false)
If V is a splat vector, return its scalar source operand by extracting that element from the source v...
LLVM_ABI SDValue getAllOnesConstant(const SDLoc &DL, EVT VT, bool IsTarget=false, bool IsOpaque=false)
LLVM_ABI MachineSDNode * getMachineNode(unsigned Opcode, const SDLoc &dl, EVT VT)
These are used for target selectors to create a new node with specified return type(s),...
LLVM_ABI void ExtractVectorElements(SDValue Op, SmallVectorImpl< SDValue > &Args, unsigned Start=0, unsigned Count=0, EVT EltVT=EVT())
Append the extracted elements from Start to Count out of the vector Op in Args.
LLVM_ABI SDValue getNeutralElement(unsigned Opcode, const SDLoc &DL, EVT VT, SDNodeFlags Flags)
Get the (commutative) neutral element for the given opcode, if it exists.
LLVM_ABI SDValue getAtomicMemset(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Value, SDValue Size, Type *SizeTy, unsigned ElemSz, bool isTailCall, MachinePointerInfo DstPtrInfo)
LLVM_ABI SDValue getAtomicLoad(ISD::LoadExtType ExtType, const SDLoc &dl, EVT MemVT, EVT VT, SDValue Chain, SDValue Ptr, MachineMemOperand *MMO)
LLVM_ABI SDNode * getNodeIfExists(unsigned Opcode, SDVTList VTList, ArrayRef< SDValue > Ops, const SDNodeFlags Flags, bool AllowCommute=false)
Get the specified node if it's already available, or else return NULL.
LLVM_ABI SDValue getPseudoProbeNode(const SDLoc &Dl, SDValue Chain, uint64_t Guid, uint64_t Index, uint32_t Attr)
Creates a PseudoProbeSDNode with function GUID Guid and the index of the block Index it is probing,...
LLVM_ABI SDValue getFreeze(SDValue V)
Return a freeze using the SDLoc of the value operand.
LLVM_ABI SDNode * SelectNodeTo(SDNode *N, unsigned MachineOpc, EVT VT)
These are used for target selectors to mutate the specified node to have the specified return type,...
LLVM_ABI void init(MachineFunction &NewMF, OptimizationRemarkEmitter &NewORE, Pass *PassPtr, const TargetLibraryInfo *LibraryInfo, const LibcallLoweringInfo *LibcallsInfo, UniformityInfo *UA, ProfileSummaryInfo *PSIin, BlockFrequencyInfo *BFIin, MachineModuleInfo &MMI, FunctionVarLocs const *FnVarLocs)
Prepare this SelectionDAG to process code in the given MachineFunction.
LLVM_ABI SelectionDAG(const TargetMachine &TM, CodeGenOptLevel)
LLVM_ABI SDValue getMemset(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI, MachinePointerInfo DstPtrInfo, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI SDValue getBitcastedSExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by first bitcasting (from potentia...
LLVM_ABI SDValue getConstantPool(const Constant *C, EVT VT, MaybeAlign Align=std::nullopt, int Offs=0, bool isT=false, unsigned TargetFlags=0)
LLVM_ABI SDValue getStridedLoadVP(ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &DL, SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, bool IsExpanding=false)
LLVM_ABI SDValue getAtomicCmpSwap(unsigned Opcode, const SDLoc &dl, EVT MemVT, SDVTList VTs, SDValue Chain, SDValue Ptr, SDValue Cmp, SDValue Swp, MachineMemOperand *MMO)
Gets a node for an atomic cmpxchg op.
LLVM_ABI SDValue makeEquivalentMemoryOrdering(SDValue OldChain, SDValue NewMemOpChain)
If an existing load has uses of its chain, create a token factor node with that chain and the new mem...
LLVM_ABI bool isConstantIntBuildVectorOrConstantInt(SDValue N, bool AllowOpaques=true) const
Test whether the given value is a constant int or similar node.
LLVM_ABI void ReplaceAllUsesOfValuesWith(const SDValue *From, const SDValue *To, unsigned Num)
Like ReplaceAllUsesOfValueWith, but for multiple values at once.
LLVM_ABI SDValue getJumpTableDebugInfo(int JTI, SDValue Chain, const SDLoc &DL)
LLVM_ABI SDValue getSymbolFunctionGlobalAddress(SDValue Op, Function **TargetFunction=nullptr)
Return a GlobalAddress of the function from the current module with name matching the given ExternalS...
LLVM_ABI std::optional< unsigned > getValidMaximumShiftAmount(SDValue V, const APInt &DemandedElts, unsigned Depth=0) const
If a SHL/SRA/SRL node V has shift amounts that are all less than the element bit-width of the shift n...
LLVM_ABI SDValue UnrollVectorOp(SDNode *N, unsigned ResNE=0)
Utility function used by legalize and lowering to "unroll" a vector operation by splitting out the sc...
LLVM_ABI SDValue getVScale(const SDLoc &DL, EVT VT, APInt MulImm)
Return a node that represents the runtime scaling 'MulImm * RuntimeVL'.
LLVM_ABI SDValue getConstantFP(double Val, const SDLoc &DL, EVT VT, bool isTarget=false)
Create a ConstantFPSDNode wrapping a constant value.
OverflowKind
Used to represent the possible overflow behavior of an operation.
static LLVM_ABI unsigned getHasPredecessorMaxSteps()
LLVM_ABI bool haveNoCommonBitsSet(SDValue A, SDValue B) const
Return true if A and B have no common bits set.
SDValue getExtractSubvector(const SDLoc &DL, EVT VT, SDValue Vec, unsigned Idx)
Return the VT typed sub-vector of Vec at Idx.
LLVM_ABI bool cannotBeOrderedNegativeFP(SDValue Op) const
Test whether the given float value is known to be positive.
LLVM_ABI SDValue getRegister(Register Reg, EVT VT)
LLVM_ABI bool calculateDivergence(SDNode *N)
LLVM_ABI std::pair< SDValue, SDValue > getStrcmp(SDValue Chain, const SDLoc &dl, SDValue S0, SDValue S1, const CallInst *CI)
Lower a strcmp operation into a target library call and return the resulting chain and call result as...
LLVM_ABI SDValue getGetFPEnv(SDValue Chain, const SDLoc &dl, SDValue Ptr, EVT MemVT, MachineMemOperand *MMO)
LLVM_ABI SDValue getAssertAlign(const SDLoc &DL, SDValue V, Align A)
Return an AssertAlignSDNode.
LLVM_ABI SDNode * mutateStrictFPToFP(SDNode *Node)
Mutate the specified strict FP node to its non-strict equivalent, unlinking the node from its chain a...
LLVM_ABI SDValue getLoad(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr)
Loads are not normal binary operators: their result type is not determined by their operands,...
LLVM_ABI bool canIgnoreSignBitOfZero(const SDUse &Use) const
Check if a use of a float value is insensitive to signed zeros.
LLVM_ABI bool SignBitIsZeroFP(SDValue Op, unsigned Depth=0) const
Return true if the sign bit of Op is known to be zero, for a floating-point value.
LLVM_ABI SDValue getMemIntrinsicNode(unsigned Opcode, const SDLoc &dl, SDVTList VTList, ArrayRef< SDValue > Ops, EVT MemVT, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags Flags=MachineMemOperand::MOLoad|MachineMemOperand::MOStore, LocationSize Size=LocationSize::precise(0), const AAMDNodes &AAInfo=AAMDNodes())
Creates a MemIntrinsicNode that may produce a result and takes a list of operands.
SDValue getInsertSubvector(const SDLoc &DL, SDValue Vec, SDValue SubVec, unsigned Idx)
Insert SubVec at the Idx element of Vec.
LLVM_ABI SDValue getBitcastedZExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by first bitcasting (from potentia...
LLVM_ABI SDValue getStepVector(const SDLoc &DL, EVT ResVT, const APInt &StepVal)
Returns a vector of type ResVT whose elements contain the linear sequence <0, Step,...
SDValue getSetCC(const SDLoc &DL, EVT VT, SDValue LHS, SDValue RHS, ISD::CondCode Cond, SDValue Chain=SDValue(), bool IsSignaling=false, SDNodeFlags Flags={})
Helper function to make it easier to build SetCC's if you just have an ISD::CondCode instead of an SD...
LLVM_ABI SDValue getAtomic(unsigned Opcode, const SDLoc &dl, EVT MemVT, SDValue Chain, SDValue Ptr, SDValue Val, MachineMemOperand *MMO)
Gets a node for an atomic op, produces result (if relevant) and chain and takes 2 operands.
LLVM_ABI SDValue getMemcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI, std::optional< bool > OverrideTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo=AAMDNodes(), BatchAAResults *BatchAA=nullptr)
LLVM_ABI Align getEVTAlign(EVT MemoryVT) const
Compute the default alignment value for the given type.
LLVM_ABI bool shouldOptForSize() const
LLVM_ABI SDValue getNOT(const SDLoc &DL, SDValue Val, EVT VT)
Create a bitwise NOT operation as (XOR Val, -1).
LLVM_ABI SDValue getVPZExtOrTrunc(const SDLoc &DL, EVT VT, SDValue Op, SDValue Mask, SDValue EVL)
Convert a vector-predicated Op, which must be an integer vector, to the vector-type VT,...
const TargetLowering & getTargetLoweringInfo() const
LLVM_ABI bool isEqualTo(SDValue A, SDValue B) const
Test whether two SDValues are known to compare equal.
static constexpr unsigned MaxRecursionDepth
LLVM_ABI SDValue getStridedStoreVP(SDValue Chain, const SDLoc &DL, SDValue Val, SDValue Ptr, SDValue Offset, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, bool IsTruncating=false, bool IsCompressing=false)
bool isGuaranteedNotToBePoison(SDValue Op, unsigned Depth=0) const
Return true if this function can prove that Op is never poison.
LLVM_ABI SDValue expandVACopy(SDNode *Node)
Expand the specified ISD::VACOPY node as the Legalize pass would.
LLVM_ABI SDValue getIndexedMaskedLoad(SDValue OrigLoad, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI void dump(bool Sorted=false) const
Dump the textual format of this DAG.
LLVM_ABI APInt computeVectorKnownZeroElements(SDValue Op, const APInt &DemandedElts, unsigned Depth=0) const
For each demanded element of a vector, see if it is known to be zero.
LLVM_ABI void AddDbgValue(SDDbgValue *DB, bool isParameter)
Add a dbg_value SDNode.
bool NewNodesMustHaveLegalTypes
When true, additional steps are taken to ensure that getConstant() and similar functions return DAG n...
LLVM_ABI std::pair< EVT, EVT > GetSplitDestVTs(const EVT &VT) const
Compute the VTs needed for the low/hi parts of a type which is split (or expanded) into two not neces...
LLVM_ABI void salvageDebugInfo(SDNode &N)
To be invoked on an SDNode that is slated to be erased.
LLVM_ABI SDNode * MorphNodeTo(SDNode *N, unsigned Opc, SDVTList VTs, ArrayRef< SDValue > Ops)
This mutates the specified node to have the specified return type, opcode, and operands.
LLVM_ABI std::pair< SDValue, SDValue > UnrollVectorOverflowOp(SDNode *N, unsigned ResNE=0)
Like UnrollVectorOp(), but for the [US](ADD|SUB|MUL)O family of opcodes.
allnodes_const_iterator allnodes_begin() const
SDValue getUNDEF(EVT VT)
Return an UNDEF node. UNDEF does not have a useful SDLoc.
LLVM_ABI SDValue getGatherVP(SDVTList VTs, EVT VT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType)
SDValue getBuildVector(EVT VT, const SDLoc &DL, ArrayRef< SDValue > Ops)
Return an ISD::BUILD_VECTOR node.
LLVM_ABI SDValue getBitcastedAnyExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by first bitcasting (from potentia...
LLVM_ABI bool isSplatValue(SDValue V, const APInt &DemandedElts, APInt &UndefElts, unsigned Depth=0) const
Test whether V has a splatted value for all the demanded elements.
LLVM_ABI void DeleteNode(SDNode *N)
Remove the specified node from the system.
LLVM_ABI SDValue getBitcast(EVT VT, SDValue V)
Return a bitcast using the SDLoc of the value operand, and casting to the provided type.
LLVM_ABI SDDbgValue * getDbgValueList(DIVariable *Var, DIExpression *Expr, ArrayRef< SDDbgOperand > Locs, ArrayRef< SDNode * > Dependencies, bool IsIndirect, const DebugLoc &DL, unsigned O, bool IsVariadic)
Creates a SDDbgValue node from a list of locations.
LLVM_ABI std::pair< SDValue, SDValue > getStrcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, const CallInst *CI)
Lower a strcpy operation into a target library call and return the resulting chain and call result as...
SDValue getSelect(const SDLoc &DL, EVT VT, SDValue Cond, SDValue LHS, SDValue RHS, SDNodeFlags Flags=SDNodeFlags())
Helper function to make it easier to build Select's if you just have operands and don't want to check...
LLVM_ABI SDValue getNegative(SDValue Val, const SDLoc &DL, EVT VT)
Create negative operation as (SUB 0, Val).
LLVM_ABI std::optional< unsigned > getValidShiftAmount(SDValue V, const APInt &DemandedElts, unsigned Depth=0) const
If a SHL/SRA/SRL node V has a uniform shift amount that is less than the element bit-width of the shi...
LLVM_ABI void setNodeMemRefs(MachineSDNode *N, ArrayRef< MachineMemOperand * > NewMemRefs)
Mutate the specified machine node's memory references to the provided list.
LLVM_ABI SDValue simplifySelect(SDValue Cond, SDValue TVal, SDValue FVal)
Try to simplify a select/vselect into 1 of its operands or a constant.
LLVM_ABI SDValue getZeroExtendInReg(SDValue Op, const SDLoc &DL, EVT VT)
Return the expression required to zero extend the Op value assuming it was the smaller SrcTy value.
LLVM_ABI bool isConstantFPBuildVectorOrConstantFP(SDValue N) const
Test whether the given value is a constant FP or similar node.
const DataLayout & getDataLayout() const
LLVM_ABI SDValue expandVAArg(SDNode *Node)
Expand the specified ISD::VAARG node as the Legalize pass would.
LLVM_ABI SDValue getTokenFactor(const SDLoc &DL, SmallVectorImpl< SDValue > &Vals)
Creates a new TokenFactor containing Vals.
LLVM_ABI bool doesNodeExist(unsigned Opcode, SDVTList VTList, ArrayRef< SDValue > Ops)
Check if a node exists without modifying its flags.
const SelectionDAGTargetInfo & getSelectionDAGInfo() const
LLVM_ABI bool areNonVolatileConsecutiveLoads(LoadSDNode *LD, LoadSDNode *Base, unsigned Bytes, int Dist) const
Return true if loads are next to each other and can be merged.
LLVM_ABI SDValue getMaskedHistogram(SDVTList VTs, EVT MemVT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType)
LLVM_ABI SDDbgLabel * getDbgLabel(DILabel *Label, const DebugLoc &DL, unsigned O)
Creates a SDDbgLabel node.
LLVM_ABI SDValue getStoreVP(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, SDValue Offset, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, bool IsTruncating=false, bool IsCompressing=false)
LLVM_ABI OverflowKind computeOverflowForUnsignedMul(SDValue N0, SDValue N1) const
Determine if the result of the unsigned mul of 2 nodes can overflow.
LLVM_ABI void copyExtraInfo(SDNode *From, SDNode *To)
Copy extra info associated with one node to another.
LLVM_ABI SDValue getConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
Create a ConstantSDNode wrapping a constant value.
LLVM_ABI SDValue getMemBasePlusOffset(SDValue Base, TypeSize Offset, const SDLoc &DL, const SDNodeFlags Flags=SDNodeFlags())
Returns sum of the base pointer and offset.
LLVM_ABI SDValue getGlobalAddress(const GlobalValue *GV, const SDLoc &DL, EVT VT, int64_t offset=0, bool isTargetGA=false, unsigned TargetFlags=0)
LLVM_ABI SDValue getVAArg(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, SDValue SV, unsigned Align)
VAArg produces a result and token chain, and takes a pointer and a source value as input.
LLVM_ABI SDValue getTruncStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, EVT SVT, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI SDValue getLoadFFVP(EVT VT, const SDLoc &DL, SDValue Chain, SDValue Ptr, SDValue Mask, SDValue EVL, MachineMemOperand *MMO)
LLVM_ABI SDValue getTypeSize(const SDLoc &DL, EVT VT, TypeSize TS)
LLVM_ABI SDValue getMDNode(const MDNode *MD)
Return an MDNodeSDNode which holds an MDNode.
LLVM_ABI void clear()
Clear state and free memory necessary to make this SelectionDAG ready to process a new block.
LLVM_ABI std::pair< SDValue, SDValue > getMemcmp(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, const CallInst *CI)
Lower a memcmp operation into a target library call and return the resulting chain and call result as...
LLVM_ABI void ReplaceAllUsesWith(SDValue From, SDValue To)
Modify anything using 'From' to use 'To' instead.
LLVM_ABI SDValue getCommutedVectorShuffle(const ShuffleVectorSDNode &SV)
Returns an ISD::VECTOR_SHUFFLE node semantically equivalent to the shuffle node in input but with swa...
LLVM_ABI std::pair< SDValue, SDValue > SplitVector(const SDValue &N, const SDLoc &DL, const EVT &LoVT, const EVT &HiVT)
Split the vector with EXTRACT_SUBVECTOR using the provided VTs and return the low/high part.
LLVM_ABI SDValue makeStateFunctionCall(unsigned LibFunc, SDValue Ptr, SDValue InChain, const SDLoc &DLoc)
Helper used to make a call to a library function that has one argument of pointer type.
LLVM_ABI bool isGuaranteedNotToBeUndefOrPoison(SDValue Op, bool PoisonOnly=false, unsigned Depth=0) const
Return true if this function can prove that Op is never poison and, if PoisonOnly is false,...
LLVM_ABI SDValue getStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
Helper function to build ISD::STORE nodes.
LLVM_ABI SDValue getSignedConstant(int64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
LLVM_ABI SDValue getIndexedLoadVP(SDValue OrigLoad, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue getSrcValue(const Value *v)
Construct a node to track a Value* through the backend.
SDValue getSplatVector(EVT VT, const SDLoc &DL, SDValue Op)
LLVM_ABI SDValue getAtomicMemcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Type *SizeTy, unsigned ElemSz, bool isTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo)
LLVM_ABI OverflowKind computeOverflowForSignedMul(SDValue N0, SDValue N1) const
Determine if the result of the signed mul of 2 nodes can overflow.
LLVM_ABI MaybeAlign InferPtrAlign(SDValue Ptr) const
Infer alignment of a load / store address.
LLVM_ABI bool MaskedValueIsAllOnes(SDValue Op, const APInt &Mask, unsigned Depth=0) const
Return true if '(Op & Mask) == Mask'.
LLVM_ABI bool SignBitIsZero(SDValue Op, unsigned Depth=0) const
Return true if the sign bit of Op is known to be zero.
LLVM_ABI void RemoveDeadNodes()
This method deletes all unreachable nodes in the SelectionDAG.
LLVM_ABI void RemoveDeadNode(SDNode *N)
Remove the specified node from the system.
LLVM_ABI void AddDbgLabel(SDDbgLabel *DB)
Add a dbg_label SDNode.
bool isConstantValueOfAnyType(SDValue N) const
LLVM_ABI SDValue getTargetExtractSubreg(int SRIdx, const SDLoc &DL, EVT VT, SDValue Operand)
A convenience function for creating TargetInstrInfo::EXTRACT_SUBREG nodes.
LLVM_ABI SDValue getBasicBlock(MachineBasicBlock *MBB)
LLVM_ABI SDValue getSExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either sign-extending or trunca...
LLVM_ABI SDDbgValue * getVRegDbgValue(DIVariable *Var, DIExpression *Expr, Register VReg, bool IsIndirect, const DebugLoc &DL, unsigned O)
Creates a VReg SDDbgValue node.
LLVM_ABI SDValue getEHLabel(const SDLoc &dl, SDValue Root, MCSymbol *Label)
LLVM_ABI SDValue getIndexedStoreVP(SDValue OrigStore, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI bool isKnownNeverZero(SDValue Op, unsigned Depth=0) const
Test whether the given SDValue is known to contain non-zero value(s).
LLVM_ABI SDValue getIndexedStore(SDValue OrigStore, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue FoldConstantArithmetic(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops, SDNodeFlags Flags=SDNodeFlags())
LLVM_ABI std::optional< unsigned > getValidMinimumShiftAmount(SDValue V, const APInt &DemandedElts, unsigned Depth=0) const
If a SHL/SRA/SRL node V has shift amounts that are all less than the element bit-width of the shift n...
LLVM_ABI SDValue getSetFPEnv(SDValue Chain, const SDLoc &dl, SDValue Ptr, EVT MemVT, MachineMemOperand *MMO)
LLVM_ABI SDValue getBoolExtOrTrunc(SDValue Op, const SDLoc &SL, EVT VT, EVT OpVT)
Convert Op, which must be of integer type, to the integer type VT, by using an extension appropriate ...
LLVM_ABI SDValue getMaskedStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Base, SDValue Offset, SDValue Mask, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, bool IsTruncating=false, bool IsCompressing=false)
LLVM_ABI SDValue getExternalSymbol(const char *Sym, EVT VT)
const TargetMachine & getTarget() const
LLVM_ABI std::pair< SDValue, SDValue > getStrictFPExtendOrRound(SDValue Op, SDValue Chain, const SDLoc &DL, EVT VT)
Convert Op, which must be a STRICT operation of float type, to the float type VT, by either extending...
LLVM_ABI std::pair< SDValue, SDValue > SplitEVL(SDValue N, EVT VecVT, const SDLoc &DL)
Split the explicit vector length parameter of a VP operation.
LLVM_ABI SDValue getPtrExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either truncating it or perform...
LLVM_ABI SDValue getVPLogicalNOT(const SDLoc &DL, SDValue Val, SDValue Mask, SDValue EVL, EVT VT)
Create a vector-predicated logical NOT operation as (VP_XOR Val, BooleanOne, Mask,...
LLVM_ABI SDValue getMaskFromElementCount(const SDLoc &DL, EVT VT, ElementCount Len)
Return a vector with the first 'Len' lanes set to true and remaining lanes set to false.
LLVM_ABI SDValue getAnyExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either any-extending or truncat...
iterator_range< allnodes_iterator > allnodes()
LLVM_ABI SDValue getBlockAddress(const BlockAddress *BA, EVT VT, int64_t Offset=0, bool isTarget=false, unsigned TargetFlags=0)
LLVM_ABI SDValue WidenVector(const SDValue &N, const SDLoc &DL)
Widen the vector up to the next power of two using INSERT_SUBVECTOR.
LLVM_ABI bool isKnownNeverZeroFloat(SDValue Op) const
Test whether the given floating point SDValue is known to never be positive or negative zero.
const LibcallLoweringInfo & getLibcalls() const
LLVM_ABI SDValue getLoadVP(ISD::MemIndexedMode AM, ISD::LoadExtType ExtType, EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, SDValue Offset, SDValue Mask, SDValue EVL, MachinePointerInfo PtrInfo, EVT MemVT, Align Alignment, MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo, const MDNode *Ranges=nullptr, bool IsExpanding=false)
LLVM_ABI SDValue getIntPtrConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI SDDbgValue * getConstantDbgValue(DIVariable *Var, DIExpression *Expr, const Value *C, const DebugLoc &DL, unsigned O)
Creates a constant SDDbgValue node.
LLVM_ABI SDValue getScatterVP(SDVTList VTs, EVT VT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType)
LLVM_ABI SDValue getValueType(EVT)
LLVM_ABI SDValue getLifetimeNode(bool IsStart, const SDLoc &dl, SDValue Chain, int FrameIndex)
Creates a LifetimeSDNode that starts (IsStart==true) or ends (IsStart==false) the lifetime of the Fra...
ArrayRef< SDDbgValue * > GetDbgValues(const SDNode *SD) const
Get the debug values which reference the given SDNode.
LLVM_ABI SDValue getNode(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDUse > Ops)
Gets or creates the specified node.
LLVM_ABI OverflowKind computeOverflowForSignedAdd(SDValue N0, SDValue N1) const
Determine if the result of the signed addition of 2 nodes can overflow.
LLVM_ABI SDValue getFPExtendOrRound(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of float type, to the float type VT, by either extending or rounding (by tr...
LLVM_ABI unsigned AssignTopologicalOrder()
Topological-sort the AllNodes list and a assign a unique node id for each node in the DAG based on th...
ilist< SDNode >::size_type allnodes_size() const
LLVM_ABI bool isKnownNeverNaN(SDValue Op, const APInt &DemandedElts, bool SNaN=false, unsigned Depth=0) const
Test whether the given SDValue (or all elements of it, if it is a vector) is known to never be NaN in...
LLVM_ABI SDValue FoldConstantBuildVector(BuildVectorSDNode *BV, const SDLoc &DL, EVT DstEltVT)
Fold BUILD_VECTOR of constants/undefs to the destination type BUILD_VECTOR of constants/undefs elemen...
LLVM_ABI SDValue getAtomicMemmove(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Type *SizeTy, unsigned ElemSz, bool isTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo)
LLVM_ABI SDValue getIndexedMaskedStore(SDValue OrigStore, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue getTruncStoreVP(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, SDValue Mask, SDValue EVL, MachinePointerInfo PtrInfo, EVT SVT, Align Alignment, MachineMemOperand::Flags MMOFlags, const AAMDNodes &AAInfo, bool IsCompressing=false)
SDValue getTargetConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isOpaque=false)
LLVM_ABI unsigned ComputeNumSignBits(SDValue Op, unsigned Depth=0) const
Return the number of times the sign bit of the register is replicated into the other bits.
LLVM_ABI bool MaskedVectorIsZero(SDValue Op, const APInt &DemandedElts, unsigned Depth=0) const
Return true if 'Op' is known to be zero in DemandedElts.
LLVM_ABI SDValue getBoolConstant(bool V, const SDLoc &DL, EVT VT, EVT OpVT)
Create a true or false constant of type VT using the target's BooleanContent for type OpVT.
LLVM_ABI SDDbgValue * getFrameIndexDbgValue(DIVariable *Var, DIExpression *Expr, unsigned FI, bool IsIndirect, const DebugLoc &DL, unsigned O)
Creates a FrameIndex SDDbgValue node.
LLVM_ABI SDValue getExtStridedLoadVP(ISD::LoadExtType ExtType, const SDLoc &DL, EVT VT, SDValue Chain, SDValue Ptr, SDValue Stride, SDValue Mask, SDValue EVL, EVT MemVT, MachineMemOperand *MMO, bool IsExpanding=false)
LLVM_ABI SDValue getMemmove(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, const CallInst *CI, std::optional< bool > OverrideTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo=AAMDNodes(), BatchAAResults *BatchAA=nullptr)
LLVM_ABI SDValue getJumpTable(int JTI, EVT VT, bool isTarget=false, unsigned TargetFlags=0)
LLVM_ABI bool isBaseWithConstantOffset(SDValue Op) const
Return true if the specified operand is an ISD::ADD with a ConstantSDNode on the right-hand side,...
LLVM_ABI SDValue getVPPtrExtOrTrunc(const SDLoc &DL, EVT VT, SDValue Op, SDValue Mask, SDValue EVL)
Convert a vector-predicated Op, which must be of integer type, to the vector-type integer type VT,...
LLVM_ABI SDValue getVectorIdxConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI void getTopologicallyOrderedNodes(SmallVectorImpl< const SDNode * > &SortedNodes) const
Get all the nodes in their topological order without modifying any states.
LLVM_ABI void ReplaceAllUsesOfValueWith(SDValue From, SDValue To)
Replace any uses of From with To, leaving uses of other values produced by From.getNode() alone.
MachineFunction & getMachineFunction() const
LLVM_ABI std::pair< SDValue, SDValue > getStrstr(SDValue Chain, const SDLoc &dl, SDValue S0, SDValue S1, const CallInst *CI)
Lower a strstr operation into a target library call and return the resulting chain and call result as...
LLVM_ABI SDValue getPtrExtendInReg(SDValue Op, const SDLoc &DL, EVT VT)
Return the expression required to extend the Op as a pointer value assuming it was the smaller SrcTy ...
LLVM_ABI bool canCreateUndefOrPoison(SDValue Op, const APInt &DemandedElts, bool PoisonOnly=false, bool ConsiderFlags=true, unsigned Depth=0) const
Return true if Op can create undef or poison from non-undef & non-poison operands.
LLVM_ABI OverflowKind computeOverflowForUnsignedAdd(SDValue N0, SDValue N1) const
Determine if the result of the unsigned addition of 2 nodes can overflow.
SDValue getPOISON(EVT VT)
Return a POISON node. POISON does not have a useful SDLoc.
SDValue getSplatBuildVector(EVT VT, const SDLoc &DL, SDValue Op)
Return a splat ISD::BUILD_VECTOR node, consisting of Op splatted to all elements.
LLVM_ABI SDValue getFrameIndex(int FI, EVT VT, bool isTarget=false)
LLVM_ABI SDValue getTruncStridedStoreVP(SDValue Chain, const SDLoc &DL, SDValue Val, SDValue Ptr, SDValue Stride, SDValue Mask, SDValue EVL, EVT SVT, MachineMemOperand *MMO, bool IsCompressing=false)
LLVM_ABI void canonicalizeCommutativeBinop(unsigned Opcode, SDValue &N1, SDValue &N2) const
Swap N1 and N2 if Opcode is a commutative binary opcode and the canonical form expects the opposite o...
LLVM_ABI KnownBits computeKnownBits(SDValue Op, unsigned Depth=0) const
Determine which bits of Op are known to be either zero or one and return them in Known.
LLVM_ABI SDValue getRegisterMask(const uint32_t *RegMask)
LLVM_ABI SDValue getZExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either zero-extending or trunca...
LLVM_ABI SDValue getCondCode(ISD::CondCode Cond)
LLVM_ABI bool MaskedValueIsZero(SDValue Op, const APInt &Mask, unsigned Depth=0) const
Return true if 'Op & Mask' is known to be zero.
LLVM_ABI bool isKnownToBeAPowerOfTwoFP(SDValue Val, unsigned Depth=0) const
Test if the given fp value is known to be an integer power-of-2, either positive or negative.
LLVM_ABI OverflowKind computeOverflowForSignedSub(SDValue N0, SDValue N1) const
Determine if the result of the signed sub of 2 nodes can overflow.
SDValue getObjectPtrOffset(const SDLoc &SL, SDValue Ptr, TypeSize Offset)
Create an add instruction with appropriate flags when used for addressing some offset of an object.
LLVMContext * getContext() const
LLVM_ABI SDValue simplifyFPBinop(unsigned Opcode, SDValue X, SDValue Y, SDNodeFlags Flags)
Try to simplify a floating-point binary operation into 1 of its operands or a constant.
const SDValue & setRoot(SDValue N)
Set the current root tag of the SelectionDAG.
LLVM_ABI bool isKnownToBeAPowerOfTwo(SDValue Val, bool OrZero=false, unsigned Depth=0) const
Test if the given value is known to have exactly one bit set.
LLVM_ABI SDValue getDeactivationSymbol(const GlobalValue *GV)
LLVM_ABI SDValue getTargetExternalSymbol(const char *Sym, EVT VT, unsigned TargetFlags=0)
LLVM_ABI SDValue getMCSymbol(MCSymbol *Sym, EVT VT)
LLVM_ABI bool isUndef(unsigned Opcode, ArrayRef< SDValue > Ops)
Return true if the result of this operation is always undefined.
LLVM_ABI SDValue CreateStackTemporary(TypeSize Bytes, Align Alignment)
Create a stack temporary based on the size in bytes and the alignment.
LLVM_ABI SDNode * UpdateNodeOperands(SDNode *N, SDValue Op)
Mutate the specified node in-place to have the specified operands.
LLVM_ABI std::pair< EVT, EVT > GetDependentSplitDestVTs(const EVT &VT, const EVT &EnvVT, bool *HiIsEmpty) const
Compute the VTs needed for the low/hi parts of a type, dependent on an enveloping VT that has been sp...
LLVM_ABI SDValue foldConstantFPMath(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDValue > Ops)
Fold floating-point operations when all operands are constants and/or undefined.
LLVM_ABI std::optional< ConstantRange > getValidShiftAmountRange(SDValue V, const APInt &DemandedElts, unsigned Depth) const
If a SHL/SRA/SRL node V has shift amounts that are all less than the element bit-width of the shift n...
LLVM_ABI SDValue FoldSymbolOffset(unsigned Opcode, EVT VT, const GlobalAddressSDNode *GA, const SDNode *N2)
LLVM_ABI SDValue getIndexedLoad(SDValue OrigLoad, const SDLoc &dl, SDValue Base, SDValue Offset, ISD::MemIndexedMode AM)
LLVM_ABI SDValue getTargetInsertSubreg(int SRIdx, const SDLoc &DL, EVT VT, SDValue Operand, SDValue Subreg)
A convenience function for creating TargetInstrInfo::INSERT_SUBREG nodes.
SDValue getEntryNode() const
Return the token chain corresponding to the entry of the function.
LLVM_ABI SDDbgValue * getDbgValue(DIVariable *Var, DIExpression *Expr, SDNode *N, unsigned R, bool IsIndirect, const DebugLoc &DL, unsigned O)
Creates a SDDbgValue node.
LLVM_ABI SDValue getMaskedLoad(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Base, SDValue Offset, SDValue Mask, SDValue Src0, EVT MemVT, MachineMemOperand *MMO, ISD::MemIndexedMode AM, ISD::LoadExtType, bool IsExpanding=false)
SDValue getSplat(EVT VT, const SDLoc &DL, SDValue Op)
Returns a node representing a splat of one value into all lanes of the provided vector type.
LLVM_ABI std::pair< SDValue, SDValue > SplitScalar(const SDValue &N, const SDLoc &DL, const EVT &LoVT, const EVT &HiVT)
Split the scalar node with EXTRACT_ELEMENT using the provided VTs and return the low/high part.
LLVM_ABI SDValue matchBinOpReduction(SDNode *Extract, ISD::NodeType &BinOp, ArrayRef< ISD::NodeType > CandidateBinOps, bool AllowPartials=false)
Match a binop + shuffle pyramid that represents a horizontal reduction over the elements of a vector ...
LLVM_ABI bool isADDLike(SDValue Op, bool NoWrap=false) const
Return true if the specified operand is an ISD::OR or ISD::XOR node that can be treated as an ISD::AD...
LLVM_ABI SDValue getVectorShuffle(EVT VT, const SDLoc &dl, SDValue N1, SDValue N2, ArrayRef< int > Mask)
Return an ISD::VECTOR_SHUFFLE node.
LLVM_ABI SDValue simplifyShift(SDValue X, SDValue Y)
Try to simplify a shift into 1 of its operands or a constant.
LLVM_ABI void transferDbgValues(SDValue From, SDValue To, unsigned OffsetInBits=0, unsigned SizeInBits=0, bool InvalidateDbg=true)
Transfer debug values from one node to another, while optionally generating fragment expressions for ...
LLVM_ABI SDValue getLogicalNOT(const SDLoc &DL, SDValue Val, EVT VT)
Create a logical NOT operation as (XOR Val, BooleanOne).
LLVM_ABI SDValue getMaskedScatter(SDVTList VTs, EVT MemVT, const SDLoc &dl, ArrayRef< SDValue > Ops, MachineMemOperand *MMO, ISD::MemIndexType IndexType, bool IsTruncating=false)
ilist< SDNode >::iterator allnodes_iterator
This SDNode is used to implement the code generator support for the llvm IR shufflevector instruction...
int getMaskElt(unsigned Idx) const
ArrayRef< int > getMask() const
static void commuteMask(MutableArrayRef< int > Mask)
Change values in a shuffle permute mask assuming the two vector operands have swapped position.
static LLVM_ABI bool isSplatMask(ArrayRef< int > Mask)
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
bool erase(PtrType Ptr)
Remove pointer from the set.
size_type count(ConstPtrType Ptr) const
count - Return 1 if the specified pointer is in the set, 0 otherwise.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void assign(size_type NumElts, ValueParamT Elt)
reference emplace_back(ArgTypes &&... Args)
void reserve(size_type N)
iterator erase(const_iterator CI)
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
This class is used to represent ISD::STORE nodes.
StringRef - Represent a constant reference to a string, i.e.
Definition StringRef.h:55
constexpr const char * data() const
data - Get a pointer to the start of the string (which may not be null terminated).
Definition StringRef.h:137
Information about stack frame layout on the target.
virtual TargetStackID::Value getStackIDForScalableVectors() const
Returns the StackID that scalable vectors should be associated with.
Align getStackAlign() const
getStackAlignment - This method returns the number of bytes to which the stack pointer must be aligne...
Completely target-dependent object reference.
unsigned getTargetFlags() const
Provides information about what library functions are available for the current target.
virtual bool shouldConvertConstantLoadToIntImm(const APInt &Imm, Type *Ty) const
Return true if it is beneficial to convert a load of a constant to just the constant itself.
const TargetMachine & getTargetMachine() const
virtual bool isZExtFree(Type *FromTy, Type *ToTy) const
Return true if any actual instruction that defines a value of type FromTy implicitly zero-extends the...
unsigned getMaxStoresPerMemcpy(bool OptSize) const
Get maximum # of store operations permitted for llvm.memcpy.
unsigned getMaxStoresPerMemset(bool OptSize) const
Get maximum # of store operations permitted for llvm.memset.
virtual bool shallExtractConstSplatVectorElementToStore(Type *VectorTy, unsigned ElemSizeInBits, unsigned &Index) const
Return true if the target shall perform extract vector element and store given that the vector is kno...
virtual bool isTruncateFree(Type *FromTy, Type *ToTy) const
Return true if it's free to truncate a value of type FromTy to type ToTy.
virtual EVT getTypeToTransformTo(LLVMContext &Context, EVT VT) const
For types supported by the target, this is an identity function.
bool isTypeLegal(EVT VT) const
Return true if the target has native support for the specified value type.
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
BooleanContent
Enum that describes how the target represents true/false values.
virtual unsigned getMaxGluedStoresPerMemcpy() const
Get maximum # of store operations to be glued together.
std::vector< ArgListEntry > ArgListTy
unsigned getMaxStoresPerMemmove(bool OptSize) const
Get maximum # of store operations permitted for llvm.memmove.
virtual bool isLegalStoreImmediate(int64_t Value) const
Return true if the specified immediate is legal for the value input of a store instruction.
static ISD::NodeType getExtendForContent(BooleanContent Content)
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
virtual bool findOptimalMemOpLowering(LLVMContext &Context, std::vector< EVT > &MemOps, unsigned Limit, const MemOp &Op, unsigned DstAS, unsigned SrcAS, const AttributeList &FuncAttributes, EVT *LargestVT=nullptr) const
Determines the optimal series of memory ops to replace the memset / memcpy.
std::pair< SDValue, SDValue > LowerCallTo(CallLoweringInfo &CLI) const
This function lowers an abstract call to a function into an actual call.
Primary interface to the complete machine description for the target machine.
virtual bool isNoopAddrSpaceCast(unsigned SrcAS, unsigned DestAS) const
Returns true if a cast between SrcAS and DestAS is a noop.
const Triple & getTargetTriple() const
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
virtual const SelectionDAGTargetInfo * getSelectionDAGInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
virtual const TargetLowering * getTargetLowering() const
bool isOSDarwin() const
Is this a "Darwin" OS (macOS, iOS, tvOS, watchOS, DriverKit, XROS, or bridgeOS).
Definition Triple.h:639
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
Definition Twine.h:82
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition TypeSize.h:343
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
bool isVectorTy() const
True if this is an instance of VectorType.
Definition Type.h:273
static LLVM_ABI IntegerType * getInt32Ty(LLVMContext &C)
Definition Type.cpp:296
static LLVM_ABI Type * getVoidTy(LLVMContext &C)
Definition Type.cpp:280
static LLVM_ABI IntegerType * getInt8Ty(LLVMContext &C)
Definition Type.cpp:294
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
Definition Type.cpp:197
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
Definition Type.cpp:230
A Use represents the edge between a Value definition and its users.
Definition Use.h:35
LLVM_ABI unsigned getOperandNo() const
Return the operand # of this use in its User.
Definition Use.cpp:35
LLVM_ABI void set(Value *Val)
Definition Value.h:907
User * getUser() const
Returns the User that contains this Use.
Definition Use.h:61
Value * getOperand(unsigned i) const
Definition User.h:207
This class is used to represent an VP_GATHER node.
This class is used to represent a VP_LOAD node.
This class is used to represent an VP_SCATTER node.
This class is used to represent a VP_STORE node.
This class is used to represent an EXPERIMENTAL_VP_STRIDED_LOAD node.
This class is used to represent an EXPERIMENTAL_VP_STRIDED_STORE node.
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:256
std::pair< iterator, bool > insert(const ValueT &V)
Definition DenseSet.h:202
bool contains(const_arg_type_t< ValueT > V) const
Check if the set contains the given element.
Definition DenseSet.h:175
constexpr bool hasKnownScalarFactor(const FixedOrScalableQuantity &RHS) const
Returns true if there exists a value X where RHS.multiplyCoefficientBy(X) will result in a value whos...
Definition TypeSize.h:269
constexpr ScalarTy getFixedValue() const
Definition TypeSize.h:200
static constexpr bool isKnownLE(const FixedOrScalableQuantity &LHS, const FixedOrScalableQuantity &RHS)
Definition TypeSize.h:230
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition TypeSize.h:168
constexpr bool isKnownEven() const
A return value of true indicates we know at compile time that the number of elements (vscale * Min) i...
Definition TypeSize.h:176
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:165
constexpr LeafTy divideCoefficientBy(ScalarTy RHS) const
We do not provide the '/' operator here because division for polynomial types does not work in the sa...
Definition TypeSize.h:252
static constexpr bool isKnownGE(const FixedOrScalableQuantity &LHS, const FixedOrScalableQuantity &RHS)
Definition TypeSize.h:237
A raw_ostream that writes to an std::string.
CallInst * Call
Changed
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
LLVM_ABI APInt clmulr(const APInt &LHS, const APInt &RHS)
Perform a reversed carry-less multiply.
Definition APInt.cpp:3212
LLVM_ABI APInt mulhu(const APInt &C1, const APInt &C2)
Performs (2*N)-bit multiplication on zero-extended operands.
Definition APInt.cpp:3142
LLVM_ABI APInt avgCeilU(const APInt &C1, const APInt &C2)
Compute the ceil of the unsigned average of C1 and C2.
Definition APInt.cpp:3129
LLVM_ABI APInt avgFloorU(const APInt &C1, const APInt &C2)
Compute the floor of the unsigned average of C1 and C2.
Definition APInt.cpp:3119
LLVM_ABI APInt fshr(const APInt &Hi, const APInt &Lo, const APInt &Shift)
Perform a funnel shift right.
Definition APInt.cpp:3193
LLVM_ABI APInt mulhs(const APInt &C1, const APInt &C2)
Performs (2*N)-bit multiplication on sign-extended operands.
Definition APInt.cpp:3134
LLVM_ABI APInt clmul(const APInt &LHS, const APInt &RHS)
Perform a carry-less multiply, also known as XOR multiplication, and return low-bits.
Definition APInt.cpp:3202
APInt abds(const APInt &A, const APInt &B)
Determine the absolute difference of two APInts considered to be signed.
Definition APInt.h:2283
LLVM_ABI APInt fshl(const APInt &Hi, const APInt &Lo, const APInt &Shift)
Perform a funnel shift left.
Definition APInt.cpp:3184
LLVM_ABI APInt ScaleBitMask(const APInt &A, unsigned NewBitWidth, bool MatchAllBits=false)
Splat/Merge neighboring bits to widen/narrow the bitmask represented by.
Definition APInt.cpp:3020
LLVM_ABI APInt clmulh(const APInt &LHS, const APInt &RHS)
Perform a carry-less multiply, and return high-bits.
Definition APInt.cpp:3217
APInt abdu(const APInt &A, const APInt &B)
Determine the absolute difference of two APInts considered to be unsigned.
Definition APInt.h:2288
LLVM_ABI APInt avgFloorS(const APInt &C1, const APInt &C2)
Compute the floor of the signed average of C1 and C2.
Definition APInt.cpp:3114
LLVM_ABI APInt avgCeilS(const APInt &C1, const APInt &C2)
Compute the ceil of the signed average of C1 and C2.
Definition APInt.cpp:3124
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
LLVM_ABI CondCode getSetCCInverse(CondCode Operation, bool isIntegerLike)
Return the operation corresponding to !(X op Y), where 'op' is a valid SetCC operation.
ISD namespace - This namespace contains an enum which represents all of the SelectionDAG node types a...
Definition ISDOpcodes.h:24
LLVM_ABI CondCode getSetCCAndOperation(CondCode Op1, CondCode Op2, EVT Type)
Return the result of a logical AND between different comparisons of identical values: ((X op1 Y) & (X...
LLVM_ABI bool isConstantSplatVectorAllOnes(const SDNode *N, bool BuildVectorOnly=false)
Return true if the specified node is a BUILD_VECTOR or SPLAT_VECTOR where all of the elements are ~0 ...
bool isNON_EXTLoad(const SDNode *N)
Returns true if the specified node is a non-extending load.
NodeType
ISD::NodeType enum - This enum defines the target-independent operators for a SelectionDAG.
Definition ISDOpcodes.h:41
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
Definition ISDOpcodes.h:819
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
Definition ISDOpcodes.h:261
@ CTLZ_ZERO_UNDEF
Definition ISDOpcodes.h:788
@ TargetConstantPool
Definition ISDOpcodes.h:189
@ MDNODE_SDNODE
MDNODE_SDNODE - This is a node that holdes an MDNode*, which is used to reference metadata in the IR.
@ STRICT_FSETCC
STRICT_FSETCC/STRICT_FSETCCS - Constrained versions of SETCC, used for floating-point operands only.
Definition ISDOpcodes.h:511
@ PTRADD
PTRADD represents pointer arithmetic semantics, for targets that opt in using shouldPreservePtrArith(...
@ DELETED_NODE
DELETED_NODE - This is an illegal value that is used to catch errors.
Definition ISDOpcodes.h:45
@ POISON
POISON - A poison node.
Definition ISDOpcodes.h:236
@ PARTIAL_REDUCE_SMLA
PARTIAL_REDUCE_[U|S]MLA(Accumulator, Input1, Input2) The partial reduction nodes sign or zero extend ...
@ VECREDUCE_SEQ_FADD
Generic reduction nodes.
@ MLOAD
Masked load and store - consecutive vector load and store operations with additional mask operand tha...
@ FGETSIGN
INT = FGETSIGN(FP) - Return the sign bit of the specified floating point value as an integer 0/1 valu...
Definition ISDOpcodes.h:538
@ SMUL_LOHI
SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing a signed/unsigned value of type i[2...
Definition ISDOpcodes.h:275
@ INSERT_SUBVECTOR
INSERT_SUBVECTOR(VECTOR1, VECTOR2, IDX) - Returns a vector with VECTOR2 inserted into VECTOR1.
Definition ISDOpcodes.h:600
@ JUMP_TABLE_DEBUG_INFO
JUMP_TABLE_DEBUG_INFO - Jumptable debug info.
@ BSWAP
Byte Swap and Counting operators.
Definition ISDOpcodes.h:779
@ TargetBlockAddress
Definition ISDOpcodes.h:191
@ DEACTIVATION_SYMBOL
Untyped node storing deactivation symbol reference (DeactivationSymbolSDNode).
@ ATOMIC_STORE
OUTCHAIN = ATOMIC_STORE(INCHAIN, val, ptr) This corresponds to "store atomic" instruction.
@ ADDC
Carry-setting nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:294
@ FMAD
FMAD - Perform a * b + c, while getting the same result as the separately rounded operations.
Definition ISDOpcodes.h:522
@ ADD
Simple integer binary arithmetic operators.
Definition ISDOpcodes.h:264
@ LOAD
LOAD and STORE have token chains as their first operand, then the same operands as an LLVM load/store...
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
Definition ISDOpcodes.h:853
@ ATOMIC_LOAD_USUB_COND
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
Definition ISDOpcodes.h:518
@ FATAN2
FATAN2 - atan2, inspired by libm.
@ INTRINSIC_VOID
OUTCHAIN = INTRINSIC_VOID(INCHAIN, INTRINSICID, arg1, arg2, ...) This node represents a target intrin...
Definition ISDOpcodes.h:220
@ GlobalAddress
Definition ISDOpcodes.h:88
@ ATOMIC_CMP_SWAP_WITH_SUCCESS
Val, Success, OUTCHAIN = ATOMIC_CMP_SWAP_WITH_SUCCESS(INCHAIN, ptr, cmp, swap) N.b.
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
Definition ISDOpcodes.h:880
@ CONCAT_VECTORS
CONCAT_VECTORS(VECTOR0, VECTOR1, ...) - Given a number of values of vector type with the same length ...
Definition ISDOpcodes.h:584
@ VECREDUCE_FMAX
FMIN/FMAX nodes can have flags, for NaN/NoNaN variants.
@ FADD
Simple binary floating point operators.
Definition ISDOpcodes.h:417
@ VECREDUCE_FMAXIMUM
FMINIMUM/FMAXIMUM nodes propatate NaNs and signed zeroes using the llvm.minimum and llvm....
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
Definition ISDOpcodes.h:747
@ SIGN_EXTEND_VECTOR_INREG
SIGN_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register sign-extension of the low ...
Definition ISDOpcodes.h:910
@ FP16_TO_FP
FP16_TO_FP, FP_TO_FP16 - These operators are used to perform promotions and truncation for half-preci...
@ FMULADD
FMULADD - Performs a * b + c, with, or without, intermediate rounding.
Definition ISDOpcodes.h:528
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
Definition ISDOpcodes.h:993
@ BUILD_PAIR
BUILD_PAIR - This is the opposite of EXTRACT_ELEMENT in some ways.
Definition ISDOpcodes.h:254
@ CLMUL
Carry-less multiplication operations.
Definition ISDOpcodes.h:774
@ FLDEXP
FLDEXP - ldexp, inspired by libm (op0 * 2**op1).
@ BUILTIN_OP_END
BUILTIN_OP_END - This must be the last enum value in this list.
@ GlobalTLSAddress
Definition ISDOpcodes.h:89
@ SRCVALUE
SRCVALUE - This is a node type that holds a Value* that is used to make reference to a value in the L...
@ EH_LABEL
EH_LABEL - Represents a label in mid basic block used to track locations needed for debug and excepti...
@ ATOMIC_LOAD_USUB_SAT
@ PARTIAL_REDUCE_UMLA
@ SIGN_EXTEND
Conversion operators.
Definition ISDOpcodes.h:844
@ AVGCEILS
AVGCEILS/AVGCEILU - Rounding averaging add - Add two integers using an integer of type i[N+2],...
Definition ISDOpcodes.h:715
@ SCALAR_TO_VECTOR
SCALAR_TO_VECTOR(VAL) - This represents the operation of loading a scalar value into element 0 of the...
Definition ISDOpcodes.h:665
@ TargetExternalSymbol
Definition ISDOpcodes.h:190
@ VECREDUCE_FADD
These reductions have relaxed evaluation order semantics, and have a single vector operand.
@ CTTZ_ZERO_UNDEF
Bit counting operators with an undefined result for zero inputs.
Definition ISDOpcodes.h:787
@ TargetJumpTable
Definition ISDOpcodes.h:188
@ TargetIndex
TargetIndex - Like a constant pool entry, but with completely target-dependent semantics.
Definition ISDOpcodes.h:198
@ PARTIAL_REDUCE_FMLA
@ PREFETCH
PREFETCH - This corresponds to a prefetch intrinsic.
@ TRUNCATE_SSAT_U
Definition ISDOpcodes.h:873
@ SETCCCARRY
Like SetCC, ops #0 and #1 are the LHS and RHS operands to compare, but op #2 is a boolean indicating ...
Definition ISDOpcodes.h:827
@ FNEG
Perform various unary floating-point operations inspired by libm.
@ BR_CC
BR_CC - Conditional branch.
@ SSUBO
Same for subtraction.
Definition ISDOpcodes.h:352
@ STEP_VECTOR
STEP_VECTOR(IMM) - Returns a scalable vector whose lanes are comprised of a linear sequence of unsign...
Definition ISDOpcodes.h:691
@ FCANONICALIZE
Returns platform specific canonical encoding of a floating point number.
Definition ISDOpcodes.h:541
@ SSUBSAT
RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2 integers with the same bit width ...
Definition ISDOpcodes.h:374
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
Definition ISDOpcodes.h:796
@ ATOMIC_LOAD
Val, OUTCHAIN = ATOMIC_LOAD(INCHAIN, ptr) This corresponds to "load atomic" instruction.
@ UNDEF
UNDEF - An undefined node.
Definition ISDOpcodes.h:233
@ EXTRACT_ELEMENT
EXTRACT_ELEMENT - This is used to get the lower or upper (determined by a Constant,...
Definition ISDOpcodes.h:247
@ SPLAT_VECTOR
SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL duplicated in all lanes.
Definition ISDOpcodes.h:672
@ AssertAlign
AssertAlign - These nodes record if a register contains a value that has a known alignment and the tr...
Definition ISDOpcodes.h:69
@ GET_ACTIVE_LANE_MASK
GET_ACTIVE_LANE_MASK - this corrosponds to the llvm.get.active.lane.mask intrinsic.
@ BasicBlock
Various leaf nodes.
Definition ISDOpcodes.h:81
@ CopyFromReg
CopyFromReg - This node indicates that the input value is a virtual or physical register that is defi...
Definition ISDOpcodes.h:230
@ SADDO
RESULT, BOOL = [SU]ADDO(LHS, RHS) - Overflow-aware nodes for addition.
Definition ISDOpcodes.h:348
@ TargetGlobalAddress
TargetGlobalAddress - Like GlobalAddress, but the DAG does no folding or anything else with this node...
Definition ISDOpcodes.h:185
@ ARITH_FENCE
ARITH_FENCE - This corresponds to a arithmetic fence intrinsic.
@ CTLS
Count leading redundant sign bits.
Definition ISDOpcodes.h:792
@ VECREDUCE_ADD
Integer reductions may have a result type larger than the vector element type.
@ MULHU
MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing an unsigned/signed value of...
Definition ISDOpcodes.h:704
@ ATOMIC_LOAD_FMAXIMUM
@ SHL
Shift and rotation operations.
Definition ISDOpcodes.h:765
@ AssertNoFPClass
AssertNoFPClass - These nodes record if a register contains a float value that is known to be not som...
Definition ISDOpcodes.h:78
@ VECTOR_SHUFFLE
VECTOR_SHUFFLE(VEC1, VEC2) - Returns a vector, of the same type as VEC1/VEC2.
Definition ISDOpcodes.h:649
@ EXTRACT_SUBVECTOR
EXTRACT_SUBVECTOR(VECTOR, IDX) - Returns a subvector from VECTOR.
Definition ISDOpcodes.h:614
@ FMINNUM_IEEE
FMINNUM_IEEE/FMAXNUM_IEEE - Perform floating-point minimumNumber or maximumNumber on two values,...
@ EntryToken
EntryToken - This is the marker used to indicate the start of a region.
Definition ISDOpcodes.h:48
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
Definition ISDOpcodes.h:576
@ CopyToReg
CopyToReg - This node has three operands: a chain, a register number to set to this value,...
Definition ISDOpcodes.h:224
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
Definition ISDOpcodes.h:850
@ TargetConstantFP
Definition ISDOpcodes.h:180
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
Definition ISDOpcodes.h:811
@ VSCALE
VSCALE(IMM) - Returns the runtime scaling factor used to calculate the number of elements within a sc...
@ ATOMIC_CMP_SWAP
Val, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmp, swap) For double-word atomic operations: ValLo,...
@ FMINNUM
FMINNUM/FMAXNUM - Perform floating-point minimum maximum on two values, following IEEE-754 definition...
@ SSHLSAT
RESULT = [US]SHLSAT(LHS, RHS) - Perform saturation left shift.
Definition ISDOpcodes.h:386
@ SMULO
Same for multiplication.
Definition ISDOpcodes.h:356
@ ATOMIC_LOAD_FMINIMUM
@ TargetFrameIndex
Definition ISDOpcodes.h:187
@ VECTOR_SPLICE_LEFT
VECTOR_SPLICE_LEFT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1, VEC2) left by OFFSET elements an...
Definition ISDOpcodes.h:653
@ ANY_EXTEND_VECTOR_INREG
ANY_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register any-extension of the low la...
Definition ISDOpcodes.h:899
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
Definition ISDOpcodes.h:888
@ SMIN
[US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned integers.
Definition ISDOpcodes.h:727
@ LIFETIME_START
This corresponds to the llvm.lifetime.
@ FP_EXTEND
X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition ISDOpcodes.h:978
@ VSELECT
Select with a vector condition (op #0) and two vector operands (ops #1 and #2), returning a vector re...
Definition ISDOpcodes.h:805
@ UADDO_CARRY
Carry-using nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:328
@ MGATHER
Masked gather and scatter - load and store operations for a vector of random addresses with additiona...
@ HANDLENODE
HANDLENODE node - Used as a handle for various purposes.
@ BF16_TO_FP
BF16_TO_FP, FP_TO_BF16 - These operators are used to perform promotions and truncation for bfloat16.
@ ATOMIC_LOAD_UDEC_WRAP
@ STRICT_FP_ROUND
X = STRICT_FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision ...
Definition ISDOpcodes.h:500
@ FMINIMUM
FMINIMUM/FMAXIMUM - NaN-propagating minimum/maximum that also treat -0.0 as less than 0....
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
Definition ISDOpcodes.h:926
@ TargetConstant
TargetConstant* - Like Constant*, but the DAG does not do any folding, simplification,...
Definition ISDOpcodes.h:179
@ STRICT_FP_EXTEND
X = STRICT_FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition ISDOpcodes.h:505
@ AND
Bitwise operators - logical and, logical or, logical xor.
Definition ISDOpcodes.h:739
@ INTRINSIC_WO_CHAIN
RESULT = INTRINSIC_WO_CHAIN(INTRINSICID, arg1, arg2, ...) This node represents a target intrinsic fun...
Definition ISDOpcodes.h:205
@ GET_FPENV_MEM
Gets the current floating-point environment.
@ PSEUDO_PROBE
Pseudo probe for AutoFDO, as a place holder in a basic block to improve the sample counts quality.
@ SCMP
[US]CMP - 3-way comparison of signed or unsigned integers.
Definition ISDOpcodes.h:735
@ AVGFLOORS
AVGFLOORS/AVGFLOORU - Averaging add - Add two integers using an integer of type i[N+1],...
Definition ISDOpcodes.h:710
@ VECTOR_SPLICE_RIGHT
VECTOR_SPLICE_RIGHT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1,VEC2) right by OFFSET elements a...
Definition ISDOpcodes.h:657
@ ADDE
Carry-using nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:304
@ SPLAT_VECTOR_PARTS
SPLAT_VECTOR_PARTS(SCALAR1, SCALAR2, ...) - Returns a vector with the scalar values joined together a...
Definition ISDOpcodes.h:681
@ FREEZE
FREEZE - FREEZE(VAL) returns an arbitrary value if VAL is UNDEF (or is evaluated to UNDEF),...
Definition ISDOpcodes.h:241
@ INSERT_VECTOR_ELT
INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element at IDX replaced with VAL.
Definition ISDOpcodes.h:565
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
Definition ISDOpcodes.h:53
@ ATOMIC_SWAP
Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt) Val, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN,...
@ ExternalSymbol
Definition ISDOpcodes.h:93
@ FFREXP
FFREXP - frexp, extract fractional and exponent component of a floating-point value.
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
Definition ISDOpcodes.h:959
@ VECTOR_COMPRESS
VECTOR_COMPRESS(Vec, Mask, Passthru) consecutively place vector elements based on mask e....
Definition ISDOpcodes.h:699
@ ZERO_EXTEND_VECTOR_INREG
ZERO_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register zero-extension of the low ...
Definition ISDOpcodes.h:921
@ ADDRSPACECAST
ADDRSPACECAST - This operator converts between pointers of different address spaces.
Definition ISDOpcodes.h:997
@ EXPERIMENTAL_VECTOR_HISTOGRAM
Experimental vector histogram intrinsic Operands: Input Chain, Inc, Mask, Base, Index,...
@ FP_TO_SINT_SAT
FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a signed or unsigned scalar integer ...
Definition ISDOpcodes.h:945
@ VECREDUCE_FMINIMUM
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
Definition ISDOpcodes.h:856
@ VAARG
VAARG - VAARG has four operands: an input chain, a pointer, a SRCVALUE, and the alignment.
@ VECREDUCE_SEQ_FMUL
@ SHL_PARTS
SHL_PARTS/SRA_PARTS/SRL_PARTS - These operators are used for expanded integer shift operations.
Definition ISDOpcodes.h:833
@ AssertSext
AssertSext, AssertZext - These nodes record if a register contains a value that has already been zero...
Definition ISDOpcodes.h:62
@ ATOMIC_LOAD_UINC_WRAP
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
Definition ISDOpcodes.h:534
@ PARTIAL_REDUCE_SUMLA
@ SADDSAT
RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2 integers with the same bit width (W)...
Definition ISDOpcodes.h:365
@ SET_FPENV_MEM
Sets the current floating point environment.
@ FMINIMUMNUM
FMINIMUMNUM/FMAXIMUMNUM - minimumnum/maximumnum that is same with FMINNUM_IEEE and FMAXNUM_IEEE besid...
@ TRUNCATE_SSAT_S
TRUNCATE_[SU]SAT_[SU] - Truncate for saturated operand [SU] located in middle, prefix for SAT means i...
Definition ISDOpcodes.h:871
@ ABDS
ABDS/ABDU - Absolute difference - Return the absolute difference between two numbers interpreted as s...
Definition ISDOpcodes.h:722
@ TRUNCATE_USAT_U
Definition ISDOpcodes.h:875
@ SADDO_CARRY
Carry-using overflow-aware nodes for multiple precision addition and subtraction.
Definition ISDOpcodes.h:338
@ INTRINSIC_W_CHAIN
RESULT,OUTCHAIN = INTRINSIC_W_CHAIN(INCHAIN, INTRINSICID, arg1, ...) This node represents a target in...
Definition ISDOpcodes.h:213
@ TargetGlobalTLSAddress
Definition ISDOpcodes.h:186
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
Definition ISDOpcodes.h:556
LLVM_ABI NodeType getOppositeSignednessMinMaxOpcode(unsigned MinMaxOpc)
Given a MinMaxOpc of ISD::(U|S)MIN or ISD::(U|S)MAX, returns the corresponding opcode with the opposi...
LLVM_ABI bool isBuildVectorOfConstantSDNodes(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR node of all ConstantSDNode or undef.
LLVM_ABI NodeType getExtForLoadExtType(bool IsFP, LoadExtType)
bool isZEXTLoad(const SDNode *N)
Returns true if the specified node is a ZEXTLOAD.
bool matchUnaryFpPredicate(SDValue Op, std::function< bool(ConstantFPSDNode *)> Match, bool AllowUndefs=false)
Hook for matching ConstantFPSDNode predicate.
bool isExtOpcode(unsigned Opcode)
LLVM_ABI bool isConstantSplatVectorAllZeros(const SDNode *N, bool BuildVectorOnly=false)
Return true if the specified node is a BUILD_VECTOR or SPLAT_VECTOR where all of the elements are 0 o...
LLVM_ABI bool isVectorShrinkable(const SDNode *N, unsigned NewEltSize, bool Signed)
Returns true if the specified node is a vector where all elements can be truncated to the specified e...
LLVM_ABI bool isVPBinaryOp(unsigned Opcode)
Whether this is a vector-predicated binary operation opcode.
LLVM_ABI CondCode getSetCCInverse(CondCode Operation, EVT Type)
Return the operation corresponding to !(X op Y), where 'op' is a valid SetCC operation.
LLVM_ABI std::optional< unsigned > getBaseOpcodeForVP(unsigned Opcode, bool hasFPExcept)
Translate this VP Opcode to its corresponding non-VP Opcode.
bool isTrueWhenEqual(CondCode Cond)
Return true if the specified condition returns true if the two operands to the condition are equal.
LLVM_ABI std::optional< unsigned > getVPMaskIdx(unsigned Opcode)
The operand position of the vector mask.
unsigned getUnorderedFlavor(CondCode Cond)
This function returns 0 if the condition is always false if an operand is a NaN, 1 if the condition i...
LLVM_ABI std::optional< unsigned > getVPExplicitVectorLengthIdx(unsigned Opcode)
The operand position of the explicit vector length parameter.
bool isEXTLoad(const SDNode *N)
Returns true if the specified node is a EXTLOAD.
LLVM_ABI bool allOperandsUndef(const SDNode *N)
Return true if the node has at least one operand and all operands of the specified node are ISD::UNDE...
LLVM_ABI bool isFreezeUndef(const SDNode *N)
Return true if the specified node is FREEZE(UNDEF).
LLVM_ABI CondCode getSetCCSwappedOperands(CondCode Operation)
Return the operation corresponding to (Y op X) when given the operation for (X op Y).
LLVM_ABI std::optional< unsigned > getVPForBaseOpcode(unsigned Opcode)
Translate this non-VP Opcode to its corresponding VP Opcode.
MemIndexType
MemIndexType enum - This enum defines how to interpret MGATHER/SCATTER's index parameter when calcula...
LLVM_ABI bool isBuildVectorAllZeros(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR where all of the elements are 0 or undef.
bool matchUnaryPredicateImpl(SDValue Op, std::function< bool(ConstNodeType *)> Match, bool AllowUndefs=false, bool AllowTruncation=false)
Attempt to match a unary predicate against a scalar/splat constant or every element of a constant BUI...
LLVM_ABI bool isConstantSplatVector(const SDNode *N, APInt &SplatValue)
Node predicates.
LLVM_ABI NodeType getInverseMinMaxOpcode(unsigned MinMaxOpc)
Given a MinMaxOpc of ISD::(U|S)MIN or ISD::(U|S)MAX, returns ISD::(U|S)MAX and ISD::(U|S)MIN,...
LLVM_ABI bool matchBinaryPredicate(SDValue LHS, SDValue RHS, std::function< bool(ConstantSDNode *, ConstantSDNode *)> Match, bool AllowUndefs=false, bool AllowTypeMismatch=false)
Attempt to match a binary predicate against a pair of scalar/splat constants or every element of a pa...
LLVM_ABI bool isVPReduction(unsigned Opcode)
Whether this is a vector-predicated reduction opcode.
bool matchUnaryPredicate(SDValue Op, std::function< bool(ConstantSDNode *)> Match, bool AllowUndefs=false, bool AllowTruncation=false)
Hook for matching ConstantSDNode predicate.
MemIndexedMode
MemIndexedMode enum - This enum defines the load / store indexed addressing modes.
LLVM_ABI bool isBuildVectorOfConstantFPSDNodes(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR node of all ConstantFPSDNode or undef.
bool isSEXTLoad(const SDNode *N)
Returns true if the specified node is a SEXTLOAD.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
LLVM_ABI bool isBuildVectorAllOnes(const SDNode *N)
Return true if the specified node is a BUILD_VECTOR where all of the elements are ~0 or undef.
LLVM_ABI NodeType getVecReduceBaseOpcode(unsigned VecReduceOpcode)
Get underlying scalar opcode for VECREDUCE opcode.
LoadExtType
LoadExtType enum - This enum defines the three variants of LOADEXT (load with extension).
LLVM_ABI bool isVPOpcode(unsigned Opcode)
Whether this is a vector-predicated Opcode.
LLVM_ABI CondCode getSetCCOrOperation(CondCode Op1, CondCode Op2, EVT Type)
Return the result of a logical OR between different comparisons of identical values: ((X op1 Y) | (X ...
BinaryOp_match< SpecificConstantMatch, SrcTy, TargetOpcode::G_SUB > m_Neg(const SrcTy &&Src)
Matches a register negated by a G_SUB.
BinaryOp_match< LHS, RHS, Instruction::And > m_And(const LHS &L, const RHS &R)
deferredval_ty< Value > m_Deferred(Value *const &V)
Like m_Specific(), but works if the specific value to match is determined as part of the same match()...
class_match< Value > m_Value()
Match an arbitrary value and ignore it.
LLVM_ABI Libcall getMEMCPY_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMCPY_ELEMENT_UNORDERED_ATOMIC - Return MEMCPY_ELEMENT_UNORDERED_ATOMIC_* value for the given ele...
LLVM_ABI Libcall getMEMSET_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMSET_ELEMENT_UNORDERED_ATOMIC - Return MEMSET_ELEMENT_UNORDERED_ATOMIC_* value for the given ele...
LLVM_ABI Libcall getMEMMOVE_ELEMENT_UNORDERED_ATOMIC(uint64_t ElementSize)
getMEMMOVE_ELEMENT_UNORDERED_ATOMIC - Return MEMMOVE_ELEMENT_UNORDERED_ATOMIC_* value for the given e...
bool sd_match(SDNode *N, const SelectionDAG *DAG, Pattern &&P)
initializer< Ty > init(const Ty &Val)
@ DW_OP_LLVM_arg
Only used in LLVM metadata.
Definition Dwarf.h:149
std::enable_if_t< detail::IsValidPointer< X, Y >::value, X * > extract(Y &&MD)
Extract a Value from Metadata.
Definition Metadata.h:668
NodeAddr< NodeBase * > Node
Definition RDFGraph.h:381
This is an optimization pass for GlobalISel generic memory operations.
Definition Types.h:26
GenericUniformityInfo< SSAContext > UniformityInfo
unsigned Log2_32_Ceil(uint32_t Value)
Return the ceil log base 2 of the specified value, 32 if the value is zero.
Definition MathExtras.h:344
@ Offset
Definition DWP.cpp:532
bool operator<(int64_t V1, const APSInt &V2)
Definition APSInt.h:360
ISD::CondCode getICmpCondCode(ICmpInst::Predicate Pred)
getICmpCondCode - Return the ISD condition code corresponding to the given LLVM IR integer condition ...
Definition Analysis.cpp:237
void fill(R &&Range, T &&Value)
Provide wrappers to std::fill which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1759
LLVM_ABI SDValue peekThroughExtractSubvectors(SDValue V)
Return the non-extracted vector source operand of V if it exists.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1739
MaybeAlign getAlign(const CallInst &I, unsigned Index)
LLVM_ABI bool isNullConstant(SDValue V)
Returns true if V is a constant integer zero.
LLVM_ABI bool isAllOnesOrAllOnesSplat(const MachineInstr &MI, const MachineRegisterInfo &MRI, bool AllowUndefs=false)
Return true if the value is a constant -1 integer or a splatted vector of a constant -1 integer (with...
Definition Utils.cpp:1612
LLVM_ABI SDValue getBitwiseNotOperand(SDValue V, SDValue Mask, bool AllowUndefs)
If V is a bitwise not, returns the inverted operand.
@ Undef
Value of the register doesn't matter.
LLVM_ABI SDValue peekThroughBitcasts(SDValue V)
Return the non-bitcasted source operand of V if it exists.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
Definition STLExtras.h:2554
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
int countr_one(T Value)
Count the number of ones from the least significant bit to the first zero bit.
Definition bit.h:293
bool isIntOrFPConstant(SDValue V)
Return true if V is either a integer or FP constant.
auto dyn_cast_if_present(const Y &Val)
dyn_cast_if_present<X> - Functionally identical to dyn_cast, except that a null (or none in the case ...
Definition Casting.h:732
LLVM_ABI bool getConstantDataArrayInfo(const Value *V, ConstantDataArraySlice &Slice, unsigned ElementSize, uint64_t Offset=0)
Returns true if the value V is a pointer into a ConstantDataArray.
LLVM_ABI bool isOneOrOneSplatFP(SDValue V, bool AllowUndefs=false)
Return true if the value is a constant floating-point value, or a splatted vector of a constant float...
int bit_width(T Value)
Returns the number of bits needed to represent Value if Value is nonzero.
Definition bit.h:303
LLVM_READONLY APFloat maximum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 maximum semantics.
Definition APFloat.h:1706
void append_range(Container &C, Range &&R)
Wrapper function to append range R to container C.
Definition STLExtras.h:2208
constexpr bool isUIntN(unsigned N, uint64_t x)
Checks if an unsigned integer fits into the given (dynamic) bit width.
Definition MathExtras.h:243
LLVM_ABI bool shouldOptimizeForSize(const MachineFunction *MF, ProfileSummaryInfo *PSI, const MachineBlockFrequencyInfo *BFI, PGSOQueryType QueryType=PGSOQueryType::Other)
Returns true if machine function MF is suggested to be size-optimized based on the profile.
iterator_range< early_inc_iterator_impl< detail::IterOfRange< RangeT > > > make_early_inc_range(RangeT &&Range)
Make a range that does early increment to allow mutation of the underlying range without disrupting i...
Definition STLExtras.h:634
auto cast_or_null(const Y &Val)
Definition Casting.h:714
LLVM_ABI bool isNullOrNullSplat(const MachineInstr &MI, const MachineRegisterInfo &MRI, bool AllowUndefs=false)
Return true if the value is a constant 0 integer or a splatted vector of a constant 0 integer (with n...
Definition Utils.cpp:1594
LLVM_ABI bool isMinSignedConstant(SDValue V)
Returns true if V is a constant min signed integer value.
LLVM_ABI ConstantFPSDNode * isConstOrConstSplatFP(SDValue N, bool AllowUndefs=false)
Returns the SDNode if it is a constant splat BuildVector or constant float.
LLVM_ABI ConstantRange getConstantRangeFromMetadata(const MDNode &RangeMD)
Parse out a conservative ConstantRange from !range metadata.
APFloat frexp(const APFloat &X, int &Exp, APFloat::roundingMode RM)
Equivalent of C standard library function.
Definition APFloat.h:1618
int countr_zero(T Val)
Count number of 0's from the least significant bit to the most stopping at the first 1.
Definition bit.h:202
auto dyn_cast_or_null(const Y &Val)
Definition Casting.h:753
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1746
LLVM_ABI bool getShuffleDemandedElts(int SrcWidth, ArrayRef< int > Mask, const APInt &DemandedElts, APInt &DemandedLHS, APInt &DemandedRHS, bool AllowUndefElts=false)
Transform a shuffle mask's output demanded element mask into demanded element masks for the 2 operand...
LLVM_READONLY APFloat maxnum(const APFloat &A, const APFloat &B)
Implements IEEE-754 2008 maxNum semantics.
Definition APFloat.h:1661
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition MathExtras.h:331
LLVM_ABI bool isBitwiseNot(SDValue V, bool AllowUndefs=false)
Returns true if V is a bitwise not operation.
LLVM_ABI SDValue peekThroughInsertVectorElt(SDValue V, const APInt &DemandedElts)
Recursively peek through INSERT_VECTOR_ELT nodes, returning the source vector operand of V,...
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition MathExtras.h:279
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
LLVM_ABI void checkForCycles(const SelectionDAG *DAG, bool force=false)
void sort(IteratorTy Start, IteratorTy End)
Definition STLExtras.h:1636
LLVM_READONLY APFloat minimumnum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 minimumNumber semantics.
Definition APFloat.h:1692
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
LLVM_ABI void computeKnownBits(const Value *V, KnownBits &Known, const DataLayout &DL, AssumptionCache *AC=nullptr, const Instruction *CxtI=nullptr, const DominatorTree *DT=nullptr, bool UseInstrInfo=true, unsigned Depth=0)
Determine which bits of V are known to be either zero or one and return them in the KnownZero/KnownOn...
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:207
LLVM_ABI SDValue peekThroughTruncates(SDValue V)
Return the non-truncated source operand of V if it exists.
bool none_of(R &&Range, UnaryPredicate P)
Provide wrappers to std::none_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1753
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
Definition Error.cpp:163
constexpr std::underlying_type_t< Enum > to_underlying(Enum E)
Returns underlying integer value of an enum.
FunctionAddr VTableAddr Count
Definition InstrProf.h:139
LLVM_ABI ConstantRange getVScaleRange(const Function *F, unsigned BitWidth)
Determine the possible constant range of vscale with the given bit width, based on the vscale_range f...
LLVM_ABI SDValue peekThroughOneUseBitcasts(SDValue V)
Return the non-bitcasted and one-use source operand of V if it exists.
CodeGenOptLevel
Code generation optimization level.
Definition CodeGen.h:82
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
LLVM_ABI bool isOneOrOneSplat(SDValue V, bool AllowUndefs=false)
Return true if the value is a constant 1 integer or a splatted vector of a constant 1 integer (with n...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
@ Other
Any other memory.
Definition ModRef.h:68
LLVM_READONLY APFloat minnum(const APFloat &A, const APFloat &B)
Implements IEEE-754 2008 minNum semantics.
Definition APFloat.h:1642
@ Mul
Product of integers.
@ Sub
Subtraction of integers.
LLVM_ABI bool isNullConstantOrUndef(SDValue V)
Returns true if V is a constant integer zero or an UNDEF node.
bool isInTailCallPosition(const CallBase &Call, const TargetMachine &TM, bool ReturnsFirstArg=false)
Test if the given instruction is in a position to be optimized with a tail-call.
Definition Analysis.cpp:539
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
LLVM_ABI ConstantSDNode * isConstOrConstSplat(SDValue N, bool AllowUndefs=false, bool AllowTruncation=false)
Returns the SDNode if it is a constant splat BuildVector or constant int.
OutputIt copy(R &&Range, OutputIt Out)
Definition STLExtras.h:1885
constexpr unsigned BitWidth
bool funcReturnsFirstArgOfCall(const CallInst &CI)
Returns true if the parent of CI returns CI's first argument after calling CI.
Definition Analysis.cpp:719
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
LLVM_ABI bool isZeroOrZeroSplat(SDValue N, bool AllowUndefs=false)
Return true if the value is a constant 0 integer or a splatted vector of a constant 0 integer (with n...
LLVM_ABI bool isOneConstant(SDValue V)
Returns true if V is a constant integer one.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
Definition STLExtras.h:1947
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
Definition Alignment.h:201
LLVM_ABI bool isNullFPConstant(SDValue V)
Returns true if V is an FP constant with a value of positive zero.
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
Definition MathExtras.h:572
unsigned Log2(Align A)
Returns the log2 of the alignment.
Definition Alignment.h:197
LLVM_ABI bool isZeroOrZeroSplatFP(SDValue N, bool AllowUndefs=false)
Return true if the value is a constant (+/-)0.0 floating-point value or a splatted vector thereof (wi...
LLVM_ABI void computeKnownBitsFromRangeMetadata(const MDNode &Ranges, KnownBits &Known)
Compute known bits from the range metadata.
LLVM_READONLY APFloat minimum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 minimum semantics.
Definition APFloat.h:1679
LLVM_READONLY APFloat maximumnum(const APFloat &A, const APFloat &B)
Implements IEEE 754-2019 maximumNumber semantics.
Definition APFloat.h:1719
LLVM_ABI bool isOnesOrOnesSplat(SDValue N, bool AllowUndefs=false)
Return true if the value is a constant 1 integer or a splatted vector of a constant 1 integer (with n...
LLVM_ABI bool isNeutralConstant(unsigned Opc, SDNodeFlags Flags, SDValue V, unsigned OperandNo)
Returns true if V is a neutral element of Opc with Flags.
LLVM_ABI bool isAllOnesConstant(SDValue V)
Returns true if V is an integer constant with all bits set.
constexpr uint64_t NextPowerOf2(uint64_t A)
Returns the next power of two (in 64-bits) that is strictly greater than A.
Definition MathExtras.h:373
LLVM_ABI void reportFatalUsageError(Error Err)
Report a fatal error that does not indicate a bug in LLVM.
Definition Error.cpp:177
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
#define N
A collection of metadata nodes that might be associated with a memory access used by the alias-analys...
Definition Metadata.h:763
MDNode * TBAAStruct
The tag for type-based alias analysis (tbaa struct).
Definition Metadata.h:783
MDNode * TBAA
The tag for type-based alias analysis.
Definition Metadata.h:780
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
Definition Alignment.h:77
Represents offset+length into a ConstantDataArray.
uint64_t Length
Length of the slice.
uint64_t Offset
Slice starts at this Offset.
void move(uint64_t Delta)
Moves the Offset and adjusts Length accordingly.
const ConstantDataArray * Array
ConstantDataArray pointer.
Extended Value Type.
Definition ValueTypes.h:35
TypeSize getStoreSize() const
Return the number of bytes overwritten by a store of the specified value type.
Definition ValueTypes.h:395
bool isSimple() const
Test if the given EVT is simple (as opposed to being extended).
Definition ValueTypes.h:137
intptr_t getRawBits() const
Definition ValueTypes.h:512
static EVT getVectorVT(LLVMContext &Context, EVT VT, unsigned NumElements, bool IsScalable=false)
Returns the EVT that represents a vector NumElements in length, where each element is of type VT.
Definition ValueTypes.h:74
EVT changeTypeToInteger() const
Return the type converted to an equivalently sized integer or vector with integer element type.
Definition ValueTypes.h:121
bool bitsGT(EVT VT) const
Return true if this has more bits than VT.
Definition ValueTypes.h:284
bool bitsLT(EVT VT) const
Return true if this has less bits than VT.
Definition ValueTypes.h:300
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
Definition ValueTypes.h:147
ElementCount getVectorElementCount() const
Definition ValueTypes.h:350
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
Definition ValueTypes.h:373
unsigned getVectorMinNumElements() const
Given a vector type, return the minimum number of elements it contains.
Definition ValueTypes.h:359
uint64_t getScalarSizeInBits() const
Definition ValueTypes.h:385
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
Definition ValueTypes.h:316
static EVT getIntegerVT(LLVMContext &Context, unsigned BitWidth)
Returns the EVT that represents an integer with the given number of bits.
Definition ValueTypes.h:65
bool isFixedLengthVector() const
Definition ValueTypes.h:181
bool isVector() const
Return true if this is a vector value type.
Definition ValueTypes.h:168
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
Definition ValueTypes.h:323
bool bitsGE(EVT VT) const
Return true if this has no less bits than VT.
Definition ValueTypes.h:292
bool bitsEq(EVT VT) const
Return true if this has the same number of bits as VT.
Definition ValueTypes.h:256
LLVM_ABI Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
bool isScalableVector() const
Return true if this is a vector type where the runtime length is machine dependent.
Definition ValueTypes.h:174
EVT getVectorElementType() const
Given a vector type, return the type of each element.
Definition ValueTypes.h:328
bool isExtended() const
Test if the given EVT is extended (as opposed to being simple).
Definition ValueTypes.h:142
LLVM_ABI const fltSemantics & getFltSemantics() const
Returns an APFloat semantics tag appropriate for the value type.
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
Definition ValueTypes.h:336
bool bitsLE(EVT VT) const
Return true if this has no more bits than VT.
Definition ValueTypes.h:308
EVT getHalfNumVectorElementsVT(LLVMContext &Context) const
Definition ValueTypes.h:453
bool isInteger() const
Return true if this is an integer or a vector integer type.
Definition ValueTypes.h:152
static KnownBits makeConstant(const APInt &C)
Create known bits from a known constant.
Definition KnownBits.h:317
LLVM_ABI KnownBits sextInReg(unsigned SrcBitWidth) const
Return known bits for a in-register sign extension of the value we're tracking.
static LLVM_ABI KnownBits mulhu(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits from zero-extended multiply-hi.
unsigned countMinSignBits() const
Returns the number of times the sign bit is replicated into the other bits.
Definition KnownBits.h:271
static LLVM_ABI KnownBits smax(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for smax(LHS, RHS).
bool isNonNegative() const
Returns true if this value is known to be non-negative.
Definition KnownBits.h:108
bool isZero() const
Returns true if value is all zero.
Definition KnownBits.h:80
void makeNonNegative()
Make this value non-negative.
Definition KnownBits.h:127
static LLVM_ABI KnownBits usub_sat(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from llvm.usub.sat(LHS, RHS)
unsigned countMinTrailingZeros() const
Returns the minimum number of trailing zero bits.
Definition KnownBits.h:258
static LLVM_ABI KnownBits ashr(const KnownBits &LHS, const KnownBits &RHS, bool ShAmtNonZero=false, bool Exact=false)
Compute known bits for ashr(LHS, RHS).
static LLVM_ABI KnownBits urem(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for urem(LHS, RHS).
bool isUnknown() const
Returns true if we don't know any bits.
Definition KnownBits.h:66
unsigned countMaxTrailingZeros() const
Returns the maximum number of trailing zero bits possible.
Definition KnownBits.h:290
static LLVM_ABI std::optional< bool > ne(const KnownBits &LHS, const KnownBits &RHS)
Determine if these known bits always give the same ICMP_NE result.
void makeNegative()
Make this value negative.
Definition KnownBits.h:122
void setAllConflict()
Make all bits known to be both zero and one.
Definition KnownBits.h:99
KnownBits trunc(unsigned BitWidth) const
Return known bits for a truncation of the value we're tracking.
Definition KnownBits.h:167
KnownBits byteSwap() const
Definition KnownBits.h:538
unsigned countMaxPopulation() const
Returns the maximum number of bits that could be one.
Definition KnownBits.h:305
void setAllZero()
Make all bits known to be zero and discard any previous information.
Definition KnownBits.h:86
KnownBits reverseBits() const
Definition KnownBits.h:542
KnownBits concat(const KnownBits &Lo) const
Concatenate the bits from Lo onto the bottom of *this.
Definition KnownBits.h:249
unsigned getBitWidth() const
Get the bit width of this value.
Definition KnownBits.h:44
static LLVM_ABI KnownBits umax(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for umax(LHS, RHS).
KnownBits zext(unsigned BitWidth) const
Return known bits for a zero extension of the value we're tracking.
Definition KnownBits.h:178
void resetAll()
Resets the known state of all bits.
Definition KnownBits.h:74
KnownBits unionWith(const KnownBits &RHS) const
Returns KnownBits information that is known to be true for either this or RHS or both.
Definition KnownBits.h:337
static LLVM_ABI KnownBits lshr(const KnownBits &LHS, const KnownBits &RHS, bool ShAmtNonZero=false, bool Exact=false)
Compute known bits for lshr(LHS, RHS).
bool isNonZero() const
Returns true if this value is known to be non-zero.
Definition KnownBits.h:111
static LLVM_ABI KnownBits abdu(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for abdu(LHS, RHS).
KnownBits extractBits(unsigned NumBits, unsigned BitPosition) const
Return a subset of the known bits from [bitPosition,bitPosition+numBits).
Definition KnownBits.h:241
static LLVM_ABI KnownBits avgFloorU(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgFloorU.
KnownBits intersectWith(const KnownBits &RHS) const
Returns KnownBits information that is known to be true for both this and RHS.
Definition KnownBits.h:327
KnownBits sext(unsigned BitWidth) const
Return known bits for a sign extension of the value we're tracking.
Definition KnownBits.h:186
static LLVM_ABI KnownBits computeForSubBorrow(const KnownBits &LHS, KnownBits RHS, const KnownBits &Borrow)
Compute known bits results from subtracting RHS from LHS with 1-bit Borrow.
KnownBits zextOrTrunc(unsigned BitWidth) const
Return known bits for a zero extension or truncation of the value we're tracking.
Definition KnownBits.h:202
APInt getMaxValue() const
Return the maximal unsigned value possible given these KnownBits.
Definition KnownBits.h:148
static LLVM_ABI KnownBits abds(KnownBits LHS, KnownBits RHS)
Compute known bits for abds(LHS, RHS).
static LLVM_ABI KnownBits smin(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for smin(LHS, RHS).
static LLVM_ABI KnownBits mulhs(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits from sign-extended multiply-hi.
static LLVM_ABI KnownBits srem(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for srem(LHS, RHS).
static LLVM_ABI KnownBits udiv(const KnownBits &LHS, const KnownBits &RHS, bool Exact=false)
Compute known bits for udiv(LHS, RHS).
static LLVM_ABI KnownBits computeForAddSub(bool Add, bool NSW, bool NUW, const KnownBits &LHS, const KnownBits &RHS)
Compute known bits resulting from adding LHS and RHS.
Definition KnownBits.cpp:61
bool isStrictlyPositive() const
Returns true if this value is known to be positive.
Definition KnownBits.h:114
static LLVM_ABI KnownBits sdiv(const KnownBits &LHS, const KnownBits &RHS, bool Exact=false)
Compute known bits for sdiv(LHS, RHS).
static LLVM_ABI KnownBits avgFloorS(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgFloorS.
static bool haveNoCommonBitsSet(const KnownBits &LHS, const KnownBits &RHS)
Return true if LHS and RHS have no common bits set.
Definition KnownBits.h:342
bool isNegative() const
Returns true if this value is known to be negative.
Definition KnownBits.h:105
LLVM_ABI KnownBits truncSSat(unsigned BitWidth) const
Truncate with signed saturation (signed input -> signed output)
static LLVM_ABI KnownBits computeForAddCarry(const KnownBits &LHS, const KnownBits &RHS, const KnownBits &Carry)
Compute known bits resulting from adding LHS, RHS and a 1-bit Carry.
Definition KnownBits.cpp:54
unsigned countMaxLeadingZeros() const
Returns the maximum number of leading zero bits possible.
Definition KnownBits.h:296
void insertBits(const KnownBits &SubBits, unsigned BitPosition)
Insert the bits from a smaller known bits starting at bitPosition.
Definition KnownBits.h:235
static LLVM_ABI KnownBits avgCeilU(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgCeilU.
static LLVM_ABI KnownBits mul(const KnownBits &LHS, const KnownBits &RHS, bool NoUndefSelfMultiply=false)
Compute known bits resulting from multiplying LHS and RHS.
KnownBits anyext(unsigned BitWidth) const
Return known bits for an "any" extension of the value we're tracking, where we don't know anything ab...
Definition KnownBits.h:173
static LLVM_ABI KnownBits clmul(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for clmul(LHS, RHS).
LLVM_ABI KnownBits abs(bool IntMinIsPoison=false) const
Compute known bits for the absolute value.
LLVM_ABI KnownBits truncUSat(unsigned BitWidth) const
Truncate with unsigned saturation (unsigned input -> unsigned output)
static LLVM_ABI KnownBits shl(const KnownBits &LHS, const KnownBits &RHS, bool NUW=false, bool NSW=false, bool ShAmtNonZero=false)
Compute known bits for shl(LHS, RHS).
static LLVM_ABI KnownBits umin(const KnownBits &LHS, const KnownBits &RHS)
Compute known bits for umin(LHS, RHS).
LLVM_ABI KnownBits truncSSatU(unsigned BitWidth) const
Truncate with signed saturation to unsigned (signed input -> unsigned output)
static LLVM_ABI KnownBits avgCeilS(const KnownBits &LHS, const KnownBits &RHS)
Compute knownbits resulting from APIntOps::avgCeilS.
This class contains a discriminated union of information about pointers in memory operands,...
LLVM_ABI bool isDereferenceable(unsigned Size, LLVMContext &C, const DataLayout &DL) const
Return true if memory region [V, V+Offset+Size) is known to be dereferenceable.
LLVM_ABI unsigned getAddrSpace() const
Return the LLVM IR address space number that this pointer points into.
PointerUnion< const Value *, const PseudoSourceValue * > V
This is the IR pointer value for the access, or it is null if unknown.
MachinePointerInfo getWithOffset(int64_t O) const
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
Definition Alignment.h:106
static MemOp Set(uint64_t Size, bool DstAlignCanChange, Align DstAlign, bool IsZeroMemset, bool IsVolatile)
static MemOp Copy(uint64_t Size, bool DstAlignCanChange, Align DstAlign, Align SrcAlign, bool IsVolatile, bool MemcpyStrSrc=false)
static StringRef getLibcallImplName(RTLIB::LibcallImpl CallImpl)
Get the libcall routine name for the specified libcall implementation.
These are IR-level optimization flags that may be propagated to SDNodes.
This represents a list of ValueType's that has been intern'd by a SelectionDAG.
unsigned int NumVTs
Clients of various APIs that cause global effects on the DAG can optionally implement this interface.
virtual void NodeDeleted(SDNode *N, SDNode *E)
The node N that was deleted and, if E is not null, an equivalent node E that replaced it.
virtual void NodeInserted(SDNode *N)
The node N that was inserted.
virtual void NodeUpdated(SDNode *N)
The node N that was updated.
This structure contains all information that is necessary for lowering calls.
CallLoweringInfo & setLibCallee(CallingConv::ID CC, Type *ResultType, SDValue Target, ArgListTy &&ArgsList)
CallLoweringInfo & setDiscardResult(bool Value=true)
CallLoweringInfo & setDebugLoc(const SDLoc &dl)
CallLoweringInfo & setTailCall(bool Value=true)
CallLoweringInfo & setChain(SDValue InChain)