LLVM 19.0.0git
LegalizeVectorOps.cpp
Go to the documentation of this file.
1//===- LegalizeVectorOps.cpp - Implement SelectionDAG::LegalizeVectors ----===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file implements the SelectionDAG::LegalizeVectors method.
10//
11// The vector legalizer looks for vector operations which might need to be
12// scalarized and legalizes them. This is a separate step from Legalize because
13// scalarizing can introduce illegal types. For example, suppose we have an
14// ISD::SDIV of type v2i64 on x86-32. The type is legal (for example, addition
15// on a v2i64 is legal), but ISD::SDIV isn't legal, so we have to unroll the
16// operation, which introduces nodes with the illegal type i64 which must be
17// expanded. Similarly, suppose we have an ISD::SRA of type v16i8 on PowerPC;
18// the operation must be unrolled, which introduces nodes with the illegal
19// type i8 which must be promoted.
20//
21// This does not legalize vector manipulations like ISD::BUILD_VECTOR,
22// or operations that happen to take a vector which are custom-lowered;
23// the legalization for such operations never produces nodes
24// with illegal types, so it's okay to put off legalizing them until
25// SelectionDAG::Legalize runs.
26//
27//===----------------------------------------------------------------------===//
28
29#include "llvm/ADT/DenseMap.h"
39#include "llvm/IR/DataLayout.h"
42#include "llvm/Support/Debug.h"
44#include <cassert>
45#include <cstdint>
46#include <iterator>
47#include <utility>
48
49using namespace llvm;
50
51#define DEBUG_TYPE "legalizevectorops"
52
53namespace {
54
55class VectorLegalizer {
56 SelectionDAG& DAG;
57 const TargetLowering &TLI;
58 bool Changed = false; // Keep track of whether anything changed
59
60 /// For nodes that are of legal width, and that have more than one use, this
61 /// map indicates what regularized operand to use. This allows us to avoid
62 /// legalizing the same thing more than once.
64
65 /// Adds a node to the translation cache.
66 void AddLegalizedOperand(SDValue From, SDValue To) {
67 LegalizedNodes.insert(std::make_pair(From, To));
68 // If someone requests legalization of the new node, return itself.
69 if (From != To)
70 LegalizedNodes.insert(std::make_pair(To, To));
71 }
72
73 /// Legalizes the given node.
74 SDValue LegalizeOp(SDValue Op);
75
76 /// Assuming the node is legal, "legalize" the results.
77 SDValue TranslateLegalizeResults(SDValue Op, SDNode *Result);
78
79 /// Make sure Results are legal and update the translation cache.
80 SDValue RecursivelyLegalizeResults(SDValue Op,
82
83 /// Wrapper to interface LowerOperation with a vector of Results.
84 /// Returns false if the target wants to use default expansion. Otherwise
85 /// returns true. If return is true and the Results are empty, then the
86 /// target wants to keep the input node as is.
87 bool LowerOperationWrapper(SDNode *N, SmallVectorImpl<SDValue> &Results);
88
89 /// Implements unrolling a VSETCC.
90 SDValue UnrollVSETCC(SDNode *Node);
91
92 /// Implement expand-based legalization of vector operations.
93 ///
94 /// This is just a high-level routine to dispatch to specific code paths for
95 /// operations to legalize them.
97
98 /// Implements expansion for FP_TO_UINT; falls back to UnrollVectorOp if
99 /// FP_TO_SINT isn't legal.
100 void ExpandFP_TO_UINT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
101
102 /// Implements expansion for UINT_TO_FLOAT; falls back to UnrollVectorOp if
103 /// SINT_TO_FLOAT and SHR on vectors isn't legal.
104 void ExpandUINT_TO_FLOAT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
105
106 /// Implement expansion for SIGN_EXTEND_INREG using SRL and SRA.
107 SDValue ExpandSEXTINREG(SDNode *Node);
108
109 /// Implement expansion for ANY_EXTEND_VECTOR_INREG.
110 ///
111 /// Shuffles the low lanes of the operand into place and bitcasts to the proper
112 /// type. The contents of the bits in the extended part of each element are
113 /// undef.
114 SDValue ExpandANY_EXTEND_VECTOR_INREG(SDNode *Node);
115
116 /// Implement expansion for SIGN_EXTEND_VECTOR_INREG.
117 ///
118 /// Shuffles the low lanes of the operand into place, bitcasts to the proper
119 /// type, then shifts left and arithmetic shifts right to introduce a sign
120 /// extension.
121 SDValue ExpandSIGN_EXTEND_VECTOR_INREG(SDNode *Node);
122
123 /// Implement expansion for ZERO_EXTEND_VECTOR_INREG.
124 ///
125 /// Shuffles the low lanes of the operand into place and blends zeros into
126 /// the remaining lanes, finally bitcasting to the proper type.
127 SDValue ExpandZERO_EXTEND_VECTOR_INREG(SDNode *Node);
128
129 /// Expand bswap of vectors into a shuffle if legal.
130 SDValue ExpandBSWAP(SDNode *Node);
131
132 /// Implement vselect in terms of XOR, AND, OR when blend is not
133 /// supported by the target.
134 SDValue ExpandVSELECT(SDNode *Node);
135 SDValue ExpandVP_SELECT(SDNode *Node);
136 SDValue ExpandVP_MERGE(SDNode *Node);
137 SDValue ExpandVP_REM(SDNode *Node);
138 SDValue ExpandSELECT(SDNode *Node);
139 std::pair<SDValue, SDValue> ExpandLoad(SDNode *N);
140 SDValue ExpandStore(SDNode *N);
141 SDValue ExpandFNEG(SDNode *Node);
142 void ExpandFSUB(SDNode *Node, SmallVectorImpl<SDValue> &Results);
143 void ExpandSETCC(SDNode *Node, SmallVectorImpl<SDValue> &Results);
144 void ExpandBITREVERSE(SDNode *Node, SmallVectorImpl<SDValue> &Results);
145 void ExpandUADDSUBO(SDNode *Node, SmallVectorImpl<SDValue> &Results);
146 void ExpandSADDSUBO(SDNode *Node, SmallVectorImpl<SDValue> &Results);
147 void ExpandMULO(SDNode *Node, SmallVectorImpl<SDValue> &Results);
148 void ExpandFixedPointDiv(SDNode *Node, SmallVectorImpl<SDValue> &Results);
149 void ExpandStrictFPOp(SDNode *Node, SmallVectorImpl<SDValue> &Results);
150 void ExpandREM(SDNode *Node, SmallVectorImpl<SDValue> &Results);
151
152 bool tryExpandVecMathCall(SDNode *Node, RTLIB::Libcall LC,
154 bool tryExpandVecMathCall(SDNode *Node, RTLIB::Libcall Call_F32,
155 RTLIB::Libcall Call_F64, RTLIB::Libcall Call_F80,
156 RTLIB::Libcall Call_F128,
157 RTLIB::Libcall Call_PPCF128,
159
160 void UnrollStrictFPOp(SDNode *Node, SmallVectorImpl<SDValue> &Results);
161
162 /// Implements vector promotion.
163 ///
164 /// This is essentially just bitcasting the operands to a different type and
165 /// bitcasting the result back to the original type.
167
168 /// Implements [SU]INT_TO_FP vector promotion.
169 ///
170 /// This is a [zs]ext of the input operand to a larger integer type.
171 void PromoteINT_TO_FP(SDNode *Node, SmallVectorImpl<SDValue> &Results);
172
173 /// Implements FP_TO_[SU]INT vector promotion of the result type.
174 ///
175 /// It is promoted to a larger integer type. The result is then
176 /// truncated back to the original type.
177 void PromoteFP_TO_INT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
178
179 /// Implements vector setcc operation promotion.
180 ///
181 /// All vector operands are promoted to a vector type with larger element
182 /// type.
183 void PromoteSETCC(SDNode *Node, SmallVectorImpl<SDValue> &Results);
184
185 void PromoteSTRICT(SDNode *Node, SmallVectorImpl<SDValue> &Results);
186
187public:
188 VectorLegalizer(SelectionDAG& dag) :
189 DAG(dag), TLI(dag.getTargetLoweringInfo()) {}
190
191 /// Begin legalizer the vector operations in the DAG.
192 bool Run();
193};
194
195} // end anonymous namespace
196
197bool VectorLegalizer::Run() {
198 // Before we start legalizing vector nodes, check if there are any vectors.
199 bool HasVectors = false;
200 for (SelectionDAG::allnodes_iterator I = DAG.allnodes_begin(),
201 E = std::prev(DAG.allnodes_end()); I != std::next(E); ++I) {
202 // Check if the values of the nodes contain vectors. We don't need to check
203 // the operands because we are going to check their values at some point.
204 HasVectors = llvm::any_of(I->values(), [](EVT T) { return T.isVector(); });
205
206 // If we found a vector node we can start the legalization.
207 if (HasVectors)
208 break;
209 }
210
211 // If this basic block has no vectors then no need to legalize vectors.
212 if (!HasVectors)
213 return false;
214
215 // The legalize process is inherently a bottom-up recursive process (users
216 // legalize their uses before themselves). Given infinite stack space, we
217 // could just start legalizing on the root and traverse the whole graph. In
218 // practice however, this causes us to run out of stack space on large basic
219 // blocks. To avoid this problem, compute an ordering of the nodes where each
220 // node is only legalized after all of its operands are legalized.
221 DAG.AssignTopologicalOrder();
222 for (SelectionDAG::allnodes_iterator I = DAG.allnodes_begin(),
223 E = std::prev(DAG.allnodes_end()); I != std::next(E); ++I)
224 LegalizeOp(SDValue(&*I, 0));
225
226 // Finally, it's possible the root changed. Get the new root.
227 SDValue OldRoot = DAG.getRoot();
228 assert(LegalizedNodes.count(OldRoot) && "Root didn't get legalized?");
229 DAG.setRoot(LegalizedNodes[OldRoot]);
230
231 LegalizedNodes.clear();
232
233 // Remove dead nodes now.
234 DAG.RemoveDeadNodes();
235
236 return Changed;
237}
238
239SDValue VectorLegalizer::TranslateLegalizeResults(SDValue Op, SDNode *Result) {
240 assert(Op->getNumValues() == Result->getNumValues() &&
241 "Unexpected number of results");
242 // Generic legalization: just pass the operand through.
243 for (unsigned i = 0, e = Op->getNumValues(); i != e; ++i)
244 AddLegalizedOperand(Op.getValue(i), SDValue(Result, i));
245 return SDValue(Result, Op.getResNo());
246}
247
249VectorLegalizer::RecursivelyLegalizeResults(SDValue Op,
251 assert(Results.size() == Op->getNumValues() &&
252 "Unexpected number of results");
253 // Make sure that the generated code is itself legal.
254 for (unsigned i = 0, e = Results.size(); i != e; ++i) {
255 Results[i] = LegalizeOp(Results[i]);
256 AddLegalizedOperand(Op.getValue(i), Results[i]);
257 }
258
259 return Results[Op.getResNo()];
260}
261
262SDValue VectorLegalizer::LegalizeOp(SDValue Op) {
263 // Note that LegalizeOp may be reentered even from single-use nodes, which
264 // means that we always must cache transformed nodes.
265 DenseMap<SDValue, SDValue>::iterator I = LegalizedNodes.find(Op);
266 if (I != LegalizedNodes.end()) return I->second;
267
268 // Legalize the operands
270 for (const SDValue &Oper : Op->op_values())
271 Ops.push_back(LegalizeOp(Oper));
272
273 SDNode *Node = DAG.UpdateNodeOperands(Op.getNode(), Ops);
274
275 bool HasVectorValueOrOp =
276 llvm::any_of(Node->values(), [](EVT T) { return T.isVector(); }) ||
277 llvm::any_of(Node->op_values(),
278 [](SDValue O) { return O.getValueType().isVector(); });
279 if (!HasVectorValueOrOp)
280 return TranslateLegalizeResults(Op, Node);
281
282 TargetLowering::LegalizeAction Action = TargetLowering::Legal;
283 EVT ValVT;
284 switch (Op.getOpcode()) {
285 default:
286 return TranslateLegalizeResults(Op, Node);
287 case ISD::LOAD: {
288 LoadSDNode *LD = cast<LoadSDNode>(Node);
289 ISD::LoadExtType ExtType = LD->getExtensionType();
290 EVT LoadedVT = LD->getMemoryVT();
291 if (LoadedVT.isVector() && ExtType != ISD::NON_EXTLOAD)
292 Action = TLI.getLoadExtAction(ExtType, LD->getValueType(0), LoadedVT);
293 break;
294 }
295 case ISD::STORE: {
296 StoreSDNode *ST = cast<StoreSDNode>(Node);
297 EVT StVT = ST->getMemoryVT();
298 MVT ValVT = ST->getValue().getSimpleValueType();
299 if (StVT.isVector() && ST->isTruncatingStore())
300 Action = TLI.getTruncStoreAction(ValVT, StVT);
301 break;
302 }
304 Action = TLI.getOperationAction(Node->getOpcode(), Node->getValueType(0));
305 // This operation lies about being legal: when it claims to be legal,
306 // it should actually be expanded.
307 if (Action == TargetLowering::Legal)
308 Action = TargetLowering::Expand;
309 break;
310#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
311 case ISD::STRICT_##DAGN:
312#include "llvm/IR/ConstrainedOps.def"
313 ValVT = Node->getValueType(0);
314 if (Op.getOpcode() == ISD::STRICT_SINT_TO_FP ||
315 Op.getOpcode() == ISD::STRICT_UINT_TO_FP)
316 ValVT = Node->getOperand(1).getValueType();
317 if (Op.getOpcode() == ISD::STRICT_FSETCC ||
318 Op.getOpcode() == ISD::STRICT_FSETCCS) {
319 MVT OpVT = Node->getOperand(1).getSimpleValueType();
320 ISD::CondCode CCCode = cast<CondCodeSDNode>(Node->getOperand(3))->get();
321 Action = TLI.getCondCodeAction(CCCode, OpVT);
322 if (Action == TargetLowering::Legal)
323 Action = TLI.getOperationAction(Node->getOpcode(), OpVT);
324 } else {
325 Action = TLI.getOperationAction(Node->getOpcode(), ValVT);
326 }
327 // If we're asked to expand a strict vector floating-point operation,
328 // by default we're going to simply unroll it. That is usually the
329 // best approach, except in the case where the resulting strict (scalar)
330 // operations would themselves use the fallback mutation to non-strict.
331 // In that specific case, just do the fallback on the vector op.
332 if (Action == TargetLowering::Expand && !TLI.isStrictFPEnabled() &&
333 TLI.getStrictFPOperationAction(Node->getOpcode(), ValVT) ==
334 TargetLowering::Legal) {
335 EVT EltVT = ValVT.getVectorElementType();
336 if (TLI.getOperationAction(Node->getOpcode(), EltVT)
337 == TargetLowering::Expand &&
338 TLI.getStrictFPOperationAction(Node->getOpcode(), EltVT)
339 == TargetLowering::Legal)
340 Action = TargetLowering::Legal;
341 }
342 break;
343 case ISD::ADD:
344 case ISD::SUB:
345 case ISD::MUL:
346 case ISD::MULHS:
347 case ISD::MULHU:
348 case ISD::SDIV:
349 case ISD::UDIV:
350 case ISD::SREM:
351 case ISD::UREM:
352 case ISD::SDIVREM:
353 case ISD::UDIVREM:
354 case ISD::FADD:
355 case ISD::FSUB:
356 case ISD::FMUL:
357 case ISD::FDIV:
358 case ISD::FREM:
359 case ISD::AND:
360 case ISD::OR:
361 case ISD::XOR:
362 case ISD::SHL:
363 case ISD::SRA:
364 case ISD::SRL:
365 case ISD::FSHL:
366 case ISD::FSHR:
367 case ISD::ROTL:
368 case ISD::ROTR:
369 case ISD::ABS:
370 case ISD::ABDS:
371 case ISD::ABDU:
372 case ISD::BSWAP:
373 case ISD::BITREVERSE:
374 case ISD::CTLZ:
375 case ISD::CTTZ:
378 case ISD::CTPOP:
379 case ISD::SELECT:
380 case ISD::VSELECT:
381 case ISD::SELECT_CC:
382 case ISD::ZERO_EXTEND:
383 case ISD::ANY_EXTEND:
384 case ISD::TRUNCATE:
385 case ISD::SIGN_EXTEND:
386 case ISD::FP_TO_SINT:
387 case ISD::FP_TO_UINT:
388 case ISD::FNEG:
389 case ISD::FABS:
390 case ISD::FMINNUM:
391 case ISD::FMAXNUM:
394 case ISD::FMINIMUM:
395 case ISD::FMAXIMUM:
396 case ISD::FCOPYSIGN:
397 case ISD::FSQRT:
398 case ISD::FSIN:
399 case ISD::FCOS:
400 case ISD::FLDEXP:
401 case ISD::FPOWI:
402 case ISD::FPOW:
403 case ISD::FLOG:
404 case ISD::FLOG2:
405 case ISD::FLOG10:
406 case ISD::FEXP:
407 case ISD::FEXP2:
408 case ISD::FEXP10:
409 case ISD::FCEIL:
410 case ISD::FTRUNC:
411 case ISD::FRINT:
412 case ISD::FNEARBYINT:
413 case ISD::FROUND:
414 case ISD::FROUNDEVEN:
415 case ISD::FFLOOR:
416 case ISD::FP_ROUND:
417 case ISD::FP_EXTEND:
419 case ISD::FMA:
424 case ISD::SMIN:
425 case ISD::SMAX:
426 case ISD::UMIN:
427 case ISD::UMAX:
428 case ISD::SMUL_LOHI:
429 case ISD::UMUL_LOHI:
430 case ISD::SADDO:
431 case ISD::UADDO:
432 case ISD::SSUBO:
433 case ISD::USUBO:
434 case ISD::SMULO:
435 case ISD::UMULO:
437 case ISD::FFREXP:
438 case ISD::SADDSAT:
439 case ISD::UADDSAT:
440 case ISD::SSUBSAT:
441 case ISD::USUBSAT:
442 case ISD::SSHLSAT:
443 case ISD::USHLSAT:
446 case ISD::MGATHER:
447 Action = TLI.getOperationAction(Node->getOpcode(), Node->getValueType(0));
448 break;
449 case ISD::SMULFIX:
450 case ISD::SMULFIXSAT:
451 case ISD::UMULFIX:
452 case ISD::UMULFIXSAT:
453 case ISD::SDIVFIX:
454 case ISD::SDIVFIXSAT:
455 case ISD::UDIVFIX:
456 case ISD::UDIVFIXSAT: {
457 unsigned Scale = Node->getConstantOperandVal(2);
458 Action = TLI.getFixedPointOperationAction(Node->getOpcode(),
459 Node->getValueType(0), Scale);
460 break;
461 }
462 case ISD::LRINT:
463 case ISD::LLRINT:
464 case ISD::SINT_TO_FP:
465 case ISD::UINT_TO_FP:
481 Action = TLI.getOperationAction(Node->getOpcode(),
482 Node->getOperand(0).getValueType());
483 break;
486 Action = TLI.getOperationAction(Node->getOpcode(),
487 Node->getOperand(1).getValueType());
488 break;
489 case ISD::SETCC: {
490 MVT OpVT = Node->getOperand(0).getSimpleValueType();
491 ISD::CondCode CCCode = cast<CondCodeSDNode>(Node->getOperand(2))->get();
492 Action = TLI.getCondCodeAction(CCCode, OpVT);
493 if (Action == TargetLowering::Legal)
494 Action = TLI.getOperationAction(Node->getOpcode(), OpVT);
495 break;
496 }
497
498#define BEGIN_REGISTER_VP_SDNODE(VPID, LEGALPOS, ...) \
499 case ISD::VPID: { \
500 EVT LegalizeVT = LEGALPOS < 0 ? Node->getValueType(-(1 + LEGALPOS)) \
501 : Node->getOperand(LEGALPOS).getValueType(); \
502 if (ISD::VPID == ISD::VP_SETCC) { \
503 ISD::CondCode CCCode = cast<CondCodeSDNode>(Node->getOperand(2))->get(); \
504 Action = TLI.getCondCodeAction(CCCode, LegalizeVT.getSimpleVT()); \
505 if (Action != TargetLowering::Legal) \
506 break; \
507 } \
508 /* Defer non-vector results to LegalizeDAG. */ \
509 if (!Node->getValueType(0).isVector()) { \
510 Action = TargetLowering::Legal; \
511 break; \
512 } \
513 Action = TLI.getOperationAction(Node->getOpcode(), LegalizeVT); \
514 } break;
515#include "llvm/IR/VPIntrinsics.def"
516 }
517
518 LLVM_DEBUG(dbgs() << "\nLegalizing vector op: "; Node->dump(&DAG));
519
520 SmallVector<SDValue, 8> ResultVals;
521 switch (Action) {
522 default: llvm_unreachable("This action is not supported yet!");
523 case TargetLowering::Promote:
524 assert((Op.getOpcode() != ISD::LOAD && Op.getOpcode() != ISD::STORE) &&
525 "This action is not supported yet!");
526 LLVM_DEBUG(dbgs() << "Promoting\n");
527 Promote(Node, ResultVals);
528 assert(!ResultVals.empty() && "No results for promotion?");
529 break;
530 case TargetLowering::Legal:
531 LLVM_DEBUG(dbgs() << "Legal node: nothing to do\n");
532 break;
533 case TargetLowering::Custom:
534 LLVM_DEBUG(dbgs() << "Trying custom legalization\n");
535 if (LowerOperationWrapper(Node, ResultVals))
536 break;
537 LLVM_DEBUG(dbgs() << "Could not custom legalize node\n");
538 [[fallthrough]];
539 case TargetLowering::Expand:
540 LLVM_DEBUG(dbgs() << "Expanding\n");
541 Expand(Node, ResultVals);
542 break;
543 }
544
545 if (ResultVals.empty())
546 return TranslateLegalizeResults(Op, Node);
547
548 Changed = true;
549 return RecursivelyLegalizeResults(Op, ResultVals);
550}
551
552// FIXME: This is very similar to TargetLowering::LowerOperationWrapper. Can we
553// merge them somehow?
554bool VectorLegalizer::LowerOperationWrapper(SDNode *Node,
556 SDValue Res = TLI.LowerOperation(SDValue(Node, 0), DAG);
557
558 if (!Res.getNode())
559 return false;
560
561 if (Res == SDValue(Node, 0))
562 return true;
563
564 // If the original node has one result, take the return value from
565 // LowerOperation as is. It might not be result number 0.
566 if (Node->getNumValues() == 1) {
567 Results.push_back(Res);
568 return true;
569 }
570
571 // If the original node has multiple results, then the return node should
572 // have the same number of results.
573 assert((Node->getNumValues() == Res->getNumValues()) &&
574 "Lowering returned the wrong number of results!");
575
576 // Places new result values base on N result number.
577 for (unsigned I = 0, E = Node->getNumValues(); I != E; ++I)
578 Results.push_back(Res.getValue(I));
579
580 return true;
581}
582
583void VectorLegalizer::PromoteSETCC(SDNode *Node,
585 MVT VecVT = Node->getOperand(0).getSimpleValueType();
586 MVT NewVecVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VecVT);
587
588 unsigned ExtOp = VecVT.isFloatingPoint() ? ISD::FP_EXTEND : ISD::ANY_EXTEND;
589
590 SDLoc DL(Node);
591 SmallVector<SDValue, 5> Operands(Node->getNumOperands());
592
593 Operands[0] = DAG.getNode(ExtOp, DL, NewVecVT, Node->getOperand(0));
594 Operands[1] = DAG.getNode(ExtOp, DL, NewVecVT, Node->getOperand(1));
595 Operands[2] = Node->getOperand(2);
596
597 if (Node->getOpcode() == ISD::VP_SETCC) {
598 Operands[3] = Node->getOperand(3); // mask
599 Operands[4] = Node->getOperand(4); // evl
600 }
601
602 SDValue Res = DAG.getNode(Node->getOpcode(), DL, Node->getSimpleValueType(0),
603 Operands, Node->getFlags());
604
605 Results.push_back(Res);
606}
607
608void VectorLegalizer::PromoteSTRICT(SDNode *Node,
610 MVT VecVT = Node->getOperand(1).getSimpleValueType();
611 MVT NewVecVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VecVT);
612
613 assert(VecVT.isFloatingPoint());
614
615 SDLoc DL(Node);
616 SmallVector<SDValue, 5> Operands(Node->getNumOperands());
618
619 for (unsigned j = 1; j != Node->getNumOperands(); ++j)
620 if (Node->getOperand(j).getValueType().isVector() &&
621 !(ISD::isVPOpcode(Node->getOpcode()) &&
622 ISD::getVPMaskIdx(Node->getOpcode()) == j)) // Skip mask operand.
623 {
624 // promote the vector operand.
625 SDValue Ext =
626 DAG.getNode(ISD::STRICT_FP_EXTEND, DL, {NewVecVT, MVT::Other},
627 {Node->getOperand(0), Node->getOperand(j)});
628 Operands[j] = Ext.getValue(0);
629 Chains.push_back(Ext.getValue(1));
630 } else
631 Operands[j] = Node->getOperand(j); // Skip no vector operand.
632
633 SDVTList VTs = DAG.getVTList(NewVecVT, Node->getValueType(1));
634
635 Operands[0] = DAG.getNode(ISD::TokenFactor, DL, MVT::Other, Chains);
636
637 SDValue Res =
638 DAG.getNode(Node->getOpcode(), DL, VTs, Operands, Node->getFlags());
639
640 SDValue Round =
641 DAG.getNode(ISD::STRICT_FP_ROUND, DL, {VecVT, MVT::Other},
642 {Res.getValue(1), Res.getValue(0),
643 DAG.getIntPtrConstant(0, DL, /*isTarget=*/true)});
644
645 Results.push_back(Round.getValue(0));
646 Results.push_back(Round.getValue(1));
647}
648
649void VectorLegalizer::Promote(SDNode *Node, SmallVectorImpl<SDValue> &Results) {
650 // For a few operations there is a specific concept for promotion based on
651 // the operand's type.
652 switch (Node->getOpcode()) {
653 case ISD::SINT_TO_FP:
654 case ISD::UINT_TO_FP:
657 // "Promote" the operation by extending the operand.
658 PromoteINT_TO_FP(Node, Results);
659 return;
660 case ISD::FP_TO_UINT:
661 case ISD::FP_TO_SINT:
664 // Promote the operation by extending the operand.
665 PromoteFP_TO_INT(Node, Results);
666 return;
667 case ISD::VP_SETCC:
668 case ISD::SETCC:
669 // Promote the operation by extending the operand.
670 PromoteSETCC(Node, Results);
671 return;
672 case ISD::STRICT_FADD:
673 case ISD::STRICT_FSUB:
674 case ISD::STRICT_FMUL:
675 case ISD::STRICT_FDIV:
677 case ISD::STRICT_FMA:
678 PromoteSTRICT(Node, Results);
679 return;
680 case ISD::FP_ROUND:
681 case ISD::FP_EXTEND:
682 // These operations are used to do promotion so they can't be promoted
683 // themselves.
684 llvm_unreachable("Don't know how to promote this operation!");
685 }
686
687 // There are currently two cases of vector promotion:
688 // 1) Bitcasting a vector of integers to a different type to a vector of the
689 // same overall length. For example, x86 promotes ISD::AND v2i32 to v1i64.
690 // 2) Extending a vector of floats to a vector of the same number of larger
691 // floats. For example, AArch64 promotes ISD::FADD on v4f16 to v4f32.
692 assert(Node->getNumValues() == 1 &&
693 "Can't promote a vector with multiple results!");
694 MVT VT = Node->getSimpleValueType(0);
695 MVT NVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VT);
696 SDLoc dl(Node);
697 SmallVector<SDValue, 4> Operands(Node->getNumOperands());
698
699 for (unsigned j = 0; j != Node->getNumOperands(); ++j) {
700 // Do not promote the mask operand of a VP OP.
701 bool SkipPromote = ISD::isVPOpcode(Node->getOpcode()) &&
702 ISD::getVPMaskIdx(Node->getOpcode()) == j;
703 if (Node->getOperand(j).getValueType().isVector() && !SkipPromote)
704 if (Node->getOperand(j)
705 .getValueType()
706 .getVectorElementType()
707 .isFloatingPoint() &&
709 Operands[j] = DAG.getNode(ISD::FP_EXTEND, dl, NVT, Node->getOperand(j));
710 else
711 Operands[j] = DAG.getNode(ISD::BITCAST, dl, NVT, Node->getOperand(j));
712 else
713 Operands[j] = Node->getOperand(j);
714 }
715
716 SDValue Res =
717 DAG.getNode(Node->getOpcode(), dl, NVT, Operands, Node->getFlags());
718
719 if ((VT.isFloatingPoint() && NVT.isFloatingPoint()) ||
722 Res = DAG.getNode(ISD::FP_ROUND, dl, VT, Res,
723 DAG.getIntPtrConstant(0, dl, /*isTarget=*/true));
724 else
725 Res = DAG.getNode(ISD::BITCAST, dl, VT, Res);
726
727 Results.push_back(Res);
728}
729
730void VectorLegalizer::PromoteINT_TO_FP(SDNode *Node,
732 // INT_TO_FP operations may require the input operand be promoted even
733 // when the type is otherwise legal.
734 bool IsStrict = Node->isStrictFPOpcode();
735 MVT VT = Node->getOperand(IsStrict ? 1 : 0).getSimpleValueType();
736 MVT NVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VT);
738 "Vectors have different number of elements!");
739
740 SDLoc dl(Node);
741 SmallVector<SDValue, 4> Operands(Node->getNumOperands());
742
743 unsigned Opc = (Node->getOpcode() == ISD::UINT_TO_FP ||
744 Node->getOpcode() == ISD::STRICT_UINT_TO_FP)
747 for (unsigned j = 0; j != Node->getNumOperands(); ++j) {
748 if (Node->getOperand(j).getValueType().isVector())
749 Operands[j] = DAG.getNode(Opc, dl, NVT, Node->getOperand(j));
750 else
751 Operands[j] = Node->getOperand(j);
752 }
753
754 if (IsStrict) {
755 SDValue Res = DAG.getNode(Node->getOpcode(), dl,
756 {Node->getValueType(0), MVT::Other}, Operands);
757 Results.push_back(Res);
758 Results.push_back(Res.getValue(1));
759 return;
760 }
761
762 SDValue Res =
763 DAG.getNode(Node->getOpcode(), dl, Node->getValueType(0), Operands);
764 Results.push_back(Res);
765}
766
767// For FP_TO_INT we promote the result type to a vector type with wider
768// elements and then truncate the result. This is different from the default
769// PromoteVector which uses bitcast to promote thus assumning that the
770// promoted vector type has the same overall size.
771void VectorLegalizer::PromoteFP_TO_INT(SDNode *Node,
773 MVT VT = Node->getSimpleValueType(0);
774 MVT NVT = TLI.getTypeToPromoteTo(Node->getOpcode(), VT);
775 bool IsStrict = Node->isStrictFPOpcode();
777 "Vectors have different number of elements!");
778
779 unsigned NewOpc = Node->getOpcode();
780 // Change FP_TO_UINT to FP_TO_SINT if possible.
781 // TODO: Should we only do this if FP_TO_UINT itself isn't legal?
782 if (NewOpc == ISD::FP_TO_UINT &&
783 TLI.isOperationLegalOrCustom(ISD::FP_TO_SINT, NVT))
784 NewOpc = ISD::FP_TO_SINT;
785
786 if (NewOpc == ISD::STRICT_FP_TO_UINT &&
787 TLI.isOperationLegalOrCustom(ISD::STRICT_FP_TO_SINT, NVT))
788 NewOpc = ISD::STRICT_FP_TO_SINT;
789
790 SDLoc dl(Node);
791 SDValue Promoted, Chain;
792 if (IsStrict) {
793 Promoted = DAG.getNode(NewOpc, dl, {NVT, MVT::Other},
794 {Node->getOperand(0), Node->getOperand(1)});
795 Chain = Promoted.getValue(1);
796 } else
797 Promoted = DAG.getNode(NewOpc, dl, NVT, Node->getOperand(0));
798
799 // Assert that the converted value fits in the original type. If it doesn't
800 // (eg: because the value being converted is too big), then the result of the
801 // original operation was undefined anyway, so the assert is still correct.
802 if (Node->getOpcode() == ISD::FP_TO_UINT ||
803 Node->getOpcode() == ISD::STRICT_FP_TO_UINT)
804 NewOpc = ISD::AssertZext;
805 else
806 NewOpc = ISD::AssertSext;
807
808 Promoted = DAG.getNode(NewOpc, dl, NVT, Promoted,
809 DAG.getValueType(VT.getScalarType()));
810 Promoted = DAG.getNode(ISD::TRUNCATE, dl, VT, Promoted);
811 Results.push_back(Promoted);
812 if (IsStrict)
813 Results.push_back(Chain);
814}
815
816std::pair<SDValue, SDValue> VectorLegalizer::ExpandLoad(SDNode *N) {
817 LoadSDNode *LD = cast<LoadSDNode>(N);
818 return TLI.scalarizeVectorLoad(LD, DAG);
819}
820
821SDValue VectorLegalizer::ExpandStore(SDNode *N) {
822 StoreSDNode *ST = cast<StoreSDNode>(N);
823 SDValue TF = TLI.scalarizeVectorStore(ST, DAG);
824 return TF;
825}
826
827void VectorLegalizer::Expand(SDNode *Node, SmallVectorImpl<SDValue> &Results) {
828 switch (Node->getOpcode()) {
829 case ISD::LOAD: {
830 std::pair<SDValue, SDValue> Tmp = ExpandLoad(Node);
831 Results.push_back(Tmp.first);
832 Results.push_back(Tmp.second);
833 return;
834 }
835 case ISD::STORE:
836 Results.push_back(ExpandStore(Node));
837 return;
839 for (unsigned i = 0, e = Node->getNumValues(); i != e; ++i)
840 Results.push_back(Node->getOperand(i));
841 return;
843 Results.push_back(ExpandSEXTINREG(Node));
844 return;
846 Results.push_back(ExpandANY_EXTEND_VECTOR_INREG(Node));
847 return;
849 Results.push_back(ExpandSIGN_EXTEND_VECTOR_INREG(Node));
850 return;
852 Results.push_back(ExpandZERO_EXTEND_VECTOR_INREG(Node));
853 return;
854 case ISD::BSWAP:
855 Results.push_back(ExpandBSWAP(Node));
856 return;
857 case ISD::VP_BSWAP:
858 Results.push_back(TLI.expandVPBSWAP(Node, DAG));
859 return;
860 case ISD::VSELECT:
861 Results.push_back(ExpandVSELECT(Node));
862 return;
863 case ISD::VP_SELECT:
864 Results.push_back(ExpandVP_SELECT(Node));
865 return;
866 case ISD::VP_SREM:
867 case ISD::VP_UREM:
868 if (SDValue Expanded = ExpandVP_REM(Node)) {
869 Results.push_back(Expanded);
870 return;
871 }
872 break;
873 case ISD::SELECT:
874 Results.push_back(ExpandSELECT(Node));
875 return;
876 case ISD::SELECT_CC: {
877 if (Node->getValueType(0).isScalableVector()) {
878 EVT CondVT = TLI.getSetCCResultType(
879 DAG.getDataLayout(), *DAG.getContext(), Node->getValueType(0));
880 SDValue SetCC =
881 DAG.getNode(ISD::SETCC, SDLoc(Node), CondVT, Node->getOperand(0),
882 Node->getOperand(1), Node->getOperand(4));
883 Results.push_back(DAG.getSelect(SDLoc(Node), Node->getValueType(0), SetCC,
884 Node->getOperand(2),
885 Node->getOperand(3)));
886 return;
887 }
888 break;
889 }
890 case ISD::FP_TO_UINT:
891 ExpandFP_TO_UINT(Node, Results);
892 return;
893 case ISD::UINT_TO_FP:
894 ExpandUINT_TO_FLOAT(Node, Results);
895 return;
896 case ISD::FNEG:
897 Results.push_back(ExpandFNEG(Node));
898 return;
899 case ISD::FSUB:
900 ExpandFSUB(Node, Results);
901 return;
902 case ISD::SETCC:
903 case ISD::VP_SETCC:
904 ExpandSETCC(Node, Results);
905 return;
906 case ISD::ABS:
907 if (SDValue Expanded = TLI.expandABS(Node, DAG)) {
908 Results.push_back(Expanded);
909 return;
910 }
911 break;
912 case ISD::ABDS:
913 case ISD::ABDU:
914 if (SDValue Expanded = TLI.expandABD(Node, DAG)) {
915 Results.push_back(Expanded);
916 return;
917 }
918 break;
919 case ISD::BITREVERSE:
920 ExpandBITREVERSE(Node, Results);
921 return;
922 case ISD::VP_BITREVERSE:
923 if (SDValue Expanded = TLI.expandVPBITREVERSE(Node, DAG)) {
924 Results.push_back(Expanded);
925 return;
926 }
927 break;
928 case ISD::CTPOP:
929 if (SDValue Expanded = TLI.expandCTPOP(Node, DAG)) {
930 Results.push_back(Expanded);
931 return;
932 }
933 break;
934 case ISD::VP_CTPOP:
935 if (SDValue Expanded = TLI.expandVPCTPOP(Node, DAG)) {
936 Results.push_back(Expanded);
937 return;
938 }
939 break;
940 case ISD::CTLZ:
942 if (SDValue Expanded = TLI.expandCTLZ(Node, DAG)) {
943 Results.push_back(Expanded);
944 return;
945 }
946 break;
947 case ISD::VP_CTLZ:
948 case ISD::VP_CTLZ_ZERO_UNDEF:
949 if (SDValue Expanded = TLI.expandVPCTLZ(Node, DAG)) {
950 Results.push_back(Expanded);
951 return;
952 }
953 break;
954 case ISD::CTTZ:
956 if (SDValue Expanded = TLI.expandCTTZ(Node, DAG)) {
957 Results.push_back(Expanded);
958 return;
959 }
960 break;
961 case ISD::VP_CTTZ:
962 case ISD::VP_CTTZ_ZERO_UNDEF:
963 if (SDValue Expanded = TLI.expandVPCTTZ(Node, DAG)) {
964 Results.push_back(Expanded);
965 return;
966 }
967 break;
968 case ISD::FSHL:
969 case ISD::VP_FSHL:
970 case ISD::FSHR:
971 case ISD::VP_FSHR:
972 if (SDValue Expanded = TLI.expandFunnelShift(Node, DAG)) {
973 Results.push_back(Expanded);
974 return;
975 }
976 break;
977 case ISD::ROTL:
978 case ISD::ROTR:
979 if (SDValue Expanded = TLI.expandROT(Node, false /*AllowVectorOps*/, DAG)) {
980 Results.push_back(Expanded);
981 return;
982 }
983 break;
984 case ISD::FMINNUM:
985 case ISD::FMAXNUM:
986 if (SDValue Expanded = TLI.expandFMINNUM_FMAXNUM(Node, DAG)) {
987 Results.push_back(Expanded);
988 return;
989 }
990 break;
991 case ISD::FMINIMUM:
992 case ISD::FMAXIMUM:
993 if (SDValue Expanded = TLI.expandFMINIMUM_FMAXIMUM(Node, DAG)) {
994 Results.push_back(Expanded);
995 return;
996 }
997 break;
998 case ISD::SMIN:
999 case ISD::SMAX:
1000 case ISD::UMIN:
1001 case ISD::UMAX:
1002 if (SDValue Expanded = TLI.expandIntMINMAX(Node, DAG)) {
1003 Results.push_back(Expanded);
1004 return;
1005 }
1006 break;
1007 case ISD::UADDO:
1008 case ISD::USUBO:
1009 ExpandUADDSUBO(Node, Results);
1010 return;
1011 case ISD::SADDO:
1012 case ISD::SSUBO:
1013 ExpandSADDSUBO(Node, Results);
1014 return;
1015 case ISD::UMULO:
1016 case ISD::SMULO:
1017 ExpandMULO(Node, Results);
1018 return;
1019 case ISD::USUBSAT:
1020 case ISD::SSUBSAT:
1021 case ISD::UADDSAT:
1022 case ISD::SADDSAT:
1023 if (SDValue Expanded = TLI.expandAddSubSat(Node, DAG)) {
1024 Results.push_back(Expanded);
1025 return;
1026 }
1027 break;
1028 case ISD::USHLSAT:
1029 case ISD::SSHLSAT:
1030 if (SDValue Expanded = TLI.expandShlSat(Node, DAG)) {
1031 Results.push_back(Expanded);
1032 return;
1033 }
1034 break;
1037 // Expand the fpsosisat if it is scalable to prevent it from unrolling below.
1038 if (Node->getValueType(0).isScalableVector()) {
1039 if (SDValue Expanded = TLI.expandFP_TO_INT_SAT(Node, DAG)) {
1040 Results.push_back(Expanded);
1041 return;
1042 }
1043 }
1044 break;
1045 case ISD::SMULFIX:
1046 case ISD::UMULFIX:
1047 if (SDValue Expanded = TLI.expandFixedPointMul(Node, DAG)) {
1048 Results.push_back(Expanded);
1049 return;
1050 }
1051 break;
1052 case ISD::SMULFIXSAT:
1053 case ISD::UMULFIXSAT:
1054 // FIXME: We do not expand SMULFIXSAT/UMULFIXSAT here yet, not sure exactly
1055 // why. Maybe it results in worse codegen compared to the unroll for some
1056 // targets? This should probably be investigated. And if we still prefer to
1057 // unroll an explanation could be helpful.
1058 break;
1059 case ISD::SDIVFIX:
1060 case ISD::UDIVFIX:
1061 ExpandFixedPointDiv(Node, Results);
1062 return;
1063 case ISD::SDIVFIXSAT:
1064 case ISD::UDIVFIXSAT:
1065 break;
1066#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
1067 case ISD::STRICT_##DAGN:
1068#include "llvm/IR/ConstrainedOps.def"
1069 ExpandStrictFPOp(Node, Results);
1070 return;
1071 case ISD::VECREDUCE_ADD:
1072 case ISD::VECREDUCE_MUL:
1073 case ISD::VECREDUCE_AND:
1074 case ISD::VECREDUCE_OR:
1075 case ISD::VECREDUCE_XOR:
1086 Results.push_back(TLI.expandVecReduce(Node, DAG));
1087 return;
1090 Results.push_back(TLI.expandVecReduceSeq(Node, DAG));
1091 return;
1092 case ISD::SREM:
1093 case ISD::UREM:
1094 ExpandREM(Node, Results);
1095 return;
1096 case ISD::VP_MERGE:
1097 Results.push_back(ExpandVP_MERGE(Node));
1098 return;
1099 case ISD::FREM:
1100 if (tryExpandVecMathCall(Node, RTLIB::REM_F32, RTLIB::REM_F64,
1101 RTLIB::REM_F80, RTLIB::REM_F128,
1102 RTLIB::REM_PPCF128, Results))
1103 return;
1104
1105 break;
1106 }
1107
1108 SDValue Unrolled = DAG.UnrollVectorOp(Node);
1109 if (Node->getNumValues() == 1) {
1110 Results.push_back(Unrolled);
1111 } else {
1112 assert(Node->getNumValues() == Unrolled->getNumValues() &&
1113 "VectorLegalizer Expand returned wrong number of results!");
1114 for (unsigned I = 0, E = Unrolled->getNumValues(); I != E; ++I)
1115 Results.push_back(Unrolled.getValue(I));
1116 }
1117}
1118
1119SDValue VectorLegalizer::ExpandSELECT(SDNode *Node) {
1120 // Lower a select instruction where the condition is a scalar and the
1121 // operands are vectors. Lower this select to VSELECT and implement it
1122 // using XOR AND OR. The selector bit is broadcasted.
1123 EVT VT = Node->getValueType(0);
1124 SDLoc DL(Node);
1125
1126 SDValue Mask = Node->getOperand(0);
1127 SDValue Op1 = Node->getOperand(1);
1128 SDValue Op2 = Node->getOperand(2);
1129
1130 assert(VT.isVector() && !Mask.getValueType().isVector()
1131 && Op1.getValueType() == Op2.getValueType() && "Invalid type");
1132
1133 // If we can't even use the basic vector operations of
1134 // AND,OR,XOR, we will have to scalarize the op.
1135 // Notice that the operation may be 'promoted' which means that it is
1136 // 'bitcasted' to another type which is handled.
1137 // Also, we need to be able to construct a splat vector using either
1138 // BUILD_VECTOR or SPLAT_VECTOR.
1139 // FIXME: Should we also permit fixed-length SPLAT_VECTOR as a fallback to
1140 // BUILD_VECTOR?
1141 if (TLI.getOperationAction(ISD::AND, VT) == TargetLowering::Expand ||
1142 TLI.getOperationAction(ISD::XOR, VT) == TargetLowering::Expand ||
1143 TLI.getOperationAction(ISD::OR, VT) == TargetLowering::Expand ||
1144 TLI.getOperationAction(VT.isFixedLengthVector() ? ISD::BUILD_VECTOR
1146 VT) == TargetLowering::Expand)
1147 return DAG.UnrollVectorOp(Node);
1148
1149 // Generate a mask operand.
1151
1152 // What is the size of each element in the vector mask.
1153 EVT BitTy = MaskTy.getScalarType();
1154
1155 Mask = DAG.getSelect(DL, BitTy, Mask, DAG.getAllOnesConstant(DL, BitTy),
1156 DAG.getConstant(0, DL, BitTy));
1157
1158 // Broadcast the mask so that the entire vector is all one or all zero.
1159 Mask = DAG.getSplat(MaskTy, DL, Mask);
1160
1161 // Bitcast the operands to be the same type as the mask.
1162 // This is needed when we select between FP types because
1163 // the mask is a vector of integers.
1164 Op1 = DAG.getNode(ISD::BITCAST, DL, MaskTy, Op1);
1165 Op2 = DAG.getNode(ISD::BITCAST, DL, MaskTy, Op2);
1166
1167 SDValue NotMask = DAG.getNOT(DL, Mask, MaskTy);
1168
1169 Op1 = DAG.getNode(ISD::AND, DL, MaskTy, Op1, Mask);
1170 Op2 = DAG.getNode(ISD::AND, DL, MaskTy, Op2, NotMask);
1171 SDValue Val = DAG.getNode(ISD::OR, DL, MaskTy, Op1, Op2);
1172 return DAG.getNode(ISD::BITCAST, DL, Node->getValueType(0), Val);
1173}
1174
1175SDValue VectorLegalizer::ExpandSEXTINREG(SDNode *Node) {
1176 EVT VT = Node->getValueType(0);
1177
1178 // Make sure that the SRA and SHL instructions are available.
1179 if (TLI.getOperationAction(ISD::SRA, VT) == TargetLowering::Expand ||
1180 TLI.getOperationAction(ISD::SHL, VT) == TargetLowering::Expand)
1181 return DAG.UnrollVectorOp(Node);
1182
1183 SDLoc DL(Node);
1184 EVT OrigTy = cast<VTSDNode>(Node->getOperand(1))->getVT();
1185
1186 unsigned BW = VT.getScalarSizeInBits();
1187 unsigned OrigBW = OrigTy.getScalarSizeInBits();
1188 SDValue ShiftSz = DAG.getConstant(BW - OrigBW, DL, VT);
1189
1190 SDValue Op = DAG.getNode(ISD::SHL, DL, VT, Node->getOperand(0), ShiftSz);
1191 return DAG.getNode(ISD::SRA, DL, VT, Op, ShiftSz);
1192}
1193
1194// Generically expand a vector anyext in register to a shuffle of the relevant
1195// lanes into the appropriate locations, with other lanes left undef.
1196SDValue VectorLegalizer::ExpandANY_EXTEND_VECTOR_INREG(SDNode *Node) {
1197 SDLoc DL(Node);
1198 EVT VT = Node->getValueType(0);
1199 int NumElements = VT.getVectorNumElements();
1200 SDValue Src = Node->getOperand(0);
1201 EVT SrcVT = Src.getValueType();
1202 int NumSrcElements = SrcVT.getVectorNumElements();
1203
1204 // *_EXTEND_VECTOR_INREG SrcVT can be smaller than VT - so insert the vector
1205 // into a larger vector type.
1206 if (SrcVT.bitsLE(VT)) {
1207 assert((VT.getSizeInBits() % SrcVT.getScalarSizeInBits()) == 0 &&
1208 "ANY_EXTEND_VECTOR_INREG vector size mismatch");
1209 NumSrcElements = VT.getSizeInBits() / SrcVT.getScalarSizeInBits();
1210 SrcVT = EVT::getVectorVT(*DAG.getContext(), SrcVT.getScalarType(),
1211 NumSrcElements);
1212 Src = DAG.getNode(ISD::INSERT_SUBVECTOR, DL, SrcVT, DAG.getUNDEF(SrcVT),
1213 Src, DAG.getVectorIdxConstant(0, DL));
1214 }
1215
1216 // Build a base mask of undef shuffles.
1217 SmallVector<int, 16> ShuffleMask;
1218 ShuffleMask.resize(NumSrcElements, -1);
1219
1220 // Place the extended lanes into the correct locations.
1221 int ExtLaneScale = NumSrcElements / NumElements;
1222 int EndianOffset = DAG.getDataLayout().isBigEndian() ? ExtLaneScale - 1 : 0;
1223 for (int i = 0; i < NumElements; ++i)
1224 ShuffleMask[i * ExtLaneScale + EndianOffset] = i;
1225
1226 return DAG.getNode(
1227 ISD::BITCAST, DL, VT,
1228 DAG.getVectorShuffle(SrcVT, DL, Src, DAG.getUNDEF(SrcVT), ShuffleMask));
1229}
1230
1231SDValue VectorLegalizer::ExpandSIGN_EXTEND_VECTOR_INREG(SDNode *Node) {
1232 SDLoc DL(Node);
1233 EVT VT = Node->getValueType(0);
1234 SDValue Src = Node->getOperand(0);
1235 EVT SrcVT = Src.getValueType();
1236
1237 // First build an any-extend node which can be legalized above when we
1238 // recurse through it.
1239 SDValue Op = DAG.getNode(ISD::ANY_EXTEND_VECTOR_INREG, DL, VT, Src);
1240
1241 // Now we need sign extend. Do this by shifting the elements. Even if these
1242 // aren't legal operations, they have a better chance of being legalized
1243 // without full scalarization than the sign extension does.
1244 unsigned EltWidth = VT.getScalarSizeInBits();
1245 unsigned SrcEltWidth = SrcVT.getScalarSizeInBits();
1246 SDValue ShiftAmount = DAG.getConstant(EltWidth - SrcEltWidth, DL, VT);
1247 return DAG.getNode(ISD::SRA, DL, VT,
1248 DAG.getNode(ISD::SHL, DL, VT, Op, ShiftAmount),
1249 ShiftAmount);
1250}
1251
1252// Generically expand a vector zext in register to a shuffle of the relevant
1253// lanes into the appropriate locations, a blend of zero into the high bits,
1254// and a bitcast to the wider element type.
1255SDValue VectorLegalizer::ExpandZERO_EXTEND_VECTOR_INREG(SDNode *Node) {
1256 SDLoc DL(Node);
1257 EVT VT = Node->getValueType(0);
1258 int NumElements = VT.getVectorNumElements();
1259 SDValue Src = Node->getOperand(0);
1260 EVT SrcVT = Src.getValueType();
1261 int NumSrcElements = SrcVT.getVectorNumElements();
1262
1263 // *_EXTEND_VECTOR_INREG SrcVT can be smaller than VT - so insert the vector
1264 // into a larger vector type.
1265 if (SrcVT.bitsLE(VT)) {
1266 assert((VT.getSizeInBits() % SrcVT.getScalarSizeInBits()) == 0 &&
1267 "ZERO_EXTEND_VECTOR_INREG vector size mismatch");
1268 NumSrcElements = VT.getSizeInBits() / SrcVT.getScalarSizeInBits();
1269 SrcVT = EVT::getVectorVT(*DAG.getContext(), SrcVT.getScalarType(),
1270 NumSrcElements);
1271 Src = DAG.getNode(ISD::INSERT_SUBVECTOR, DL, SrcVT, DAG.getUNDEF(SrcVT),
1272 Src, DAG.getVectorIdxConstant(0, DL));
1273 }
1274
1275 // Build up a zero vector to blend into this one.
1276 SDValue Zero = DAG.getConstant(0, DL, SrcVT);
1277
1278 // Shuffle the incoming lanes into the correct position, and pull all other
1279 // lanes from the zero vector.
1280 auto ShuffleMask = llvm::to_vector<16>(llvm::seq<int>(0, NumSrcElements));
1281
1282 int ExtLaneScale = NumSrcElements / NumElements;
1283 int EndianOffset = DAG.getDataLayout().isBigEndian() ? ExtLaneScale - 1 : 0;
1284 for (int i = 0; i < NumElements; ++i)
1285 ShuffleMask[i * ExtLaneScale + EndianOffset] = NumSrcElements + i;
1286
1287 return DAG.getNode(ISD::BITCAST, DL, VT,
1288 DAG.getVectorShuffle(SrcVT, DL, Zero, Src, ShuffleMask));
1289}
1290
1291static void createBSWAPShuffleMask(EVT VT, SmallVectorImpl<int> &ShuffleMask) {
1292 int ScalarSizeInBytes = VT.getScalarSizeInBits() / 8;
1293 for (int I = 0, E = VT.getVectorNumElements(); I != E; ++I)
1294 for (int J = ScalarSizeInBytes - 1; J >= 0; --J)
1295 ShuffleMask.push_back((I * ScalarSizeInBytes) + J);
1296}
1297
1298SDValue VectorLegalizer::ExpandBSWAP(SDNode *Node) {
1299 EVT VT = Node->getValueType(0);
1300
1301 // Scalable vectors can't use shuffle expansion.
1302 if (VT.isScalableVector())
1303 return TLI.expandBSWAP(Node, DAG);
1304
1305 // Generate a byte wise shuffle mask for the BSWAP.
1306 SmallVector<int, 16> ShuffleMask;
1307 createBSWAPShuffleMask(VT, ShuffleMask);
1308 EVT ByteVT = EVT::getVectorVT(*DAG.getContext(), MVT::i8, ShuffleMask.size());
1309
1310 // Only emit a shuffle if the mask is legal.
1311 if (TLI.isShuffleMaskLegal(ShuffleMask, ByteVT)) {
1312 SDLoc DL(Node);
1313 SDValue Op = DAG.getNode(ISD::BITCAST, DL, ByteVT, Node->getOperand(0));
1314 Op = DAG.getVectorShuffle(ByteVT, DL, Op, DAG.getUNDEF(ByteVT), ShuffleMask);
1315 return DAG.getNode(ISD::BITCAST, DL, VT, Op);
1316 }
1317
1318 // If we have the appropriate vector bit operations, it is better to use them
1319 // than unrolling and expanding each component.
1320 if (TLI.isOperationLegalOrCustom(ISD::SHL, VT) &&
1321 TLI.isOperationLegalOrCustom(ISD::SRL, VT) &&
1322 TLI.isOperationLegalOrCustomOrPromote(ISD::AND, VT) &&
1323 TLI.isOperationLegalOrCustomOrPromote(ISD::OR, VT))
1324 return TLI.expandBSWAP(Node, DAG);
1325
1326 // Otherwise unroll.
1327 return DAG.UnrollVectorOp(Node);
1328}
1329
1330void VectorLegalizer::ExpandBITREVERSE(SDNode *Node,
1332 EVT VT = Node->getValueType(0);
1333
1334 // We can't unroll or use shuffles for scalable vectors.
1335 if (VT.isScalableVector()) {
1336 Results.push_back(TLI.expandBITREVERSE(Node, DAG));
1337 return;
1338 }
1339
1340 // If we have the scalar operation, it's probably cheaper to unroll it.
1341 if (TLI.isOperationLegalOrCustom(ISD::BITREVERSE, VT.getScalarType())) {
1342 SDValue Tmp = DAG.UnrollVectorOp(Node);
1343 Results.push_back(Tmp);
1344 return;
1345 }
1346
1347 // If the vector element width is a whole number of bytes, test if its legal
1348 // to BSWAP shuffle the bytes and then perform the BITREVERSE on the byte
1349 // vector. This greatly reduces the number of bit shifts necessary.
1350 unsigned ScalarSizeInBits = VT.getScalarSizeInBits();
1351 if (ScalarSizeInBits > 8 && (ScalarSizeInBits % 8) == 0) {
1352 SmallVector<int, 16> BSWAPMask;
1353 createBSWAPShuffleMask(VT, BSWAPMask);
1354
1355 EVT ByteVT = EVT::getVectorVT(*DAG.getContext(), MVT::i8, BSWAPMask.size());
1356 if (TLI.isShuffleMaskLegal(BSWAPMask, ByteVT) &&
1357 (TLI.isOperationLegalOrCustom(ISD::BITREVERSE, ByteVT) ||
1358 (TLI.isOperationLegalOrCustom(ISD::SHL, ByteVT) &&
1359 TLI.isOperationLegalOrCustom(ISD::SRL, ByteVT) &&
1360 TLI.isOperationLegalOrCustomOrPromote(ISD::AND, ByteVT) &&
1361 TLI.isOperationLegalOrCustomOrPromote(ISD::OR, ByteVT)))) {
1362 SDLoc DL(Node);
1363 SDValue Op = DAG.getNode(ISD::BITCAST, DL, ByteVT, Node->getOperand(0));
1364 Op = DAG.getVectorShuffle(ByteVT, DL, Op, DAG.getUNDEF(ByteVT),
1365 BSWAPMask);
1366 Op = DAG.getNode(ISD::BITREVERSE, DL, ByteVT, Op);
1367 Op = DAG.getNode(ISD::BITCAST, DL, VT, Op);
1368 Results.push_back(Op);
1369 return;
1370 }
1371 }
1372
1373 // If we have the appropriate vector bit operations, it is better to use them
1374 // than unrolling and expanding each component.
1375 if (TLI.isOperationLegalOrCustom(ISD::SHL, VT) &&
1376 TLI.isOperationLegalOrCustom(ISD::SRL, VT) &&
1377 TLI.isOperationLegalOrCustomOrPromote(ISD::AND, VT) &&
1378 TLI.isOperationLegalOrCustomOrPromote(ISD::OR, VT)) {
1379 Results.push_back(TLI.expandBITREVERSE(Node, DAG));
1380 return;
1381 }
1382
1383 // Otherwise unroll.
1384 SDValue Tmp = DAG.UnrollVectorOp(Node);
1385 Results.push_back(Tmp);
1386}
1387
1388SDValue VectorLegalizer::ExpandVSELECT(SDNode *Node) {
1389 // Implement VSELECT in terms of XOR, AND, OR
1390 // on platforms which do not support blend natively.
1391 SDLoc DL(Node);
1392
1393 SDValue Mask = Node->getOperand(0);
1394 SDValue Op1 = Node->getOperand(1);
1395 SDValue Op2 = Node->getOperand(2);
1396
1397 EVT VT = Mask.getValueType();
1398
1399 // If we can't even use the basic vector operations of
1400 // AND,OR,XOR, we will have to scalarize the op.
1401 // Notice that the operation may be 'promoted' which means that it is
1402 // 'bitcasted' to another type which is handled.
1403 if (TLI.getOperationAction(ISD::AND, VT) == TargetLowering::Expand ||
1404 TLI.getOperationAction(ISD::XOR, VT) == TargetLowering::Expand ||
1405 TLI.getOperationAction(ISD::OR, VT) == TargetLowering::Expand)
1406 return DAG.UnrollVectorOp(Node);
1407
1408 // This operation also isn't safe with AND, OR, XOR when the boolean type is
1409 // 0/1 and the select operands aren't also booleans, as we need an all-ones
1410 // vector constant to mask with.
1411 // FIXME: Sign extend 1 to all ones if that's legal on the target.
1412 auto BoolContents = TLI.getBooleanContents(Op1.getValueType());
1413 if (BoolContents != TargetLowering::ZeroOrNegativeOneBooleanContent &&
1414 !(BoolContents == TargetLowering::ZeroOrOneBooleanContent &&
1415 Op1.getValueType().getVectorElementType() == MVT::i1))
1416 return DAG.UnrollVectorOp(Node);
1417
1418 // If the mask and the type are different sizes, unroll the vector op. This
1419 // can occur when getSetCCResultType returns something that is different in
1420 // size from the operand types. For example, v4i8 = select v4i32, v4i8, v4i8.
1421 if (VT.getSizeInBits() != Op1.getValueSizeInBits())
1422 return DAG.UnrollVectorOp(Node);
1423
1424 // Bitcast the operands to be the same type as the mask.
1425 // This is needed when we select between FP types because
1426 // the mask is a vector of integers.
1427 Op1 = DAG.getNode(ISD::BITCAST, DL, VT, Op1);
1428 Op2 = DAG.getNode(ISD::BITCAST, DL, VT, Op2);
1429
1430 SDValue NotMask = DAG.getNOT(DL, Mask, VT);
1431
1432 Op1 = DAG.getNode(ISD::AND, DL, VT, Op1, Mask);
1433 Op2 = DAG.getNode(ISD::AND, DL, VT, Op2, NotMask);
1434 SDValue Val = DAG.getNode(ISD::OR, DL, VT, Op1, Op2);
1435 return DAG.getNode(ISD::BITCAST, DL, Node->getValueType(0), Val);
1436}
1437
1438SDValue VectorLegalizer::ExpandVP_SELECT(SDNode *Node) {
1439 // Implement VP_SELECT in terms of VP_XOR, VP_AND and VP_OR on platforms which
1440 // do not support it natively.
1441 SDLoc DL(Node);
1442
1443 SDValue Mask = Node->getOperand(0);
1444 SDValue Op1 = Node->getOperand(1);
1445 SDValue Op2 = Node->getOperand(2);
1446 SDValue EVL = Node->getOperand(3);
1447
1448 EVT VT = Mask.getValueType();
1449
1450 // If we can't even use the basic vector operations of
1451 // VP_AND,VP_OR,VP_XOR, we will have to scalarize the op.
1452 if (TLI.getOperationAction(ISD::VP_AND, VT) == TargetLowering::Expand ||
1453 TLI.getOperationAction(ISD::VP_XOR, VT) == TargetLowering::Expand ||
1454 TLI.getOperationAction(ISD::VP_OR, VT) == TargetLowering::Expand)
1455 return DAG.UnrollVectorOp(Node);
1456
1457 // This operation also isn't safe when the operands aren't also booleans.
1458 if (Op1.getValueType().getVectorElementType() != MVT::i1)
1459 return DAG.UnrollVectorOp(Node);
1460
1461 SDValue Ones = DAG.getAllOnesConstant(DL, VT);
1462 SDValue NotMask = DAG.getNode(ISD::VP_XOR, DL, VT, Mask, Ones, Ones, EVL);
1463
1464 Op1 = DAG.getNode(ISD::VP_AND, DL, VT, Op1, Mask, Ones, EVL);
1465 Op2 = DAG.getNode(ISD::VP_AND, DL, VT, Op2, NotMask, Ones, EVL);
1466 return DAG.getNode(ISD::VP_OR, DL, VT, Op1, Op2, Ones, EVL);
1467}
1468
1469SDValue VectorLegalizer::ExpandVP_MERGE(SDNode *Node) {
1470 // Implement VP_MERGE in terms of VSELECT. Construct a mask where vector
1471 // indices less than the EVL/pivot are true. Combine that with the original
1472 // mask for a full-length mask. Use a full-length VSELECT to select between
1473 // the true and false values.
1474 SDLoc DL(Node);
1475
1476 SDValue Mask = Node->getOperand(0);
1477 SDValue Op1 = Node->getOperand(1);
1478 SDValue Op2 = Node->getOperand(2);
1479 SDValue EVL = Node->getOperand(3);
1480
1481 EVT MaskVT = Mask.getValueType();
1482 bool IsFixedLen = MaskVT.isFixedLengthVector();
1483
1484 EVT EVLVecVT = EVT::getVectorVT(*DAG.getContext(), EVL.getValueType(),
1485 MaskVT.getVectorElementCount());
1486
1487 // If we can't construct the EVL mask efficiently, it's better to unroll.
1488 if ((IsFixedLen &&
1489 !TLI.isOperationLegalOrCustom(ISD::BUILD_VECTOR, EVLVecVT)) ||
1490 (!IsFixedLen &&
1491 (!TLI.isOperationLegalOrCustom(ISD::STEP_VECTOR, EVLVecVT) ||
1492 !TLI.isOperationLegalOrCustom(ISD::SPLAT_VECTOR, EVLVecVT))))
1493 return DAG.UnrollVectorOp(Node);
1494
1495 // If using a SETCC would result in a different type than the mask type,
1496 // unroll.
1497 if (TLI.getSetCCResultType(DAG.getDataLayout(), *DAG.getContext(),
1498 EVLVecVT) != MaskVT)
1499 return DAG.UnrollVectorOp(Node);
1500
1501 SDValue StepVec = DAG.getStepVector(DL, EVLVecVT);
1502 SDValue SplatEVL = DAG.getSplat(EVLVecVT, DL, EVL);
1503 SDValue EVLMask =
1504 DAG.getSetCC(DL, MaskVT, StepVec, SplatEVL, ISD::CondCode::SETULT);
1505
1506 SDValue FullMask = DAG.getNode(ISD::AND, DL, MaskVT, Mask, EVLMask);
1507 return DAG.getSelect(DL, Node->getValueType(0), FullMask, Op1, Op2);
1508}
1509
1510SDValue VectorLegalizer::ExpandVP_REM(SDNode *Node) {
1511 // Implement VP_SREM/UREM in terms of VP_SDIV/VP_UDIV, VP_MUL, VP_SUB.
1512 EVT VT = Node->getValueType(0);
1513
1514 unsigned DivOpc = Node->getOpcode() == ISD::VP_SREM ? ISD::VP_SDIV : ISD::VP_UDIV;
1515
1516 if (!TLI.isOperationLegalOrCustom(DivOpc, VT) ||
1517 !TLI.isOperationLegalOrCustom(ISD::VP_MUL, VT) ||
1518 !TLI.isOperationLegalOrCustom(ISD::VP_SUB, VT))
1519 return SDValue();
1520
1521 SDLoc DL(Node);
1522
1523 SDValue Dividend = Node->getOperand(0);
1524 SDValue Divisor = Node->getOperand(1);
1525 SDValue Mask = Node->getOperand(2);
1526 SDValue EVL = Node->getOperand(3);
1527
1528 // X % Y -> X-X/Y*Y
1529 SDValue Div = DAG.getNode(DivOpc, DL, VT, Dividend, Divisor, Mask, EVL);
1530 SDValue Mul = DAG.getNode(ISD::VP_MUL, DL, VT, Divisor, Div, Mask, EVL);
1531 return DAG.getNode(ISD::VP_SUB, DL, VT, Dividend, Mul, Mask, EVL);
1532}
1533
1534void VectorLegalizer::ExpandFP_TO_UINT(SDNode *Node,
1536 // Attempt to expand using TargetLowering.
1537 SDValue Result, Chain;
1538 if (TLI.expandFP_TO_UINT(Node, Result, Chain, DAG)) {
1539 Results.push_back(Result);
1540 if (Node->isStrictFPOpcode())
1541 Results.push_back(Chain);
1542 return;
1543 }
1544
1545 // Otherwise go ahead and unroll.
1546 if (Node->isStrictFPOpcode()) {
1547 UnrollStrictFPOp(Node, Results);
1548 return;
1549 }
1550
1551 Results.push_back(DAG.UnrollVectorOp(Node));
1552}
1553
1554void VectorLegalizer::ExpandUINT_TO_FLOAT(SDNode *Node,
1556 bool IsStrict = Node->isStrictFPOpcode();
1557 unsigned OpNo = IsStrict ? 1 : 0;
1558 SDValue Src = Node->getOperand(OpNo);
1559 EVT VT = Src.getValueType();
1560 SDLoc DL(Node);
1561
1562 // Attempt to expand using TargetLowering.
1564 SDValue Chain;
1565 if (TLI.expandUINT_TO_FP(Node, Result, Chain, DAG)) {
1566 Results.push_back(Result);
1567 if (IsStrict)
1568 Results.push_back(Chain);
1569 return;
1570 }
1571
1572 // Make sure that the SINT_TO_FP and SRL instructions are available.
1573 if (((!IsStrict && TLI.getOperationAction(ISD::SINT_TO_FP, VT) ==
1574 TargetLowering::Expand) ||
1575 (IsStrict && TLI.getOperationAction(ISD::STRICT_SINT_TO_FP, VT) ==
1576 TargetLowering::Expand)) ||
1577 TLI.getOperationAction(ISD::SRL, VT) == TargetLowering::Expand) {
1578 if (IsStrict) {
1579 UnrollStrictFPOp(Node, Results);
1580 return;
1581 }
1582
1583 Results.push_back(DAG.UnrollVectorOp(Node));
1584 return;
1585 }
1586
1587 unsigned BW = VT.getScalarSizeInBits();
1588 assert((BW == 64 || BW == 32) &&
1589 "Elements in vector-UINT_TO_FP must be 32 or 64 bits wide");
1590
1591 SDValue HalfWord = DAG.getConstant(BW / 2, DL, VT);
1592
1593 // Constants to clear the upper part of the word.
1594 // Notice that we can also use SHL+SHR, but using a constant is slightly
1595 // faster on x86.
1596 uint64_t HWMask = (BW == 64) ? 0x00000000FFFFFFFF : 0x0000FFFF;
1597 SDValue HalfWordMask = DAG.getConstant(HWMask, DL, VT);
1598
1599 // Two to the power of half-word-size.
1600 SDValue TWOHW =
1601 DAG.getConstantFP(1ULL << (BW / 2), DL, Node->getValueType(0));
1602
1603 // Clear upper part of LO, lower HI
1604 SDValue HI = DAG.getNode(ISD::SRL, DL, VT, Src, HalfWord);
1605 SDValue LO = DAG.getNode(ISD::AND, DL, VT, Src, HalfWordMask);
1606
1607 if (IsStrict) {
1608 // Convert hi and lo to floats
1609 // Convert the hi part back to the upper values
1610 // TODO: Can any fast-math-flags be set on these nodes?
1612 {Node->getValueType(0), MVT::Other},
1613 {Node->getOperand(0), HI});
1614 fHI = DAG.getNode(ISD::STRICT_FMUL, DL, {Node->getValueType(0), MVT::Other},
1615 {fHI.getValue(1), fHI, TWOHW});
1617 {Node->getValueType(0), MVT::Other},
1618 {Node->getOperand(0), LO});
1619
1620 SDValue TF = DAG.getNode(ISD::TokenFactor, DL, MVT::Other, fHI.getValue(1),
1621 fLO.getValue(1));
1622
1623 // Add the two halves
1624 SDValue Result =
1625 DAG.getNode(ISD::STRICT_FADD, DL, {Node->getValueType(0), MVT::Other},
1626 {TF, fHI, fLO});
1627
1628 Results.push_back(Result);
1629 Results.push_back(Result.getValue(1));
1630 return;
1631 }
1632
1633 // Convert hi and lo to floats
1634 // Convert the hi part back to the upper values
1635 // TODO: Can any fast-math-flags be set on these nodes?
1636 SDValue fHI = DAG.getNode(ISD::SINT_TO_FP, DL, Node->getValueType(0), HI);
1637 fHI = DAG.getNode(ISD::FMUL, DL, Node->getValueType(0), fHI, TWOHW);
1638 SDValue fLO = DAG.getNode(ISD::SINT_TO_FP, DL, Node->getValueType(0), LO);
1639
1640 // Add the two halves
1641 Results.push_back(
1642 DAG.getNode(ISD::FADD, DL, Node->getValueType(0), fHI, fLO));
1643}
1644
1645SDValue VectorLegalizer::ExpandFNEG(SDNode *Node) {
1646 if (TLI.isOperationLegalOrCustom(ISD::FSUB, Node->getValueType(0))) {
1647 SDLoc DL(Node);
1648 SDValue Zero = DAG.getConstantFP(-0.0, DL, Node->getValueType(0));
1649 // TODO: If FNEG had fast-math-flags, they'd get propagated to this FSUB.
1650 return DAG.getNode(ISD::FSUB, DL, Node->getValueType(0), Zero,
1651 Node->getOperand(0));
1652 }
1653 return DAG.UnrollVectorOp(Node);
1654}
1655
1656void VectorLegalizer::ExpandFSUB(SDNode *Node,
1658 // For floating-point values, (a-b) is the same as a+(-b). If FNEG is legal,
1659 // we can defer this to operation legalization where it will be lowered as
1660 // a+(-b).
1661 EVT VT = Node->getValueType(0);
1662 if (TLI.isOperationLegalOrCustom(ISD::FNEG, VT) &&
1663 TLI.isOperationLegalOrCustom(ISD::FADD, VT))
1664 return; // Defer to LegalizeDAG
1665
1666 SDValue Tmp = DAG.UnrollVectorOp(Node);
1667 Results.push_back(Tmp);
1668}
1669
1670void VectorLegalizer::ExpandSETCC(SDNode *Node,
1672 bool NeedInvert = false;
1673 bool IsVP = Node->getOpcode() == ISD::VP_SETCC;
1674 bool IsStrict = Node->getOpcode() == ISD::STRICT_FSETCC ||
1675 Node->getOpcode() == ISD::STRICT_FSETCCS;
1676 bool IsSignaling = Node->getOpcode() == ISD::STRICT_FSETCCS;
1677 unsigned Offset = IsStrict ? 1 : 0;
1678
1679 SDValue Chain = IsStrict ? Node->getOperand(0) : SDValue();
1680 SDValue LHS = Node->getOperand(0 + Offset);
1681 SDValue RHS = Node->getOperand(1 + Offset);
1682 SDValue CC = Node->getOperand(2 + Offset);
1683
1684 MVT OpVT = LHS.getSimpleValueType();
1685 ISD::CondCode CCCode = cast<CondCodeSDNode>(CC)->get();
1686
1687 if (TLI.getCondCodeAction(CCCode, OpVT) != TargetLowering::Expand) {
1688 if (IsStrict) {
1689 UnrollStrictFPOp(Node, Results);
1690 return;
1691 }
1692 Results.push_back(UnrollVSETCC(Node));
1693 return;
1694 }
1695
1696 SDValue Mask, EVL;
1697 if (IsVP) {
1698 Mask = Node->getOperand(3 + Offset);
1699 EVL = Node->getOperand(4 + Offset);
1700 }
1701
1702 SDLoc dl(Node);
1703 bool Legalized =
1704 TLI.LegalizeSetCCCondCode(DAG, Node->getValueType(0), LHS, RHS, CC, Mask,
1705 EVL, NeedInvert, dl, Chain, IsSignaling);
1706
1707 if (Legalized) {
1708 // If we expanded the SETCC by swapping LHS and RHS, or by inverting the
1709 // condition code, create a new SETCC node.
1710 if (CC.getNode()) {
1711 if (IsStrict) {
1712 LHS = DAG.getNode(Node->getOpcode(), dl, Node->getVTList(),
1713 {Chain, LHS, RHS, CC}, Node->getFlags());
1714 Chain = LHS.getValue(1);
1715 } else if (IsVP) {
1716 LHS = DAG.getNode(ISD::VP_SETCC, dl, Node->getValueType(0),
1717 {LHS, RHS, CC, Mask, EVL}, Node->getFlags());
1718 } else {
1719 LHS = DAG.getNode(ISD::SETCC, dl, Node->getValueType(0), LHS, RHS, CC,
1720 Node->getFlags());
1721 }
1722 }
1723
1724 // If we expanded the SETCC by inverting the condition code, then wrap
1725 // the existing SETCC in a NOT to restore the intended condition.
1726 if (NeedInvert) {
1727 if (!IsVP)
1728 LHS = DAG.getLogicalNOT(dl, LHS, LHS->getValueType(0));
1729 else
1730 LHS = DAG.getVPLogicalNOT(dl, LHS, Mask, EVL, LHS->getValueType(0));
1731 }
1732 } else {
1733 assert(!IsStrict && "Don't know how to expand for strict nodes.");
1734
1735 // Otherwise, SETCC for the given comparison type must be completely
1736 // illegal; expand it into a SELECT_CC.
1737 EVT VT = Node->getValueType(0);
1738 LHS =
1739 DAG.getNode(ISD::SELECT_CC, dl, VT, LHS, RHS,
1740 DAG.getBoolConstant(true, dl, VT, LHS.getValueType()),
1741 DAG.getBoolConstant(false, dl, VT, LHS.getValueType()), CC);
1742 LHS->setFlags(Node->getFlags());
1743 }
1744
1745 Results.push_back(LHS);
1746 if (IsStrict)
1747 Results.push_back(Chain);
1748}
1749
1750void VectorLegalizer::ExpandUADDSUBO(SDNode *Node,
1752 SDValue Result, Overflow;
1753 TLI.expandUADDSUBO(Node, Result, Overflow, DAG);
1754 Results.push_back(Result);
1755 Results.push_back(Overflow);
1756}
1757
1758void VectorLegalizer::ExpandSADDSUBO(SDNode *Node,
1760 SDValue Result, Overflow;
1761 TLI.expandSADDSUBO(Node, Result, Overflow, DAG);
1762 Results.push_back(Result);
1763 Results.push_back(Overflow);
1764}
1765
1766void VectorLegalizer::ExpandMULO(SDNode *Node,
1768 SDValue Result, Overflow;
1769 if (!TLI.expandMULO(Node, Result, Overflow, DAG))
1770 std::tie(Result, Overflow) = DAG.UnrollVectorOverflowOp(Node);
1771
1772 Results.push_back(Result);
1773 Results.push_back(Overflow);
1774}
1775
1776void VectorLegalizer::ExpandFixedPointDiv(SDNode *Node,
1778 SDNode *N = Node;
1779 if (SDValue Expanded = TLI.expandFixedPointDiv(N->getOpcode(), SDLoc(N),
1780 N->getOperand(0), N->getOperand(1), N->getConstantOperandVal(2), DAG))
1781 Results.push_back(Expanded);
1782}
1783
1784void VectorLegalizer::ExpandStrictFPOp(SDNode *Node,
1786 if (Node->getOpcode() == ISD::STRICT_UINT_TO_FP) {
1787 ExpandUINT_TO_FLOAT(Node, Results);
1788 return;
1789 }
1790 if (Node->getOpcode() == ISD::STRICT_FP_TO_UINT) {
1791 ExpandFP_TO_UINT(Node, Results);
1792 return;
1793 }
1794
1795 if (Node->getOpcode() == ISD::STRICT_FSETCC ||
1796 Node->getOpcode() == ISD::STRICT_FSETCCS) {
1797 ExpandSETCC(Node, Results);
1798 return;
1799 }
1800
1801 UnrollStrictFPOp(Node, Results);
1802}
1803
1804void VectorLegalizer::ExpandREM(SDNode *Node,
1806 assert((Node->getOpcode() == ISD::SREM || Node->getOpcode() == ISD::UREM) &&
1807 "Expected REM node");
1808
1810 if (!TLI.expandREM(Node, Result, DAG))
1811 Result = DAG.UnrollVectorOp(Node);
1812 Results.push_back(Result);
1813}
1814
1815// Try to expand libm nodes into vector math routine calls. Callers provide the
1816// LibFunc equivalent of the passed in Node, which is used to lookup mappings
1817// within TargetLibraryInfo. The only mappings considered are those where the
1818// result and all operands are the same vector type. While predicated nodes are
1819// not supported, we will emit calls to masked routines by passing in an all
1820// true mask.
1821bool VectorLegalizer::tryExpandVecMathCall(SDNode *Node, RTLIB::Libcall LC,
1823 // Chain must be propagated but currently strict fp operations are down
1824 // converted to their none strict counterpart.
1825 assert(!Node->isStrictFPOpcode() && "Unexpected strict fp operation!");
1826
1827 const char *LCName = TLI.getLibcallName(LC);
1828 if (!LCName)
1829 return false;
1830 LLVM_DEBUG(dbgs() << "Looking for vector variant of " << LCName << "\n");
1831
1832 EVT VT = Node->getValueType(0);
1834
1835 // Lookup a vector function equivalent to the specified libcall. Prefer
1836 // unmasked variants but we will generate a mask if need be.
1837 const TargetLibraryInfo &TLibInfo = DAG.getLibInfo();
1838 const VecDesc *VD = TLibInfo.getVectorMappingInfo(LCName, VL, false);
1839 if (!VD)
1840 VD = TLibInfo.getVectorMappingInfo(LCName, VL, /*Masked=*/true);
1841 if (!VD)
1842 return false;
1843
1844 LLVMContext *Ctx = DAG.getContext();
1845 Type *Ty = VT.getTypeForEVT(*Ctx);
1846 Type *ScalarTy = Ty->getScalarType();
1847
1848 // Construct a scalar function type based on Node's operands.
1850 for (unsigned i = 0; i < Node->getNumOperands(); ++i) {
1851 assert(Node->getOperand(i).getValueType() == VT &&
1852 "Expected matching vector types!");
1853 ArgTys.push_back(ScalarTy);
1854 }
1855 FunctionType *ScalarFTy = FunctionType::get(ScalarTy, ArgTys, false);
1856
1857 // Generate call information for the vector function.
1858 const std::string MangledName = VD->getVectorFunctionABIVariantString();
1859 auto OptVFInfo = VFABI::tryDemangleForVFABI(MangledName, ScalarFTy);
1860 if (!OptVFInfo)
1861 return false;
1862
1863 LLVM_DEBUG(dbgs() << "Found vector variant " << VD->getVectorFnName()
1864 << "\n");
1865
1866 // Sanity check just in case OptVFInfo has unexpected parameters.
1867 if (OptVFInfo->Shape.Parameters.size() !=
1868 Node->getNumOperands() + VD->isMasked())
1869 return false;
1870
1871 // Collect vector call operands.
1872
1873 SDLoc DL(Node);
1876 Entry.IsSExt = false;
1877 Entry.IsZExt = false;
1878
1879 unsigned OpNum = 0;
1880 for (auto &VFParam : OptVFInfo->Shape.Parameters) {
1881 if (VFParam.ParamKind == VFParamKind::GlobalPredicate) {
1882 EVT MaskVT = TLI.getSetCCResultType(DAG.getDataLayout(), *Ctx, VT);
1883 Entry.Node = DAG.getBoolConstant(true, DL, MaskVT, VT);
1884 Entry.Ty = MaskVT.getTypeForEVT(*Ctx);
1885 Args.push_back(Entry);
1886 continue;
1887 }
1888
1889 // Only vector operands are supported.
1890 if (VFParam.ParamKind != VFParamKind::Vector)
1891 return false;
1892
1893 Entry.Node = Node->getOperand(OpNum++);
1894 Entry.Ty = Ty;
1895 Args.push_back(Entry);
1896 }
1897
1898 // Emit a call to the vector function.
1899 SDValue Callee = DAG.getExternalSymbol(VD->getVectorFnName().data(),
1900 TLI.getPointerTy(DAG.getDataLayout()));
1902 CLI.setDebugLoc(DL)
1903 .setChain(DAG.getEntryNode())
1904 .setLibCallee(CallingConv::C, Ty, Callee, std::move(Args));
1905
1906 std::pair<SDValue, SDValue> CallResult = TLI.LowerCallTo(CLI);
1907 Results.push_back(CallResult.first);
1908 return true;
1909}
1910
1911/// Try to expand the node to a vector libcall based on the result type.
1912bool VectorLegalizer::tryExpandVecMathCall(
1913 SDNode *Node, RTLIB::Libcall Call_F32, RTLIB::Libcall Call_F64,
1914 RTLIB::Libcall Call_F80, RTLIB::Libcall Call_F128,
1917 Node->getValueType(0).getVectorElementType(), Call_F32, Call_F64,
1918 Call_F80, Call_F128, Call_PPCF128);
1919
1920 if (LC == RTLIB::UNKNOWN_LIBCALL)
1921 return false;
1922
1923 return tryExpandVecMathCall(Node, LC, Results);
1924}
1925
1926void VectorLegalizer::UnrollStrictFPOp(SDNode *Node,
1928 EVT VT = Node->getValueType(0);
1929 EVT EltVT = VT.getVectorElementType();
1930 unsigned NumElems = VT.getVectorNumElements();
1931 unsigned NumOpers = Node->getNumOperands();
1932 const TargetLowering &TLI = DAG.getTargetLoweringInfo();
1933
1934 EVT TmpEltVT = EltVT;
1935 if (Node->getOpcode() == ISD::STRICT_FSETCC ||
1936 Node->getOpcode() == ISD::STRICT_FSETCCS)
1937 TmpEltVT = TLI.getSetCCResultType(DAG.getDataLayout(),
1938 *DAG.getContext(), TmpEltVT);
1939
1940 EVT ValueVTs[] = {TmpEltVT, MVT::Other};
1941 SDValue Chain = Node->getOperand(0);
1942 SDLoc dl(Node);
1943
1944 SmallVector<SDValue, 32> OpValues;
1945 SmallVector<SDValue, 32> OpChains;
1946 for (unsigned i = 0; i < NumElems; ++i) {
1948 SDValue Idx = DAG.getVectorIdxConstant(i, dl);
1949
1950 // The Chain is the first operand.
1951 Opers.push_back(Chain);
1952
1953 // Now process the remaining operands.
1954 for (unsigned j = 1; j < NumOpers; ++j) {
1955 SDValue Oper = Node->getOperand(j);
1956 EVT OperVT = Oper.getValueType();
1957
1958 if (OperVT.isVector())
1959 Oper = DAG.getNode(ISD::EXTRACT_VECTOR_ELT, dl,
1960 OperVT.getVectorElementType(), Oper, Idx);
1961
1962 Opers.push_back(Oper);
1963 }
1964
1965 SDValue ScalarOp = DAG.getNode(Node->getOpcode(), dl, ValueVTs, Opers);
1966 SDValue ScalarResult = ScalarOp.getValue(0);
1967 SDValue ScalarChain = ScalarOp.getValue(1);
1968
1969 if (Node->getOpcode() == ISD::STRICT_FSETCC ||
1970 Node->getOpcode() == ISD::STRICT_FSETCCS)
1971 ScalarResult = DAG.getSelect(dl, EltVT, ScalarResult,
1972 DAG.getAllOnesConstant(dl, EltVT),
1973 DAG.getConstant(0, dl, EltVT));
1974
1975 OpValues.push_back(ScalarResult);
1976 OpChains.push_back(ScalarChain);
1977 }
1978
1979 SDValue Result = DAG.getBuildVector(VT, dl, OpValues);
1980 SDValue NewChain = DAG.getNode(ISD::TokenFactor, dl, MVT::Other, OpChains);
1981
1982 Results.push_back(Result);
1983 Results.push_back(NewChain);
1984}
1985
1986SDValue VectorLegalizer::UnrollVSETCC(SDNode *Node) {
1987 EVT VT = Node->getValueType(0);
1988 unsigned NumElems = VT.getVectorNumElements();
1989 EVT EltVT = VT.getVectorElementType();
1990 SDValue LHS = Node->getOperand(0);
1991 SDValue RHS = Node->getOperand(1);
1992 SDValue CC = Node->getOperand(2);
1993 EVT TmpEltVT = LHS.getValueType().getVectorElementType();
1994 SDLoc dl(Node);
1995 SmallVector<SDValue, 8> Ops(NumElems);
1996 for (unsigned i = 0; i < NumElems; ++i) {
1997 SDValue LHSElem = DAG.getNode(ISD::EXTRACT_VECTOR_ELT, dl, TmpEltVT, LHS,
1998 DAG.getVectorIdxConstant(i, dl));
1999 SDValue RHSElem = DAG.getNode(ISD::EXTRACT_VECTOR_ELT, dl, TmpEltVT, RHS,
2000 DAG.getVectorIdxConstant(i, dl));
2001 Ops[i] = DAG.getNode(ISD::SETCC, dl,
2002 TLI.getSetCCResultType(DAG.getDataLayout(),
2003 *DAG.getContext(), TmpEltVT),
2004 LHSElem, RHSElem, CC);
2005 Ops[i] = DAG.getSelect(dl, EltVT, Ops[i], DAG.getAllOnesConstant(dl, EltVT),
2006 DAG.getConstant(0, dl, EltVT));
2007 }
2008 return DAG.getBuildVector(VT, dl, Ops);
2009}
2010
2012 return VectorLegalizer(*this).Run();
2013}
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis Results
BlockVerifier::State From
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
#define LLVM_DEBUG(X)
Definition: Debug.h:101
This file defines the DenseMap class.
static void createBSWAPShuffleMask(EVT VT, SmallVectorImpl< int > &ShuffleMask)
#define I(x, y, z)
Definition: MD5.cpp:58
mir Rename Register Operands
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
This file defines the SmallVector class.
This file describes how to lower LLVM code to machine code.
Value * RHS
Value * LHS
BinaryOperator * Mul
DEMANGLE_DUMP_METHOD void dump() const
This class represents an Operation in the Expression.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
Definition: DenseMap.h:220
size_t size() const
Definition: Function.h:808
This is an important class for using LLVM in a threaded context.
Definition: LLVMContext.h:67
This class is used to represent ISD::LOAD nodes.
Machine Value Type.
unsigned getVectorNumElements() const
bool isVector() const
Return true if this is a vector value type.
MVT getVectorElementType() const
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
MVT getScalarType() const
If this is a vector, return the element type, otherwise return this.
MutableArrayRef - Represent a mutable reference to an array (0 or more elements consecutively in memo...
Definition: ArrayRef.h:307
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
unsigned getNumValues() const
Return the number of values defined/returned by this operator.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
TypeSize getValueSizeInBits() const
Returns the size of the value in bits.
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
Definition: SelectionDAG.h:225
bool LegalizeVectors()
This transforms the SelectionDAG into a SelectionDAG that only uses vector math operations supported ...
const TargetLowering & getTargetLoweringInfo() const
Definition: SelectionDAG.h:478
ilist< SDNode >::iterator allnodes_iterator
Definition: SelectionDAG.h:534
bool empty() const
Definition: SmallVector.h:94
size_t size() const
Definition: SmallVector.h:91
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
Definition: SmallVector.h:586
void resize(size_type N)
Definition: SmallVector.h:651
void push_back(const T &Elt)
Definition: SmallVector.h:426
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1209
This class is used to represent ISD::STORE nodes.
constexpr const char * data() const
data - Get a pointer to the start of the string (which may not be null terminated).
Definition: StringRef.h:131
Provides information about what library functions are available for the current target.
const VecDesc * getVectorMappingInfo(StringRef F, const ElementCount &VF, bool Masked) const
LegalizeAction
This enum indicates whether operations are valid for a target, and if not, what action should be used...
virtual EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const
Return the ValueType of the result of SETCC operations.
std::vector< ArgListEntry > ArgListTy
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
The instances of the Type class are immutable: once they are created, they are never changed.
Definition: Type.h:45
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition: Type.h:348
Provides info so a possible vectorization of a function can be computed.
bool isMasked() const
std::string getVectorFunctionABIVariantString() const
Returns a vector function ABI variant string on the form: ZGV<isa><mask><vlen><vparams><scalarname>(<...
StringRef getVectorFnName() const
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
constexpr char Args[]
Key for Kernel::Metadata::mArgs.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
Definition: BitmaskEnum.h:121
@ C
The default llvm calling convention, compatible with C.
Definition: CallingConv.h:34
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
Definition: ISDOpcodes.h:751
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
Definition: ISDOpcodes.h:237
@ CTLZ_ZERO_UNDEF
Definition: ISDOpcodes.h:724
@ STRICT_FSETCC
STRICT_FSETCC/STRICT_FSETCCS - Constrained versions of SETCC, used for floating-point operands only.
Definition: ISDOpcodes.h:477
@ VECREDUCE_SEQ_FADD
Generic reduction nodes.
Definition: ISDOpcodes.h:1346
@ VECREDUCE_SMIN
Definition: ISDOpcodes.h:1377
@ SMUL_LOHI
SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing a signed/unsigned value of type i[2...
Definition: ISDOpcodes.h:251
@ INSERT_SUBVECTOR
INSERT_SUBVECTOR(VECTOR1, VECTOR2, IDX) - Returns a vector with VECTOR2 inserted into VECTOR1.
Definition: ISDOpcodes.h:560
@ BSWAP
Byte Swap and Counting operators.
Definition: ISDOpcodes.h:715
@ SMULFIX
RESULT = [US]MULFIX(LHS, RHS, SCALE) - Perform fixed point multiplication on 2 integers with the same...
Definition: ISDOpcodes.h:368
@ FMAXNUM_IEEE
Definition: ISDOpcodes.h:986
@ ADD
Simple integer binary arithmetic operators.
Definition: ISDOpcodes.h:240
@ LOAD
LOAD and STORE have token chains as their first operand, then the same operands as an LLVM load/store...
Definition: ISDOpcodes.h:1038
@ SMULFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
Definition: ISDOpcodes.h:374
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
Definition: ISDOpcodes.h:784
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
Definition: ISDOpcodes.h:484
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
Definition: ISDOpcodes.h:791
@ VECREDUCE_FMAX
FMIN/FMAX nodes can have flags, for NaN/NoNaN variants.
Definition: ISDOpcodes.h:1362
@ FADD
Simple binary floating point operators.
Definition: ISDOpcodes.h:391
@ VECREDUCE_FMAXIMUM
FMINIMUM/FMAXIMUM nodes propatate NaNs and signed zeroes using the llvm.minimum and llvm....
Definition: ISDOpcodes.h:1366
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
Definition: ISDOpcodes.h:689
@ SIGN_EXTEND_VECTOR_INREG
SIGN_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register sign-extension of the low ...
Definition: ISDOpcodes.h:821
@ SDIVREM
SDIVREM/UDIVREM - Divide two integers and produce both a quotient and remainder result.
Definition: ISDOpcodes.h:256
@ VECREDUCE_SMAX
Definition: ISDOpcodes.h:1376
@ STRICT_FSETCCS
Definition: ISDOpcodes.h:478
@ FPTRUNC_ROUND
Definition: ISDOpcodes.h:481
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
Definition: ISDOpcodes.h:904
@ FLDEXP
FLDEXP - ldexp, inspired by libm (op0 * 2**op1).
Definition: ISDOpcodes.h:940
@ SDIVFIX
RESULT = [US]DIVFIX(LHS, RHS, SCALE) - Perform fixed point division on 2 integers with the same width...
Definition: ISDOpcodes.h:381
@ STRICT_FSQRT
Constrained versions of libm-equivalent floating point intrinsics.
Definition: ISDOpcodes.h:412
@ SIGN_EXTEND
Conversion operators.
Definition: ISDOpcodes.h:775
@ STRICT_UINT_TO_FP
Definition: ISDOpcodes.h:451
@ VECREDUCE_FADD
These reductions have relaxed evaluation order semantics, and have a single vector operand.
Definition: ISDOpcodes.h:1359
@ CTTZ_ZERO_UNDEF
Bit counting operators with an undefined result for zero inputs.
Definition: ISDOpcodes.h:723
@ VECREDUCE_FMIN
Definition: ISDOpcodes.h:1363
@ FNEG
Perform various unary floating-point operations inspired by libm.
Definition: ISDOpcodes.h:931
@ SSUBO
Same for subtraction.
Definition: ISDOpcodes.h:328
@ STEP_VECTOR
STEP_VECTOR(IMM) - Returns a scalable vector whose lanes are comprised of a linear sequence of unsign...
Definition: ISDOpcodes.h:647
@ FCANONICALIZE
Returns platform specific canonical encoding of a floating point number.
Definition: ISDOpcodes.h:501
@ SSUBSAT
RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2 integers with the same bit width ...
Definition: ISDOpcodes.h:350
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
Definition: ISDOpcodes.h:728
@ VECREDUCE_UMAX
Definition: ISDOpcodes.h:1378
@ SPLAT_VECTOR
SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL duplicated in all lanes.
Definition: ISDOpcodes.h:628
@ SADDO
RESULT, BOOL = [SU]ADDO(LHS, RHS) - Overflow-aware nodes for addition.
Definition: ISDOpcodes.h:324
@ VECREDUCE_ADD
Integer reductions may have a result type larger than the vector element type.
Definition: ISDOpcodes.h:1371
@ MULHU
MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing an unsigned/signed value of...
Definition: ISDOpcodes.h:652
@ SHL
Shift and rotation operations.
Definition: ISDOpcodes.h:706
@ FMINNUM_IEEE
FMINNUM_IEEE/FMAXNUM_IEEE - Perform floating-point minimumNumber or maximumNumber on two values,...
Definition: ISDOpcodes.h:985
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
Definition: ISDOpcodes.h:536
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
Definition: ISDOpcodes.h:781
@ FP_TO_UINT_SAT
Definition: ISDOpcodes.h:857
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
Definition: ISDOpcodes.h:743
@ FMINNUM
FMINNUM/FMAXNUM - Perform floating-point minimum or maximum on two values.
Definition: ISDOpcodes.h:972
@ SSHLSAT
RESULT = [US]SHLSAT(LHS, RHS) - Perform saturation left shift.
Definition: ISDOpcodes.h:360
@ SMULO
Same for multiplication.
Definition: ISDOpcodes.h:332
@ ANY_EXTEND_VECTOR_INREG
ANY_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register any-extension of the low la...
Definition: ISDOpcodes.h:810
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
Definition: ISDOpcodes.h:799
@ SMIN
[US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned integers.
Definition: ISDOpcodes.h:675
@ SDIVFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
Definition: ISDOpcodes.h:387
@ FP_EXTEND
X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition: ISDOpcodes.h:889
@ VSELECT
Select with a vector condition (op #0) and two vector operands (ops #1 and #2), returning a vector re...
Definition: ISDOpcodes.h:737
@ STRICT_SINT_TO_FP
STRICT_[US]INT_TO_FP - Convert a signed or unsigned integer to a floating point value.
Definition: ISDOpcodes.h:450
@ VECREDUCE_UMIN
Definition: ISDOpcodes.h:1379
@ STRICT_FP_TO_UINT
Definition: ISDOpcodes.h:444
@ STRICT_FP_ROUND
X = STRICT_FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision ...
Definition: ISDOpcodes.h:466
@ STRICT_FP_TO_SINT
STRICT_FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
Definition: ISDOpcodes.h:443
@ FMINIMUM
FMINIMUM/FMAXIMUM - NaN-propagating minimum/maximum that also treat -0.0 as less than 0....
Definition: ISDOpcodes.h:991
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
Definition: ISDOpcodes.h:837
@ STRICT_FP_EXTEND
X = STRICT_FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
Definition: ISDOpcodes.h:471
@ AND
Bitwise operators - logical and, logical or, logical xor.
Definition: ISDOpcodes.h:681
@ VECREDUCE_FMUL
Definition: ISDOpcodes.h:1360
@ STRICT_FADD
Constrained versions of the binary floating point operators.
Definition: ISDOpcodes.h:401
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
Definition: ISDOpcodes.h:52
@ FFREXP
FFREXP - frexp, extract fractional and exponent component of a floating-point value.
Definition: ISDOpcodes.h:945
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
Definition: ISDOpcodes.h:870
@ ZERO_EXTEND_VECTOR_INREG
ZERO_EXTEND_VECTOR_INREG(Vector) - This operator represents an in-register zero-extension of the low ...
Definition: ISDOpcodes.h:832
@ FP_TO_SINT_SAT
FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a signed or unsigned scalar integer ...
Definition: ISDOpcodes.h:856
@ VECREDUCE_FMINIMUM
Definition: ISDOpcodes.h:1367
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
Definition: ISDOpcodes.h:787
@ VECREDUCE_SEQ_FMUL
Definition: ISDOpcodes.h:1347
@ AssertSext
AssertSext, AssertZext - These nodes record if a register contains a value that has already been zero...
Definition: ISDOpcodes.h:61
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
Definition: ISDOpcodes.h:494
@ SADDSAT
RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2 integers with the same bit width (W)...
Definition: ISDOpcodes.h:341
@ AssertZext
Definition: ISDOpcodes.h:62
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
Definition: ISDOpcodes.h:516
std::optional< unsigned > getVPMaskIdx(unsigned Opcode)
The operand position of the vector mask.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
Definition: ISDOpcodes.h:1535
LoadExtType
LoadExtType enum - This enum defines the three variants of LOADEXT (load with extension).
Definition: ISDOpcodes.h:1515
bool isVPOpcode(unsigned Opcode)
Whether this is a vector-predicated Opcode.
Libcall
RTLIB::Libcall enum - This enum defines all of the runtime library calls the backend can emit.
Libcall getFPLibCall(EVT VT, Libcall Call_F32, Libcall Call_F64, Libcall Call_F80, Libcall Call_F128, Libcall Call_PPCF128)
GetFPLibCall - Helper to return the right libcall for the given floating point type,...
ManagedStatic< cl::opt< FnT >, OptCreatorT > Action
std::optional< VFInfo > tryDemangleForVFABI(StringRef MangledName, const FunctionType *FTy)
Function to construct a VFInfo out of a mangled names in the following format:
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
@ Offset
Definition: DWP.cpp:456
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1729
raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition: Debug.cpp:163
#define N
Extended Value Type.
Definition: ValueTypes.h:34
EVT changeVectorElementTypeToInteger() const
Return a vector with the same number of elements as this vector, but with the element type converted ...
Definition: ValueTypes.h:93
static EVT getVectorVT(LLVMContext &Context, EVT VT, unsigned NumElements, bool IsScalable=false)
Returns the EVT that represents a vector NumElements in length, where each element is of type VT.
Definition: ValueTypes.h:73
ElementCount getVectorElementCount() const
Definition: ValueTypes.h:340
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
Definition: ValueTypes.h:358
uint64_t getScalarSizeInBits() const
Definition: ValueTypes.h:370
bool isFixedLengthVector() const
Definition: ValueTypes.h:177
bool isVector() const
Return true if this is a vector value type.
Definition: ValueTypes.h:167
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
Definition: ValueTypes.h:313
Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
Definition: ValueTypes.cpp:202
bool isScalableVector() const
Return true if this is a vector type where the runtime length is machine dependent.
Definition: ValueTypes.h:173
EVT getVectorElementType() const
Given a vector type, return the type of each element.
Definition: ValueTypes.h:318
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
Definition: ValueTypes.h:326
bool bitsLE(EVT VT) const
Return true if this has no more bits than VT.
Definition: ValueTypes.h:298
This represents a list of ValueType's that has been intern'd by a SelectionDAG.
This structure contains all information that is necessary for lowering calls.