LLVM 22.0.0git
RISCVISelLowering.h
Go to the documentation of this file.
1//===-- RISCVISelLowering.h - RISC-V DAG Lowering Interface -----*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file defines the interfaces that RISC-V uses to lower LLVM code into a
10// selection DAG.
11//
12//===----------------------------------------------------------------------===//
13
14#ifndef LLVM_LIB_TARGET_RISCV_RISCVISELLOWERING_H
15#define LLVM_LIB_TARGET_RISCV_RISCVISELLOWERING_H
16
17#include "RISCV.h"
18#include "RISCVCallingConv.h"
22#include <optional>
23
24namespace llvm {
25class InstructionCost;
26class RISCVSubtarget;
28
30 const RISCVSubtarget &Subtarget;
31
32public:
33 explicit RISCVTargetLowering(const TargetMachine &TM,
34 const RISCVSubtarget &STI);
35
36 const RISCVSubtarget &getSubtarget() const { return Subtarget; }
37
38 bool getTgtMemIntrinsic(IntrinsicInfo &Info, const CallBase &I,
40 unsigned Intrinsic) const override;
41 bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty,
42 unsigned AS,
43 Instruction *I = nullptr) const override;
44 bool isLegalICmpImmediate(int64_t Imm) const override;
45 bool isLegalAddImmediate(int64_t Imm) const override;
46 bool isTruncateFree(Type *SrcTy, Type *DstTy) const override;
47 bool isTruncateFree(EVT SrcVT, EVT DstVT) const override;
48 bool isTruncateFree(SDValue Val, EVT VT2) const override;
49 bool isZExtFree(SDValue Val, EVT VT2) const override;
50 bool isSExtCheaperThanZExt(EVT SrcVT, EVT DstVT) const override;
51 bool signExtendConstant(const ConstantInt *CI) const override;
52 bool isCheapToSpeculateCttz(Type *Ty) const override;
53 bool isCheapToSpeculateCtlz(Type *Ty) const override;
54 bool isMaskAndCmp0FoldingBeneficial(const Instruction &AndI) const override;
55 bool hasAndNotCompare(SDValue Y) const override;
56 bool hasAndNot(SDValue Y) const override;
57 bool hasBitTest(SDValue X, SDValue Y) const override;
60 unsigned OldShiftOpcode, unsigned NewShiftOpcode,
61 SelectionDAG &DAG) const override;
62 bool shouldScalarizeBinop(SDValue VecOp) const override;
63 bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override;
64 int getLegalZfaFPImm(const APFloat &Imm, EVT VT) const;
65 bool isFPImmLegal(const APFloat &Imm, EVT VT,
66 bool ForCodeSize) const override;
67 bool isExtractSubvectorCheap(EVT ResVT, EVT SrcVT,
68 unsigned Index) const override;
69
70 bool isIntDivCheap(EVT VT, AttributeList Attr) const override;
71
72 bool preferScalarizeSplat(SDNode *N) const override;
73
74 /// Customize the preferred legalization strategy for certain types.
75 LegalizeTypeAction getPreferredVectorAction(MVT VT) const override;
76
77 bool softPromoteHalfType() const override { return true; }
78
79 /// Return the register type for a given MVT, ensuring vectors are treated
80 /// as a series of gpr sized integers.
82 EVT VT) const override;
83
84 /// Return the number of registers for a given MVT, for inline assembly
85 unsigned
86 getNumRegisters(LLVMContext &Context, EVT VT,
87 std::optional<MVT> RegisterVT = std::nullopt) const override;
88
89 /// Return the number of registers for a given MVT, ensuring vectors are
90 /// treated as a series of gpr sized integers.
93 EVT VT) const override;
94
96 CallingConv::ID CC, EVT VT,
97 EVT &IntermediateVT,
98 unsigned &NumIntermediates,
99 MVT &RegisterVT) const override;
100
101 bool shouldFoldSelectWithIdentityConstant(unsigned BinOpcode, EVT VT,
102 unsigned SelectOpcode, SDValue X,
103 SDValue Y) const override;
104
105 /// Return true if the given shuffle mask can be codegen'd directly, or if it
106 /// should be stack expanded.
107 bool isShuffleMaskLegal(ArrayRef<int> M, EVT VT) const override;
108
109 bool isMultiStoresCheaperThanBitsMerge(EVT LTy, EVT HTy) const override {
110 // If the pair to store is a mixture of float and int values, we will
111 // save two bitwise instructions and one float-to-int instruction and
112 // increase one store instruction. There is potentially a more
113 // significant benefit because it avoids the float->int domain switch
114 // for input value. So It is more likely a win.
115 if ((LTy.isFloatingPoint() && HTy.isInteger()) ||
116 (LTy.isInteger() && HTy.isFloatingPoint()))
117 return true;
118 // If the pair only contains int values, we will save two bitwise
119 // instructions and increase one store instruction (costing one more
120 // store buffer). Since the benefit is more blurred we leave such a pair
121 // out until we get testcase to prove it is a win.
122 return false;
123 }
124
125 bool
127 unsigned DefinedValues) const override;
128
129 bool shouldExpandCttzElements(EVT VT) const override;
130
131 /// Return the cost of LMUL for linear operations.
133
138
139 // Provide custom lowering hooks for some operations.
140 SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override;
142 SelectionDAG &DAG) const override;
143
144 SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override;
145
147 const APInt &DemandedElts,
148 TargetLoweringOpt &TLO) const override;
149
151 KnownBits &Known,
152 const APInt &DemandedElts,
153 const SelectionDAG &DAG,
154 unsigned Depth) const override;
156 const APInt &DemandedElts,
157 const SelectionDAG &DAG,
158 unsigned Depth) const override;
159
161 const APInt &DemandedElts,
162 KnownBits &Known,
163 TargetLoweringOpt &TLO,
164 unsigned Depth) const override;
165
167 const APInt &DemandedElts,
168 const SelectionDAG &DAG,
169 bool PoisonOnly, bool ConsiderFlags,
170 unsigned Depth) const override;
171
172 const Constant *getTargetConstantFromLoad(LoadSDNode *LD) const override;
173
175 getTargetMMOFlags(const Instruction &I) const override;
176
178 getTargetMMOFlags(const MemSDNode &Node) const override;
179
180 bool
182 const MemSDNode &NodeY) const override;
183
184 ConstraintType getConstraintType(StringRef Constraint) const override;
185
187 getInlineAsmMemConstraint(StringRef ConstraintCode) const override;
188
189 std::pair<unsigned, const TargetRegisterClass *>
191 StringRef Constraint, MVT VT) const override;
192
194 std::vector<SDValue> &Ops,
195 SelectionDAG &DAG) const override;
196
199 MachineBasicBlock *BB) const override;
200
202 SDNode *Node) const override;
203
205 EVT VT) const override;
206
207 bool shouldFormOverflowOp(unsigned Opcode, EVT VT,
208 bool MathUsed) const override {
209 if (VT == MVT::i8 || VT == MVT::i16)
210 return false;
211
212 return TargetLowering::shouldFormOverflowOp(Opcode, VT, MathUsed);
213 }
214
215 bool storeOfVectorConstantIsCheap(bool IsZero, EVT MemVT, unsigned NumElem,
216 unsigned AddrSpace) const override {
217 // If we can replace 4 or more scalar stores, there will be a reduction
218 // in instructions even after we add a vector constant load.
219 return NumElem >= 4;
220 }
221
222 bool convertSetCCLogicToBitwiseLogic(EVT VT) const override {
223 return VT.isScalarInteger();
224 }
225 bool convertSelectOfConstantsToMath(EVT VT) const override { return true; }
226
227 bool isCtpopFast(EVT VT) const override;
228
229 unsigned getCustomCtpopCost(EVT VT, ISD::CondCode Cond) const override;
230
231 bool preferZeroCompareBranch() const override { return true; }
232
233 // Note that one specific case requires fence insertion for an
234 // AtomicCmpXchgInst but is handled via the RISCVZacasABIFix pass rather
235 // than this hook due to limitations in the interface here.
236 bool shouldInsertFencesForAtomic(const Instruction *I) const override;
237
239 AtomicOrdering Ord) const override;
241 AtomicOrdering Ord) const override;
242
244 EVT VT) const override;
245
247 return ISD::SIGN_EXTEND;
248 }
249
251 ISD::NodeType getExtendForAtomicRMWArg(unsigned Op) const override;
252
254 unsigned KeptBits) const override;
255
258 unsigned ExpansionFactor) const override {
262 ExpansionFactor);
263 }
264
266 CombineLevel Level) const override;
267
268 /// If a physical register, this returns the register that receives the
269 /// exception address on entry to an EH pad.
271 getExceptionPointerRegister(const Constant *PersonalityFn) const override;
272
273 /// If a physical register, this returns the register that receives the
274 /// exception typeid on entry to a landing pad.
276 getExceptionSelectorRegister(const Constant *PersonalityFn) const override;
277
278 bool shouldExtendTypeInLibCall(EVT Type) const override;
279 bool shouldSignExtendTypeInLibCall(Type *Ty, bool IsSigned) const override;
280
281 /// Returns the register with the specified architectural or ABI name. This
282 /// method is necessary to lower the llvm.read_register.* and
283 /// llvm.write_register.* intrinsics. Allocatable registers must be reserved
284 /// with the clang -ffixed-xX flag for access to be allowed.
285 Register getRegisterByName(const char *RegName, LLT VT,
286 const MachineFunction &MF) const override;
287
288 // Lower incoming arguments, copy physregs into vregs
290 bool IsVarArg,
292 const SDLoc &DL, SelectionDAG &DAG,
293 SmallVectorImpl<SDValue> &InVals) const override;
295 bool IsVarArg,
297 LLVMContext &Context, const Type *RetTy) const override;
298 SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg,
300 const SmallVectorImpl<SDValue> &OutVals, const SDLoc &DL,
301 SelectionDAG &DAG) const override;
303 SmallVectorImpl<SDValue> &InVals) const override;
304
306 Type *Ty) const override;
307 bool isUsedByReturnOnly(SDNode *N, SDValue &Chain) const override;
308 bool mayBeEmittedAsTailCall(const CallInst *CI) const override;
309 bool shouldConsiderGEPOffsetSplit() const override { return true; }
310
311 bool decomposeMulByConstant(LLVMContext &Context, EVT VT,
312 SDValue C) const override;
313
315 SDValue ConstNode) const override;
316
318 shouldExpandAtomicRMWInIR(AtomicRMWInst *AI) const override;
320 Value *AlignedAddr, Value *Incr,
321 Value *Mask, Value *ShiftAmt,
322 AtomicOrdering Ord) const override;
327 Value *AlignedAddr, Value *CmpVal,
328 Value *NewVal, Value *Mask,
329 AtomicOrdering Ord) const override;
330
331 /// Returns true if the target allows unaligned memory accesses of the
332 /// specified type.
334 EVT VT, unsigned AddrSpace = 0, Align Alignment = Align(1),
336 unsigned *Fast = nullptr) const override;
337
338 EVT getOptimalMemOpType(LLVMContext &Context, const MemOp &Op,
339 const AttributeList &FuncAttributes) const override;
340
342 SelectionDAG & DAG, const SDLoc &DL, SDValue Val, SDValue *Parts,
343 unsigned NumParts, MVT PartVT, std::optional<CallingConv::ID> CC)
344 const override;
345
347 SelectionDAG & DAG, const SDLoc &DL, const SDValue *Parts,
348 unsigned NumParts, MVT PartVT, EVT ValueVT,
349 std::optional<CallingConv::ID> CC) const override;
350
351 // Return the value of VLMax for the given vector type (i.e. SEW and LMUL)
352 SDValue computeVLMax(MVT VecVT, const SDLoc &DL, SelectionDAG &DAG) const;
353
354 static RISCVVType::VLMUL getLMUL(MVT VT);
355 inline static unsigned computeVLMAX(unsigned VectorBits, unsigned EltSize,
356 unsigned MinSize) {
357 // Original equation:
358 // VLMAX = (VectorBits / EltSize) * LMUL
359 // where LMUL = MinSize / RISCV::RVVBitsPerBlock
360 // The following equations have been reordered to prevent loss of precision
361 // when calculating fractional LMUL.
362 return ((VectorBits / EltSize) * MinSize) / RISCV::RVVBitsPerBlock;
363 }
364
365 // Return inclusive (low, high) bounds on the value of VLMAX for the
366 // given scalable container type given known bounds on VLEN.
367 static std::pair<unsigned, unsigned>
368 computeVLMAXBounds(MVT ContainerVT, const RISCVSubtarget &Subtarget);
369
370 /// Given a vector (either fixed or scalable), return the scalable vector
371 /// corresponding to a vector register (i.e. an m1 register group).
372 static MVT getM1VT(MVT VT) {
373 unsigned EltSizeInBits = VT.getVectorElementType().getSizeInBits();
374 assert(EltSizeInBits <= RISCV::RVVBitsPerBlock && "Unexpected vector MVT");
376 RISCV::RVVBitsPerBlock / EltSizeInBits);
377 }
378
379 static unsigned getRegClassIDForLMUL(RISCVVType::VLMUL LMul);
380 static unsigned getSubregIndexByMVT(MVT VT, unsigned Index);
381 static unsigned getRegClassIDForVecVT(MVT VT);
382 static std::pair<unsigned, unsigned>
384 unsigned InsertExtractIdx,
385 const RISCVRegisterInfo *TRI);
387
388 bool shouldRemoveExtendFromGSIndex(SDValue Extend, EVT DataVT) const override;
389
390 bool isLegalElementTypeForRVV(EVT ScalarTy) const;
391
392 bool shouldConvertFpToSat(unsigned Op, EVT FPVT, EVT VT) const override;
393
394 unsigned getJumpTableEncoding() const override;
395
397 const MachineBasicBlock *MBB,
398 unsigned uid,
399 MCContext &Ctx) const override;
400
401 bool isVScaleKnownToBeAPowerOfTwo() const override;
402
404 ISD::MemIndexedMode &AM, SelectionDAG &DAG) const;
407 SelectionDAG &DAG) const override;
410 SelectionDAG &DAG) const override;
411
413 uint64_t ElemSize) const override {
414 // Scaled addressing not supported on indexed load/stores
415 return Scale == 1;
416 }
417
418 /// If the target has a standard location for the stack protector cookie,
419 /// returns the address of that location. Otherwise, returns nullptr.
420 Value *getIRStackGuard(IRBuilderBase &IRB) const override;
421
422 /// Returns whether or not generating a interleaved load/store intrinsic for
423 /// this type will be legal.
424 bool isLegalInterleavedAccessType(VectorType *VTy, unsigned Factor,
425 Align Alignment, unsigned AddrSpace,
426 const DataLayout &) const;
427
428 /// Return true if a stride load store of the given result type and
429 /// alignment is legal.
430 bool isLegalStridedLoadStore(EVT DataType, Align Alignment) const;
431
432 /// Return true if a fault-only-first load of the given result type and
433 /// alignment is legal.
434 bool isLegalFirstFaultLoad(EVT DataType, Align Alignment) const;
435
436 unsigned getMaxSupportedInterleaveFactor() const override { return 8; }
437
438 bool fallBackToDAGISel(const Instruction &Inst) const override;
439
440 bool lowerInterleavedLoad(Instruction *Load, Value *Mask,
442 ArrayRef<unsigned> Indices, unsigned Factor,
443 const APInt &GapMask) const override;
444
445 bool lowerInterleavedStore(Instruction *Store, Value *Mask,
446 ShuffleVectorInst *SVI, unsigned Factor,
447 const APInt &GapMask) const override;
448
450 IntrinsicInst *DI) const override;
451
453 Instruction *Store, Value *Mask,
454 ArrayRef<Value *> InterleaveValues) const override;
455
456 bool supportKCFIBundles() const override { return true; }
457
459 int JTI, SelectionDAG &DAG) const override;
460
463 const TargetInstrInfo *TII) const override;
464
465 /// True if stack clash protection is enabled for this functions.
466 bool hasInlineStackProbe(const MachineFunction &MF) const override;
467
468 unsigned getStackProbeSize(const MachineFunction &MF, Align StackAlign) const;
469
471 MachineBasicBlock *MBB) const;
472
474
475 bool shouldFoldMaskToVariableShiftPair(SDValue Y) const override;
476
477 /// Control the following reassociation of operands: (op (op x, c1), y) -> (op
478 /// (op x, y), c1) where N0 is (op x, c1) and N1 is y.
480 SDValue N1) const override;
481
482 /// Match a mask which "spreads" the leading elements of a vector evenly
483 /// across the result. Factor is the spread amount, and Index is the
484 /// offset applied.
485 static bool isSpreadMask(ArrayRef<int> Mask, unsigned Factor,
486 unsigned &Index);
487
488private:
489 void analyzeInputArgs(MachineFunction &MF, CCState &CCInfo,
490 const SmallVectorImpl<ISD::InputArg> &Ins, bool IsRet,
491 RISCVCCAssignFn Fn) const;
492 void analyzeOutputArgs(MachineFunction &MF, CCState &CCInfo,
494 bool IsRet, CallLoweringInfo *CLI,
495 RISCVCCAssignFn Fn) const;
496
497 template <class NodeTy>
498 SDValue getAddr(NodeTy *N, SelectionDAG &DAG, bool IsLocal = true,
499 bool IsExternWeak = false) const;
500 SDValue getStaticTLSAddr(GlobalAddressSDNode *N, SelectionDAG &DAG,
501 bool UseGOT) const;
502 SDValue getDynamicTLSAddr(GlobalAddressSDNode *N, SelectionDAG &DAG) const;
503 SDValue getTLSDescAddr(GlobalAddressSDNode *N, SelectionDAG &DAG) const;
504
505 SDValue lowerConstantFP(SDValue Op, SelectionDAG &DAG) const;
506 SDValue lowerGlobalAddress(SDValue Op, SelectionDAG &DAG) const;
507 SDValue lowerBlockAddress(SDValue Op, SelectionDAG &DAG) const;
508 SDValue lowerConstantPool(SDValue Op, SelectionDAG &DAG) const;
509 SDValue lowerJumpTable(SDValue Op, SelectionDAG &DAG) const;
510 SDValue lowerGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const;
512 SDValue lowerBRCOND(SDValue Op, SelectionDAG &DAG) const;
513 SDValue lowerVASTART(SDValue Op, SelectionDAG &DAG) const;
516 SDValue lowerShiftLeftParts(SDValue Op, SelectionDAG &DAG) const;
517 SDValue lowerShiftRightParts(SDValue Op, SelectionDAG &DAG, bool IsSRA) const;
518 SDValue lowerSPLAT_VECTOR_PARTS(SDValue Op, SelectionDAG &DAG) const;
519 SDValue lowerVectorMaskSplat(SDValue Op, SelectionDAG &DAG) const;
520 SDValue lowerVectorMaskExt(SDValue Op, SelectionDAG &DAG,
521 int64_t ExtTrueVal) const;
522 SDValue lowerVectorMaskTruncLike(SDValue Op, SelectionDAG &DAG) const;
523 SDValue lowerVectorTruncLike(SDValue Op, SelectionDAG &DAG) const;
524 SDValue lowerVectorFPExtendOrRoundLike(SDValue Op, SelectionDAG &DAG) const;
525 SDValue lowerINSERT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const;
526 SDValue lowerEXTRACT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const;
527 SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const;
529 SDValue LowerINTRINSIC_VOID(SDValue Op, SelectionDAG &DAG) const;
530 SDValue lowerVPREDUCE(SDValue Op, SelectionDAG &DAG) const;
531 SDValue lowerVECREDUCE(SDValue Op, SelectionDAG &DAG) const;
532 SDValue lowerVectorMaskVecReduction(SDValue Op, SelectionDAG &DAG,
533 bool IsVP) const;
534 SDValue lowerFPVECREDUCE(SDValue Op, SelectionDAG &DAG) const;
535 SDValue lowerINSERT_SUBVECTOR(SDValue Op, SelectionDAG &DAG) const;
536 SDValue lowerEXTRACT_SUBVECTOR(SDValue Op, SelectionDAG &DAG) const;
537 SDValue lowerVECTOR_DEINTERLEAVE(SDValue Op, SelectionDAG &DAG) const;
538 SDValue lowerVECTOR_INTERLEAVE(SDValue Op, SelectionDAG &DAG) const;
539 SDValue lowerSTEP_VECTOR(SDValue Op, SelectionDAG &DAG) const;
540 SDValue lowerVECTOR_REVERSE(SDValue Op, SelectionDAG &DAG) const;
541 SDValue lowerVECTOR_SPLICE(SDValue Op, SelectionDAG &DAG) const;
542 SDValue lowerABS(SDValue Op, SelectionDAG &DAG) const;
543 SDValue lowerMaskedLoad(SDValue Op, SelectionDAG &DAG) const;
544 SDValue lowerLoadFF(SDValue Op, SelectionDAG &DAG) const;
545 SDValue lowerMaskedStore(SDValue Op, SelectionDAG &DAG) const;
546 SDValue lowerVectorCompress(SDValue Op, SelectionDAG &DAG) const;
547 SDValue lowerFixedLengthVectorFCOPYSIGNToRVV(SDValue Op,
548 SelectionDAG &DAG) const;
549 SDValue lowerMaskedGather(SDValue Op, SelectionDAG &DAG) const;
550 SDValue lowerMaskedScatter(SDValue Op, SelectionDAG &DAG) const;
551 SDValue lowerFixedLengthVectorLoadToRVV(SDValue Op, SelectionDAG &DAG) const;
552 SDValue lowerFixedLengthVectorStoreToRVV(SDValue Op, SelectionDAG &DAG) const;
553 SDValue lowerToScalableOp(SDValue Op, SelectionDAG &DAG) const;
554 SDValue LowerIS_FPCLASS(SDValue Op, SelectionDAG &DAG) const;
555 SDValue lowerVPOp(SDValue Op, SelectionDAG &DAG) const;
556 SDValue lowerLogicVPOp(SDValue Op, SelectionDAG &DAG) const;
557 SDValue lowerVPExtMaskOp(SDValue Op, SelectionDAG &DAG) const;
558 SDValue lowerVPSetCCMaskOp(SDValue Op, SelectionDAG &DAG) const;
559 SDValue lowerVPMergeMask(SDValue Op, SelectionDAG &DAG) const;
560 SDValue lowerVPSplatExperimental(SDValue Op, SelectionDAG &DAG) const;
561 SDValue lowerVPSpliceExperimental(SDValue Op, SelectionDAG &DAG) const;
562 SDValue lowerVPReverseExperimental(SDValue Op, SelectionDAG &DAG) const;
563 SDValue lowerVPFPIntConvOp(SDValue Op, SelectionDAG &DAG) const;
564 SDValue lowerVPStridedLoad(SDValue Op, SelectionDAG &DAG) const;
565 SDValue lowerVPStridedStore(SDValue Op, SelectionDAG &DAG) const;
566 SDValue lowerVPCttzElements(SDValue Op, SelectionDAG &DAG) const;
567 SDValue lowerGET_ROUNDING(SDValue Op, SelectionDAG &DAG) const;
568 SDValue lowerSET_ROUNDING(SDValue Op, SelectionDAG &DAG) const;
569 SDValue lowerGET_FPENV(SDValue Op, SelectionDAG &DAG) const;
570 SDValue lowerSET_FPENV(SDValue Op, SelectionDAG &DAG) const;
571 SDValue lowerRESET_FPENV(SDValue Op, SelectionDAG &DAG) const;
572 SDValue lowerGET_FPMODE(SDValue Op, SelectionDAG &DAG) const;
573 SDValue lowerSET_FPMODE(SDValue Op, SelectionDAG &DAG) const;
574 SDValue lowerRESET_FPMODE(SDValue Op, SelectionDAG &DAG) const;
575
576 SDValue lowerEH_DWARF_CFA(SDValue Op, SelectionDAG &DAG) const;
577 SDValue lowerCTLZ_CTTZ_ZERO_UNDEF(SDValue Op, SelectionDAG &DAG) const;
578
579 SDValue lowerStrictFPExtendOrRoundLike(SDValue Op, SelectionDAG &DAG) const;
580
581 SDValue lowerVectorStrictFSetcc(SDValue Op, SelectionDAG &DAG) const;
582
583 SDValue lowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG) const;
584
585 SDValue expandUnalignedRVVLoad(SDValue Op, SelectionDAG &DAG) const;
586 SDValue expandUnalignedRVVStore(SDValue Op, SelectionDAG &DAG) const;
587
588 SDValue expandUnalignedVPLoad(SDValue Op, SelectionDAG &DAG) const;
589 SDValue expandUnalignedVPStore(SDValue Op, SelectionDAG &DAG) const;
590
591 SDValue lowerINIT_TRAMPOLINE(SDValue Op, SelectionDAG &DAG) const;
592 SDValue lowerADJUST_TRAMPOLINE(SDValue Op, SelectionDAG &DAG) const;
593 SDValue lowerPARTIAL_REDUCE_MLA(SDValue Op, SelectionDAG &DAG) const;
594
595 SDValue lowerXAndesBfHCvtBFloat16Load(SDValue Op, SelectionDAG &DAG) const;
596 SDValue lowerXAndesBfHCvtBFloat16Store(SDValue Op, SelectionDAG &DAG) const;
597
598 bool isEligibleForTailCallOptimization(
599 CCState &CCInfo, CallLoweringInfo &CLI, MachineFunction &MF,
600 const SmallVector<CCValAssign, 16> &ArgLocs) const;
601
602 /// Generate error diagnostics if any register used by CC has been marked
603 /// reserved.
604 void validateCCReservedRegs(
605 const SmallVectorImpl<std::pair<llvm::Register, llvm::SDValue>> &Regs,
606 MachineFunction &MF) const;
607
608 bool useRVVForFixedLengthVectorVT(MVT VT) const;
609
610 MVT getVPExplicitVectorLengthTy() const override;
611
612 bool shouldExpandGetVectorLength(EVT TripCountVT, unsigned VF,
613 bool IsScalable) const override;
614
615 /// RVV code generation for fixed length vectors does not lower all
616 /// BUILD_VECTORs. This makes BUILD_VECTOR legalisation a source of stores to
617 /// merge. However, merging them creates a BUILD_VECTOR that is just as
618 /// illegal as the original, thus leading to an infinite legalisation loop.
619 /// NOTE: Once BUILD_VECTOR can be custom lowered for all legal vector types,
620 /// this override can be removed.
621 bool mergeStoresAfterLegalization(EVT VT) const override;
622
623 /// Disable normalizing
624 /// select(N0&N1, X, Y) => select(N0, select(N1, X, Y), Y) and
625 /// select(N0|N1, X, Y) => select(N0, select(N1, X, Y, Y))
626 /// RISC-V doesn't have flags so it's better to perform the and/or in a GPR.
627 bool shouldNormalizeToSelectSequence(LLVMContext &, EVT) const override {
628 return false;
629 }
630
631 /// Disables storing and loading vectors by default when there are function
632 /// calls between the load and store, since these are more expensive than just
633 /// using scalars
634 bool shouldMergeStoreOfLoadsOverCall(EVT SrcVT, EVT MergedVT) const override {
635 return !MergedVT.isVector() || SrcVT.isVector();
636 }
637
638 /// For available scheduling models FDIV + two independent FMULs are much
639 /// faster than two FDIVs.
640 unsigned combineRepeatedFPDivisors() const override;
641
642 SDValue BuildSDIVPow2(SDNode *N, const APInt &Divisor, SelectionDAG &DAG,
643 SmallVectorImpl<SDNode *> &Created) const override;
644
645 bool shouldFoldSelectWithSingleBitTest(EVT VT,
646 const APInt &AndMask) const override;
647
648 unsigned getMinimumJumpTableEntries() const override;
649
650 SDValue emitFlushICache(SelectionDAG &DAG, SDValue InChain, SDValue Start,
651 SDValue End, SDValue Flags, SDLoc DL) const;
652
653 std::pair<const TargetRegisterClass *, uint8_t>
654 findRepresentativeClass(const TargetRegisterInfo *TRI, MVT VT) const override;
655};
656
657namespace RISCVVIntrinsicsTable {
658
660 unsigned IntrinsicID;
664 bool hasScalarOperand() const {
665 // 0xF is not valid. See NoScalarOperand in IntrinsicsRISCV.td.
666 return ScalarOperand != 0xF;
667 }
668 bool hasVLOperand() const {
669 // 0x1F is not valid. See NoVLOperand in IntrinsicsRISCV.td.
670 return VLOperand != 0x1F;
671 }
672};
673
674using namespace RISCV;
675
676#define GET_RISCVVIntrinsicsTable_DECL
677#include "RISCVGenSearchableTables.inc"
678#undef GET_RISCVVIntrinsicsTable_DECL
679
680} // end namespace RISCVVIntrinsicsTable
681
682} // end namespace llvm
683
684#endif
return SDValue()
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
Function Alias Analysis Results
Analysis containing CSE Info
Definition CSEInfo.cpp:27
const HexagonInstrInfo * TII
IRTranslator LLVM IR MI
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
#define RegName(no)
#define I(x, y, z)
Definition MD5.cpp:57
Register const TargetRegisterInfo * TRI
static SDValue lowerSELECT(SDValue Op, SelectionDAG &DAG)
static bool useRVVForFixedLengthVectorVT(MVT VT, const RISCVSubtarget &Subtarget)
const SmallVectorImpl< MachineOperand > & Cond
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
This file describes how to lower LLVM code to machine code.
static SDValue lowerFRAMEADDR(SDValue Op, SelectionDAG &DAG, const VETargetLowering &TLI, const VESubtarget *Subtarget)
static SDValue lowerRETURNADDR(SDValue Op, SelectionDAG &DAG, const VETargetLowering &TLI, const VESubtarget *Subtarget)
static SDValue LowerINTRINSIC_W_CHAIN(SDValue Op, const X86Subtarget &Subtarget, SelectionDAG &DAG)
Class for arbitrary precision integers.
Definition APInt.h:78
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
An instruction that atomically checks whether a specified value is in a memory location,...
an instruction that atomically reads a memory location, combines it with another value,...
CCState - This class holds information needed while lowering arguments and return values.
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
This class represents a function call, abstracting a target machine's calling convention.
This is the shared class of boolean and integer constants.
Definition Constants.h:87
This is an important base class in LLVM.
Definition Constant.h:43
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:63
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Definition Function.h:703
Common base class shared among various IRBuilders.
Definition IRBuilder.h:114
A wrapper class for inspecting calls to intrinsic functions.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
This class is used to represent ISD::LOAD nodes.
Context object for machine code objects.
Definition MCContext.h:83
Base class for the full range of assembler expressions which are needed for parsing.
Definition MCExpr.h:34
Machine Value Type.
static MVT getScalableVectorVT(MVT VT, unsigned NumElements)
TypeSize getSizeInBits() const
Returns the size of the specified MVT in bits.
MVT getVectorElementType() const
Instructions::iterator instr_iterator
Function & getFunction()
Return the LLVM function that this machine code represents.
Representation of each machine instruction.
Flags
Flags values. These may be or'd together.
This is an abstract virtual class for memory operations.
static std::pair< unsigned, unsigned > computeVLMAXBounds(MVT ContainerVT, const RISCVSubtarget &Subtarget)
static std::pair< unsigned, unsigned > decomposeSubvectorInsertExtractToSubRegs(MVT VecVT, MVT SubVecVT, unsigned InsertExtractIdx, const RISCVRegisterInfo *TRI)
ArrayRef< MCPhysReg > getRoundingControlRegisters() const override
Returns a 0 terminated array of rounding control registers that can be attached into strict FP call.
static MVT getM1VT(MVT VT)
Given a vector (either fixed or scalable), return the scalable vector corresponding to a vector regis...
InstructionCost getVRGatherVVCost(MVT VT) const
Return the cost of a vrgather.vv instruction for the type VT.
bool getIndexedAddressParts(SDNode *Op, SDValue &Base, SDValue &Offset, ISD::MemIndexedMode &AM, SelectionDAG &DAG) const
static unsigned getSubregIndexByMVT(MVT VT, unsigned Index)
Value * getIRStackGuard(IRBuilderBase &IRB) const override
If the target has a standard location for the stack protector cookie, returns the address of that loc...
bool shouldConvertFpToSat(unsigned Op, EVT FPVT, EVT VT) const override
Should we generate fp_to_si_sat and fp_to_ui_sat from type FPVT to type VT from min(max(fptoi)) satur...
bool storeOfVectorConstantIsCheap(bool IsZero, EVT MemVT, unsigned NumElem, unsigned AddrSpace) const override
Return true if it is expected to be cheaper to do a store of vector constant with the given size and ...
InlineAsm::ConstraintCode getInlineAsmMemConstraint(StringRef ConstraintCode) const override
SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, const SmallVectorImpl< SDValue > &OutVals, const SDLoc &DL, SelectionDAG &DAG) const override
This hook must be implemented to lower outgoing return values, described by the Outs array,...
unsigned getMaxSupportedInterleaveFactor() const override
Get the maximum supported factor for interleaved memory accesses.
bool mayBeEmittedAsTailCall(const CallInst *CI) const override
Return true if the target may be able emit the call instruction as a tail call.
RISCVTargetLowering(const TargetMachine &TM, const RISCVSubtarget &STI)
MachineBasicBlock * EmitInstrWithCustomInserter(MachineInstr &MI, MachineBasicBlock *BB) const override
This method should be implemented by targets that mark instructions with the 'usesCustomInserter' fla...
Instruction * emitLeadingFence(IRBuilderBase &Builder, Instruction *Inst, AtomicOrdering Ord) const override
Inserts in the IR a target-specific intrinsic specifying a fence.
ISD::NodeType getExtendForAtomicOps() const override
Returns how the platform's atomic operations are extended (ZERO_EXTEND, SIGN_EXTEND,...
bool isTruncateFree(Type *SrcTy, Type *DstTy) const override
Return true if it's free to truncate a value of type FromTy to type ToTy.
bool preferZeroCompareBranch() const override
Return true if the heuristic to prefer icmp eq zero should be used in code gen prepare.
bool shouldRemoveExtendFromGSIndex(SDValue Extend, EVT DataVT) const override
Value * emitMaskedAtomicRMWIntrinsic(IRBuilderBase &Builder, AtomicRMWInst *AI, Value *AlignedAddr, Value *Incr, Value *Mask, Value *ShiftAmt, AtomicOrdering Ord) const override
Perform a masked atomicrmw using a target-specific intrinsic.
bool allowsMisalignedMemoryAccesses(EVT VT, unsigned AddrSpace=0, Align Alignment=Align(1), MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *Fast=nullptr) const override
Returns true if the target allows unaligned memory accesses of the specified type.
const Constant * getTargetConstantFromLoad(LoadSDNode *LD) const override
This method returns the constant pool value that will be loaded by LD.
const RISCVSubtarget & getSubtarget() const
bool lowerDeinterleaveIntrinsicToLoad(Instruction *Load, Value *Mask, IntrinsicInst *DI) const override
Lower a deinterleave intrinsic to a target specific load intrinsic.
TargetLowering::ShiftLegalizationStrategy preferredShiftLegalizationStrategy(SelectionDAG &DAG, SDNode *N, unsigned ExpansionFactor) const override
SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override
This method will be invoked for all target nodes and for any target-independent nodes that the target...
bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override
Return true if folding a constant offset with the given GlobalAddress is legal.
void computeKnownBitsForTargetNode(const SDValue Op, KnownBits &Known, const APInt &DemandedElts, const SelectionDAG &DAG, unsigned Depth) const override
Determine which of the bits specified in Mask are known to be either zero or one and return them in t...
bool preferScalarizeSplat(SDNode *N) const override
bool shouldExtendTypeInLibCall(EVT Type) const override
Returns true if arguments should be extended in lib calls.
bool isLegalAddImmediate(int64_t Imm) const override
Return true if the specified immediate is legal add immediate, that is the target has add instruction...
bool shouldSignExtendTypeInLibCall(Type *Ty, bool IsSigned) const override
Returns true if arguments should be sign-extended in lib calls.
const MCExpr * LowerCustomJumpTableEntry(const MachineJumpTableInfo *MJTI, const MachineBasicBlock *MBB, unsigned uid, MCContext &Ctx) const override
InstructionCost getVRGatherVICost(MVT VT) const
Return the cost of a vrgather.vi (or vx) instruction for the type VT.
bool shouldConvertConstantLoadToIntImm(const APInt &Imm, Type *Ty) const override
Return true if it is beneficial to convert a load of a constant to just the constant itself.
bool targetShrinkDemandedConstant(SDValue Op, const APInt &DemandedBits, const APInt &DemandedElts, TargetLoweringOpt &TLO) const override
bool shouldExpandBuildVectorWithShuffles(EVT VT, unsigned DefinedValues) const override
bool lowerInterleavedStore(Instruction *Store, Value *Mask, ShuffleVectorInst *SVI, unsigned Factor, const APInt &GapMask) const override
Lower an interleaved store into a vssegN intrinsic.
MVT getRegisterTypeForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const override
Return the register type for a given MVT, ensuring vectors are treated as a series of gpr sized integ...
bool decomposeMulByConstant(LLVMContext &Context, EVT VT, SDValue C) const override
Return true if it is profitable to transform an integer multiplication-by-constant into simpler opera...
bool CanLowerReturn(CallingConv::ID CallConv, MachineFunction &MF, bool IsVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, LLVMContext &Context, const Type *RetTy) const override
This hook should be implemented to check whether the return values described by the Outs array can fi...
bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty, unsigned AS, Instruction *I=nullptr) const override
Return true if the addressing mode represented by AM is legal for this target, for a load/store of th...
bool hasAndNotCompare(SDValue Y) const override
Return true if the target should transform: (X & Y) == Y ---> (~X & Y) == 0 (X & Y) !...
bool shouldScalarizeBinop(SDValue VecOp) const override
Try to convert an extract element of a vector binary operation into an extract element followed by a ...
ISD::NodeType getExtendForAtomicRMWArg(unsigned Op) const override
Returns how the platform's atomic rmw operations expect their input argument to be extended (ZERO_EXT...
bool isDesirableToCommuteWithShift(const SDNode *N, CombineLevel Level) const override
Return true if it is profitable to move this shift by a constant amount through its operand,...
bool areTwoSDNodeTargetMMOFlagsMergeable(const MemSDNode &NodeX, const MemSDNode &NodeY) const override
Return true if it is valid to merge the TargetMMOFlags in two SDNodes.
bool hasBitTest(SDValue X, SDValue Y) const override
Return true if the target has a bit-test instruction: (X & (1 << Y)) ==/!= 0 This knowledge can be us...
static unsigned computeVLMAX(unsigned VectorBits, unsigned EltSize, unsigned MinSize)
bool shouldExpandCttzElements(EVT VT) const override
Return true if the @llvm.experimental.cttz.elts intrinsic should be expanded using generic code in Se...
bool isCheapToSpeculateCtlz(Type *Ty) const override
Return true if it is cheap to speculate a call to intrinsic ctlz.
Value * emitMaskedAtomicCmpXchgIntrinsic(IRBuilderBase &Builder, AtomicCmpXchgInst *CI, Value *AlignedAddr, Value *CmpVal, Value *NewVal, Value *Mask, AtomicOrdering Ord) const override
Perform a masked cmpxchg using a target-specific intrinsic.
bool isFPImmLegal(const APFloat &Imm, EVT VT, bool ForCodeSize) const override
Returns true if the target can instruction select the specified FP immediate natively.
bool convertSelectOfConstantsToMath(EVT VT) const override
Return true if a select of constants (select Cond, C1, C2) should be transformed into simple math ops...
InstructionCost getLMULCost(MVT VT) const
Return the cost of LMUL for linear operations.
bool supportKCFIBundles() const override
Return true if the target supports kcfi operand bundles.
unsigned getJumpTableEncoding() const override
Return the entry encoding for a jump table in the current function.
bool isMulAddWithConstProfitable(SDValue AddNode, SDValue ConstNode) const override
Return true if it may be profitable to transform (mul (add x, c1), c2) -> (add (mul x,...
InstructionCost getVSlideVICost(MVT VT) const
Return the cost of a vslidedown.vi or vslideup.vi instruction for the type VT.
bool fallBackToDAGISel(const Instruction &Inst) const override
EVT getOptimalMemOpType(LLVMContext &Context, const MemOp &Op, const AttributeList &FuncAttributes) const override
Returns the target specific optimal type for load and store operations as a result of memset,...
EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const override
Return the ValueType of the result of SETCC operations.
bool isCtpopFast(EVT VT) const override
Return true if ctpop instruction is fast.
unsigned ComputeNumSignBitsForTargetNode(SDValue Op, const APInt &DemandedElts, const SelectionDAG &DAG, unsigned Depth) const override
This method can be implemented by targets that want to expose additional information about sign bits ...
MVT getContainerForFixedLengthVector(MVT VT) const
static unsigned getRegClassIDForVecVT(MVT VT)
Register getExceptionPointerRegister(const Constant *PersonalityFn) const override
If a physical register, this returns the register that receives the exception address on entry to an ...
TargetLowering::AtomicExpansionKind shouldExpandAtomicRMWInIR(AtomicRMWInst *AI) const override
Returns how the IR-level AtomicExpand pass should expand the given AtomicRMW, if at all.
bool isExtractSubvectorCheap(EVT ResVT, EVT SrcVT, unsigned Index) const override
Return true if EXTRACT_SUBVECTOR is cheap for extracting this result type from this source type with ...
std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const override
Given a physical register constraint (e.g.
MachineBasicBlock * emitDynamicProbedAlloc(MachineInstr &MI, MachineBasicBlock *MBB) const
MachineMemOperand::Flags getTargetMMOFlags(const Instruction &I) const override
This callback is used to inspect load/store instructions and add target-specific MachineMemOperand fl...
SDValue computeVLMax(MVT VecVT, const SDLoc &DL, SelectionDAG &DAG) const
bool signExtendConstant(const ConstantInt *CI) const override
Return true if this constant should be sign extended when promoting to a larger type.
bool shouldTransformSignedTruncationCheck(EVT XVT, unsigned KeptBits) const override
Should we tranform the IR-optimal check for whether given truncation down into KeptBits would be trun...
bool shouldProduceAndByConstByHoistingConstFromShiftsLHSOfAnd(SDValue X, ConstantSDNode *XC, ConstantSDNode *CC, SDValue Y, unsigned OldShiftOpcode, unsigned NewShiftOpcode, SelectionDAG &DAG) const override
Given the pattern (X & (C l>>/<< Y)) ==/!= 0 return true if it should be transformed into: ((X <</l>>...
bool hasInlineStackProbe(const MachineFunction &MF) const override
True if stack clash protection is enabled for this functions.
bool hasAndNot(SDValue Y) const override
Return true if the target has a bitwise and-not operation: X = ~A & B This can be used to simplify se...
Register getRegisterByName(const char *RegName, LLT VT, const MachineFunction &MF) const override
Returns the register with the specified architectural or ABI name.
InstructionCost getVSlideVXCost(MVT VT) const
Return the cost of a vslidedown.vx or vslideup.vx instruction for the type VT.
bool isReassocProfitable(SelectionDAG &DAG, SDValue N0, SDValue N1) const override
Control the following reassociation of operands: (op (op x, c1), y) -> (op (op x, y),...
bool shouldFoldMaskToVariableShiftPair(SDValue Y) const override
There are two ways to clear extreme bits (either low or high): Mask: x & (-1 << y) (the instcombine c...
SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override
This callback is invoked for operations that are unsupported by the target, which are registered to u...
bool isUsedByReturnOnly(SDNode *N, SDValue &Chain) const override
Return true if result of the specified node is used by a return node only.
bool softPromoteHalfType() const override
bool isFMAFasterThanFMulAndFAdd(const MachineFunction &MF, EVT VT) const override
Return true if an FMA operation is faster than a pair of fmul and fadd instructions.
TargetLowering::AtomicExpansionKind shouldExpandAtomicCmpXchgInIR(AtomicCmpXchgInst *CI) const override
Returns how the given atomic cmpxchg should be expanded by the IR-level AtomicExpand pass.
Register getExceptionSelectorRegister(const Constant *PersonalityFn) const override
If a physical register, this returns the register that receives the exception typeid on entry to a la...
bool convertSetCCLogicToBitwiseLogic(EVT VT) const override
Use bitwise logic to make pairs of compares more efficient.
unsigned getCustomCtpopCost(EVT VT, ISD::CondCode Cond) const override
Return the maximum number of "x & (x - 1)" operations that can be done instead of deferring to a cust...
void AdjustInstrPostInstrSelection(MachineInstr &MI, SDNode *Node) const override
This method should be implemented by targets that mark instructions with the 'hasPostISelHook' flag.
bool isShuffleMaskLegal(ArrayRef< int > M, EVT VT) const override
Return true if the given shuffle mask can be codegen'd directly, or if it should be stack expanded.
bool isCheapToSpeculateCttz(Type *Ty) const override
Return true if it is cheap to speculate a call to intrinsic cttz.
bool isLegalICmpImmediate(int64_t Imm) const override
Return true if the specified immediate is legal icmp immediate, that is the target has icmp instructi...
ISD::NodeType getExtendForAtomicCmpSwapArg() const override
Returns how the platform's atomic compare and swap expects its comparison value to be extended (ZERO_...
bool lowerInterleavedLoad(Instruction *Load, Value *Mask, ArrayRef< ShuffleVectorInst * > Shuffles, ArrayRef< unsigned > Indices, unsigned Factor, const APInt &GapMask) const override
Lower an interleaved load into a vlsegN intrinsic.
bool lowerInterleaveIntrinsicToStore(Instruction *Store, Value *Mask, ArrayRef< Value * > InterleaveValues) const override
Lower an interleave intrinsic to a target specific store intrinsic.
bool isLegalScaleForGatherScatter(uint64_t Scale, uint64_t ElemSize) const override
SDValue LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg, const SmallVectorImpl< ISD::InputArg > &Ins, const SDLoc &DL, SelectionDAG &DAG, SmallVectorImpl< SDValue > &InVals) const override
This hook must be implemented to lower the incoming (formal) arguments, described by the Ins array,...
void ReplaceNodeResults(SDNode *N, SmallVectorImpl< SDValue > &Results, SelectionDAG &DAG) const override
This callback is invoked when a node result type is illegal for the target, and the operation was reg...
unsigned getVectorTypeBreakdownForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT, EVT &IntermediateVT, unsigned &NumIntermediates, MVT &RegisterVT) const override
Certain targets such as MIPS require that some types such as vectors are always broken down into scal...
bool isLegalElementTypeForRVV(EVT ScalarTy) const
bool isLegalFirstFaultLoad(EVT DataType, Align Alignment) const
Return true if a fault-only-first load of the given result type and alignment is legal.
bool isVScaleKnownToBeAPowerOfTwo() const override
Return true only if vscale must be a power of two.
int getLegalZfaFPImm(const APFloat &Imm, EVT VT) const
void LowerAsmOperandForConstraint(SDValue Op, StringRef Constraint, std::vector< SDValue > &Ops, SelectionDAG &DAG) const override
Lower the specified operand into the Ops vector.
bool splitValueIntoRegisterParts(SelectionDAG &DAG, const SDLoc &DL, SDValue Val, SDValue *Parts, unsigned NumParts, MVT PartVT, std::optional< CallingConv::ID > CC) const override
Target-specific splitting of values into parts that fit a register storing a legal type.
Instruction * emitTrailingFence(IRBuilderBase &Builder, Instruction *Inst, AtomicOrdering Ord) const override
unsigned getNumRegistersForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const override
Return the number of registers for a given MVT, ensuring vectors are treated as a series of gpr sized...
ConstraintType getConstraintType(StringRef Constraint) const override
getConstraintType - Given a constraint letter, return the type of constraint it is for this target.
bool shouldConsiderGEPOffsetSplit() const override
MachineInstr * EmitKCFICheck(MachineBasicBlock &MBB, MachineBasicBlock::instr_iterator &MBBI, const TargetInstrInfo *TII) const override
bool isLegalInterleavedAccessType(VectorType *VTy, unsigned Factor, Align Alignment, unsigned AddrSpace, const DataLayout &) const
Returns whether or not generating a interleaved load/store intrinsic for this type will be legal.
bool canCreateUndefOrPoisonForTargetNode(SDValue Op, const APInt &DemandedElts, const SelectionDAG &DAG, bool PoisonOnly, bool ConsiderFlags, unsigned Depth) const override
Return true if Op can create undef or poison from non-undef & non-poison operands.
bool isIntDivCheap(EVT VT, AttributeList Attr) const override
Return true if integer divide is usually cheaper than a sequence of several shifts,...
LegalizeTypeAction getPreferredVectorAction(MVT VT) const override
Customize the preferred legalization strategy for certain types.
SDValue expandIndirectJTBranch(const SDLoc &dl, SDValue Value, SDValue Addr, int JTI, SelectionDAG &DAG) const override
Expands target specific indirect branch for the case of JumpTable expansion.
static unsigned getRegClassIDForLMUL(RISCVVType::VLMUL LMul)
unsigned getNumRegisters(LLVMContext &Context, EVT VT, std::optional< MVT > RegisterVT=std::nullopt) const override
Return the number of registers for a given MVT, for inline assembly.
bool isMultiStoresCheaperThanBitsMerge(EVT LTy, EVT HTy) const override
Return true if it is cheaper to split the store of a merged int val from a pair of smaller values int...
bool getPostIndexedAddressParts(SDNode *N, SDNode *Op, SDValue &Base, SDValue &Offset, ISD::MemIndexedMode &AM, SelectionDAG &DAG) const override
Returns true by value, base pointer and offset pointer and addressing mode by reference if this node ...
bool shouldFormOverflowOp(unsigned Opcode, EVT VT, bool MathUsed) const override
Try to convert math with an overflow comparison into the corresponding DAG node operation.
bool getPreIndexedAddressParts(SDNode *N, SDValue &Base, SDValue &Offset, ISD::MemIndexedMode &AM, SelectionDAG &DAG) const override
Returns true by value, base pointer and offset pointer and addressing mode by reference if the node's...
SDValue joinRegisterPartsIntoValue(SelectionDAG &DAG, const SDLoc &DL, const SDValue *Parts, unsigned NumParts, MVT PartVT, EVT ValueVT, std::optional< CallingConv::ID > CC) const override
Target-specific combining of register parts into its original value.
bool isMaskAndCmp0FoldingBeneficial(const Instruction &AndI) const override
Return if the target supports combining a chain like:
bool isSExtCheaperThanZExt(EVT SrcVT, EVT DstVT) const override
Return true if sign-extension from FromTy to ToTy is cheaper than zero-extension.
bool isLegalStridedLoadStore(EVT DataType, Align Alignment) const
Return true if a stride load store of the given result type and alignment is legal.
bool getTgtMemIntrinsic(IntrinsicInfo &Info, const CallBase &I, MachineFunction &MF, unsigned Intrinsic) const override
Given an intrinsic, checks if on the target the intrinsic will need to map to a MemIntrinsicNode (tou...
static bool isSpreadMask(ArrayRef< int > Mask, unsigned Factor, unsigned &Index)
Match a mask which "spreads" the leading elements of a vector evenly across the result.
static RISCVVType::VLMUL getLMUL(MVT VT)
SDValue LowerCall(TargetLowering::CallLoweringInfo &CLI, SmallVectorImpl< SDValue > &InVals) const override
This hook must be implemented to lower calls into the specified DAG.
bool SimplifyDemandedBitsForTargetNode(SDValue Op, const APInt &DemandedBits, const APInt &DemandedElts, KnownBits &Known, TargetLoweringOpt &TLO, unsigned Depth) const override
Attempt to simplify any target nodes based on the demanded bits/elts, returning true on success.
bool isZExtFree(SDValue Val, EVT VT2) const override
Return true if zero-extending the specific node Val to type VT2 is free (either because it's implicit...
bool shouldFoldSelectWithIdentityConstant(unsigned BinOpcode, EVT VT, unsigned SelectOpcode, SDValue X, SDValue Y) const override
Return true if pulling a binary operation into a select with an identity constant is profitable.
unsigned getStackProbeSize(const MachineFunction &MF, Align StackAlign) const
bool shouldInsertFencesForAtomic(const Instruction *I) const override
Whether AtomicExpandPass should automatically insert fences and reduce ordering for this atomic.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
MachineFunction & getMachineFunction() const
This instruction constructs a fixed permutation of two input vectors.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
StringRef - Represent a constant reference to a string, i.e.
Definition StringRef.h:55
TargetInstrInfo - Interface to description of machine instruction set.
virtual bool shouldFormOverflowOp(unsigned Opcode, EVT VT, bool MathUsed) const
Try to convert math with an overflow comparison into the corresponding DAG node operation.
ShiftLegalizationStrategy
Return the preferred strategy to legalize tihs SHIFT instruction, with ExpansionFactor being the recu...
virtual unsigned getMinimumJumpTableEntries() const
Return lower limit for number of blocks in a jump table.
virtual ShiftLegalizationStrategy preferredShiftLegalizationStrategy(SelectionDAG &DAG, SDNode *N, unsigned ExpansionFactor) const
AtomicExpansionKind
Enum that specifies what an atomic load/AtomicRMWInst is expanded to, if at all.
virtual MVT getVPExplicitVectorLengthTy() const
Returns the type to be used for the EVL/AVL operand of VP nodes: ISD::VP_ADD, ISD::VP_SUB,...
TargetLowering(const TargetLowering &)=delete
virtual unsigned combineRepeatedFPDivisors() const
Indicate whether this target prefers to combine FDIVs with the same divisor.
Primary interface to the complete machine description for the target machine.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
LLVM Value Representation.
Definition Value.h:75
Base class of all SIMD vector types.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
@ Fast
Attempts to make calls as fast as possible (e.g.
Definition CallingConv.h:41
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
NodeType
ISD::NodeType enum - This enum defines the target-independent operators for a SelectionDAG.
Definition ISDOpcodes.h:41
@ SIGN_EXTEND
Conversion operators.
Definition ISDOpcodes.h:832
MemIndexedMode
MemIndexedMode enum - This enum defines the load / store indexed addressing modes.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
This namespace contains an enum with a value for every intrinsic/builtin function known by LLVM.
static constexpr unsigned RVVBitsPerBlock
This is an optimization pass for GlobalISel generic memory operations.
@ Offset
Definition DWP.cpp:532
bool RISCVCCAssignFn(unsigned ValNo, MVT ValVT, MVT LocVT, CCValAssign::LocInfo LocInfo, ISD::ArgFlagsTy ArgFlags, CCState &State, bool IsRet, Type *OrigTy)
RISCVCCAssignFn - This target-specific function extends the default CCValAssign with additional infor...
AtomicOrdering
Atomic ordering for LLVM's memory model.
CombineLevel
Definition DAGCombine.h:15
DWARFExpression::Operation Op
#define N
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
Extended Value Type.
Definition ValueTypes.h:35
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
Definition ValueTypes.h:147
bool isScalarInteger() const
Return true if this is an integer, but not a vector.
Definition ValueTypes.h:157
bool isInteger() const
Return true if this is an integer or a vector integer type.
Definition ValueTypes.h:152
This structure contains all information that is necessary for lowering calls.