LLVM 23.0.0git
SIISelLowering.h
Go to the documentation of this file.
1//===-- SIISelLowering.h - SI DAG Lowering Interface ------------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9/// \file
10/// SI DAG Lowering interface definition
11//
12//===----------------------------------------------------------------------===//
13
14#ifndef LLVM_LIB_TARGET_AMDGPU_SIISELLOWERING_H
15#define LLVM_LIB_TARGET_AMDGPU_SIISELLOWERING_H
16
18#include "AMDGPUISelLowering.h"
19#include "SIDefines.h"
21
22namespace llvm {
23
24class GCNSubtarget;
26class SIRegisterInfo;
27
28namespace AMDGPU {
30}
31
33private:
34 const GCNSubtarget *Subtarget;
35
36public:
39 EVT VT) const override;
42 EVT VT) const override;
43
45 LLVMContext &Context, CallingConv::ID CC, EVT VT, EVT &IntermediateVT,
46 unsigned &NumIntermediates, MVT &RegisterVT) const override;
47
49
50private:
51 SDValue lowerKernArgParameterPtr(SelectionDAG &DAG, const SDLoc &SL,
52 SDValue Chain, uint64_t Offset) const;
53 SDValue getImplicitArgPtr(SelectionDAG &DAG, const SDLoc &SL) const;
54 SDValue getLDSKernelId(SelectionDAG &DAG, const SDLoc &SL) const;
55 SDValue lowerKernargMemParameter(SelectionDAG &DAG, EVT VT, EVT MemVT,
56 const SDLoc &SL, SDValue Chain,
57 uint64_t Offset, Align Alignment,
58 bool Signed,
59 const ISD::InputArg *Arg = nullptr) const;
60 SDValue loadImplicitKernelArgument(SelectionDAG &DAG, MVT VT, const SDLoc &DL,
61 Align Alignment,
62 ImplicitParameter Param) const;
63
64 SDValue convertABITypeToValueType(SelectionDAG &DAG, SDValue Val,
65 CCValAssign &VA, const SDLoc &SL) const;
66
67 SDValue lowerStackParameter(SelectionDAG &DAG, CCValAssign &VA,
68 const SDLoc &SL, SDValue Chain,
69 const ISD::InputArg &Arg) const;
70 SDValue lowerWorkGroupId(
71 SelectionDAG &DAG, const SIMachineFunctionInfo &MFI, EVT VT,
74 AMDGPUFunctionArgInfo::PreloadedValue ClusterWorkGroupIdPV) const;
75 SDValue getPreloadedValue(SelectionDAG &DAG,
76 const SIMachineFunctionInfo &MFI,
77 EVT VT,
79
80 SDValue LowerGlobalAddress(AMDGPUMachineFunctionInfo *MFI, SDValue Op,
81 SelectionDAG &DAG) const override;
82 SDValue LowerExternalSymbol(SDValue Op, SelectionDAG &DAG) const;
83
84 SDValue lowerImplicitZextParam(SelectionDAG &DAG, SDValue Op,
85 MVT VT, unsigned Offset) const;
86 SDValue lowerImage(SDValue Op, const AMDGPU::ImageDimIntrinsicInfo *Intr,
87 SelectionDAG &DAG, bool WithChain) const;
88 SDValue lowerSBuffer(EVT VT, SDLoc DL, SDValue Rsrc, SDValue Offset,
89 SDValue CachePolicy, SelectionDAG &DAG) const;
90
91 SDValue lowerRawBufferAtomicIntrin(SDValue Op, SelectionDAG &DAG,
92 unsigned NewOpcode) const;
93 SDValue lowerStructBufferAtomicIntrin(SDValue Op, SelectionDAG &DAG,
94 unsigned NewOpcode) const;
95
96 SDValue lowerWaveID(SelectionDAG &DAG, SDValue Op) const;
97 SDValue lowerConstHwRegRead(SelectionDAG &DAG, SDValue Op,
98 AMDGPU::Hwreg::Id HwReg, unsigned LowBit,
99 unsigned Width) const;
100 SDValue lowerWorkitemID(SelectionDAG &DAG, SDValue Op, unsigned Dim,
101 const ArgDescriptor &ArgDesc) const;
102
103 SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const;
104 SDValue LowerINTRINSIC_W_CHAIN(SDValue Op, SelectionDAG &DAG) const;
105 SDValue LowerINTRINSIC_VOID(SDValue Op, SelectionDAG &DAG) const;
106
107 // The raw.tbuffer and struct.tbuffer intrinsics have two offset args: offset
108 // (the offset that is included in bounds checking and swizzling, to be split
109 // between the instruction's voffset and immoffset fields) and soffset (the
110 // offset that is excluded from bounds checking and swizzling, to go in the
111 // instruction's soffset field). This function takes the first kind of
112 // offset and figures out how to split it between voffset and immoffset.
113 std::pair<SDValue, SDValue> splitBufferOffsets(SDValue Offset,
114 SelectionDAG &DAG) const;
115
116 SDValue widenLoad(LoadSDNode *Ld, DAGCombinerInfo &DCI) const;
117 SDValue LowerLOAD(SDValue Op, SelectionDAG &DAG) const;
118 SDValue LowerSELECT(SDValue Op, SelectionDAG &DAG) const;
119 SDValue lowerFastUnsafeFDIV(SDValue Op, SelectionDAG &DAG) const;
120 SDValue lowerFastUnsafeFDIV64(SDValue Op, SelectionDAG &DAG) const;
121 SDValue lowerFDIV_FAST(SDValue Op, SelectionDAG &DAG) const;
122 SDValue LowerFDIV16(SDValue Op, SelectionDAG &DAG) const;
123 SDValue LowerFDIV32(SDValue Op, SelectionDAG &DAG) const;
124 SDValue LowerFDIV64(SDValue Op, SelectionDAG &DAG) const;
125 SDValue LowerFDIV(SDValue Op, SelectionDAG &DAG) const;
126 SDValue LowerFFREXP(SDValue Op, SelectionDAG &DAG) const;
127 SDValue LowerSTORE(SDValue Op, SelectionDAG &DAG) const;
128 SDValue LowerTrig(SDValue Op, SelectionDAG &DAG) const;
129 SDValue lowerFSQRTF16(SDValue Op, SelectionDAG &DAG) const;
130 SDValue lowerFSQRTF32(SDValue Op, SelectionDAG &DAG) const;
131 SDValue lowerFSQRTF64(SDValue Op, SelectionDAG &DAG) const;
132 SDValue LowerATOMIC_CMP_SWAP(SDValue Op, SelectionDAG &DAG) const;
133 SDValue LowerBRCOND(SDValue Op, SelectionDAG &DAG) const;
134 SDValue LowerRETURNADDR(SDValue Op, SelectionDAG &DAG) const;
135 SDValue LowerSPONENTRY(SDValue Op, SelectionDAG &DAG) const;
136 SDValue adjustLoadValueType(unsigned Opcode, MemSDNode *M,
138 bool IsIntrinsic = false) const;
139
140 SDValue lowerIntrinsicLoad(MemSDNode *M, bool IsFormat, SelectionDAG &DAG,
141 ArrayRef<SDValue> Ops) const;
142
143 // Call DAG.getMemIntrinsicNode for a load, but first widen a dwordx3 type to
144 // dwordx4 if on SI.
145 SDValue getMemIntrinsicNode(unsigned Opcode, const SDLoc &DL, SDVTList VTList,
147 MachineMemOperand *MMO, SelectionDAG &DAG) const;
148
149 SDValue handleD16VData(SDValue VData, SelectionDAG &DAG,
150 bool ImageStore = false) const;
151
152 /// Converts \p Op, which must be of floating point type, to the
153 /// floating point type \p VT, by either extending or truncating it.
154 SDValue getFPExtOrFPRound(SelectionDAG &DAG,
155 SDValue Op,
156 const SDLoc &DL,
157 EVT VT) const;
158
159 SDValue convertArgType(
160 SelectionDAG &DAG, EVT VT, EVT MemVT, const SDLoc &SL, SDValue Val,
161 bool Signed, const ISD::InputArg *Arg = nullptr) const;
162
163 /// Custom lowering for ISD::FP_ROUND for MVT::f16.
164 SDValue lowerFP_ROUND(SDValue Op, SelectionDAG &DAG) const;
165 SDValue splitFP_ROUNDVectorOp(SDValue Op, SelectionDAG &DAG) const;
166 SDValue lowerFMINNUM_FMAXNUM(SDValue Op, SelectionDAG &DAG) const;
167 SDValue lowerFMINIMUMNUM_FMAXIMUMNUM(SDValue Op, SelectionDAG &DAG) const;
168 SDValue lowerFMINIMUM_FMAXIMUM(SDValue Op, SelectionDAG &DAG) const;
169 SDValue lowerFLDEXP(SDValue Op, SelectionDAG &DAG) const;
170 SDValue promoteUniformOpToI32(SDValue Op, DAGCombinerInfo &DCI) const;
171 SDValue lowerFCOPYSIGN(SDValue Op, SelectionDAG &DAG) const;
172 SDValue lowerMUL(SDValue Op, SelectionDAG &DAG) const;
173 SDValue lowerXMULO(SDValue Op, SelectionDAG &DAG) const;
174 SDValue lowerXMUL_LOHI(SDValue Op, SelectionDAG &DAG) const;
175
176 SDValue getSegmentAperture(unsigned AS, const SDLoc &DL,
177 SelectionDAG &DAG) const;
178
179 SDValue lowerADDRSPACECAST(SDValue Op, SelectionDAG &DAG) const;
180 SDValue lowerINSERT_SUBVECTOR(SDValue Op, SelectionDAG &DAG) const;
181 SDValue lowerINSERT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const;
182 SDValue lowerEXTRACT_VECTOR_ELT(SDValue Op, SelectionDAG &DAG) const;
183 SDValue lowerVECTOR_SHUFFLE(SDValue Op, SelectionDAG &DAG) const;
184 SDValue lowerSCALAR_TO_VECTOR(SDValue Op, SelectionDAG &DAG) const;
185 SDValue lowerBUILD_VECTOR(SDValue Op, SelectionDAG &DAG) const;
186
187 SDValue lowerTRAP(SDValue Op, SelectionDAG &DAG) const;
188 SDValue lowerTrapEndpgm(SDValue Op, SelectionDAG &DAG) const;
189 SDValue lowerTrapHsaQueuePtr(SDValue Op, SelectionDAG &DAG) const;
190 SDValue lowerTrapHsa(SDValue Op, SelectionDAG &DAG) const;
191 SDValue lowerDEBUGTRAP(SDValue Op, SelectionDAG &DAG) const;
192 SDValue LowerINLINEASM(SDValue Op, SelectionDAG &DAG) const;
193
194 SDNode *adjustWritemask(MachineSDNode *&N, SelectionDAG &DAG) const;
195
196 SDValue performUCharToFloatCombine(SDNode *N,
197 DAGCombinerInfo &DCI) const;
198 SDValue performFCopySignCombine(SDNode *N, DAGCombinerInfo &DCI) const;
199
200 SDValue performSHLPtrCombine(SDNode *N,
201 unsigned AS,
202 EVT MemVT,
203 DAGCombinerInfo &DCI) const;
204
205 SDValue performMemSDNodeCombine(MemSDNode *N, DAGCombinerInfo &DCI) const;
206
207 SDValue splitBinaryBitConstantOp(DAGCombinerInfo &DCI, const SDLoc &SL,
208 unsigned Opc, SDValue LHS,
209 const ConstantSDNode *CRHS) const;
210
211 SDValue performAndCombine(SDNode *N, DAGCombinerInfo &DCI) const;
212 SDValue performOrCombine(SDNode *N, DAGCombinerInfo &DCI) const;
213 SDValue performXorCombine(SDNode *N, DAGCombinerInfo &DCI) const;
214 SDValue performZeroOrAnyExtendCombine(SDNode *N, DAGCombinerInfo &DCI) const;
215 SDValue performSignExtendInRegCombine(SDNode *N, DAGCombinerInfo &DCI) const;
216 SDValue performClassCombine(SDNode *N, DAGCombinerInfo &DCI) const;
217 SDValue getCanonicalConstantFP(SelectionDAG &DAG, const SDLoc &SL, EVT VT,
218 const APFloat &C) const;
219 SDValue performFCanonicalizeCombine(SDNode *N, DAGCombinerInfo &DCI) const;
220
221 SDValue performFPMed3ImmCombine(SelectionDAG &DAG, const SDLoc &SL,
222 SDValue Op0, SDValue Op1,
223 bool IsKnownNoNaNs) const;
224 SDValue performIntMed3ImmCombine(SelectionDAG &DAG, const SDLoc &SL,
225 SDValue Src, SDValue MinVal, SDValue MaxVal,
226 bool Signed) const;
227 SDValue performMinMaxCombine(SDNode *N, DAGCombinerInfo &DCI) const;
228 SDValue performFMed3Combine(SDNode *N, DAGCombinerInfo &DCI) const;
229 SDValue performCvtPkRTZCombine(SDNode *N, DAGCombinerInfo &DCI) const;
230 SDValue performExtractVectorEltCombine(SDNode *N, DAGCombinerInfo &DCI) const;
231 SDValue performInsertVectorEltCombine(SDNode *N, DAGCombinerInfo &DCI) const;
232 SDValue performFPRoundCombine(SDNode *N, DAGCombinerInfo &DCI) const;
233 SDValue performSelectCombine(SDNode *N, DAGCombinerInfo &DCI) const;
234
235 SDValue reassociateScalarOps(SDNode *N, SelectionDAG &DAG) const;
236 unsigned getFusedOpcode(const SelectionDAG &DAG,
237 const SDNode *N0, const SDNode *N1) const;
238 SDValue tryFoldToMad64_32(SDNode *N, DAGCombinerInfo &DCI) const;
239 SDValue foldAddSub64WithZeroLowBitsTo32(SDNode *N,
240 DAGCombinerInfo &DCI) const;
241
242 SDValue performAddCombine(SDNode *N, DAGCombinerInfo &DCI) const;
243 SDValue performPtrAddCombine(SDNode *N, DAGCombinerInfo &DCI) const;
244 SDValue performAddCarrySubCarryCombine(SDNode *N, DAGCombinerInfo &DCI) const;
245 SDValue performSubCombine(SDNode *N, DAGCombinerInfo &DCI) const;
246 SDValue performFAddCombine(SDNode *N, DAGCombinerInfo &DCI) const;
247 SDValue performFSubCombine(SDNode *N, DAGCombinerInfo &DCI) const;
248 SDValue performFDivCombine(SDNode *N, DAGCombinerInfo &DCI) const;
249 SDValue performFMulCombine(SDNode *N, DAGCombinerInfo &DCI) const;
250 SDValue performFMACombine(SDNode *N, DAGCombinerInfo &DCI) const;
251 SDValue performSetCCCombine(SDNode *N, DAGCombinerInfo &DCI) const;
252 SDValue performCvtF32UByteNCombine(SDNode *N, DAGCombinerInfo &DCI) const;
253 SDValue performClampCombine(SDNode *N, DAGCombinerInfo &DCI) const;
254 SDValue performRcpCombine(SDNode *N, DAGCombinerInfo &DCI) const;
255
256 bool isLegalMUBUFAddressingMode(const AddrMode &AM) const;
257
258 unsigned isCFIntrinsic(const SDNode *Intr) const;
259
260public:
261 /// \returns True if fixup needs to be emitted for given global value \p GV,
262 /// false otherwise.
263 bool shouldEmitFixup(const GlobalValue *GV) const;
264
265 /// \returns True if GOT relocation needs to be emitted for given global value
266 /// \p GV, false otherwise.
267 bool shouldEmitGOTReloc(const GlobalValue *GV) const;
268
269 /// \returns True if PC-relative relocation needs to be emitted for given
270 /// global value \p GV, false otherwise.
271 bool shouldEmitPCReloc(const GlobalValue *GV) const;
272
273 /// \returns true if this should use a literal constant for an LDS address,
274 /// and not emit a relocation for an LDS global.
275 bool shouldUseLDSConstAddress(const GlobalValue *GV) const;
276
277 /// Check if EXTRACT_VECTOR_ELT/INSERT_VECTOR_ELT (<n x e>, var-idx) should be
278 /// expanded into a set of cmp/select instructions.
279 static bool shouldExpandVectorDynExt(unsigned EltSize, unsigned NumElem,
280 bool IsDivergentIdx,
281 const GCNSubtarget *Subtarget);
282
283 bool shouldExpandVectorDynExt(SDNode *N) const;
284
285 bool shouldPreservePtrArith(const Function &F, EVT PtrVT) const override;
286
288 EVT PtrVT) const override;
289
290private:
291 /// Returns true if the first real instruction in MBB is 8 bytes and could
292 /// be split by a 32-byte fetch window boundary. Used on GFX950 to avoid
293 /// instruction fetch delays.
294 bool needsFetchWindowAlignment(const MachineBasicBlock &MBB) const;
295
296 // Analyze a combined offset from an amdgcn_s_buffer_load intrinsic and store
297 // the three offsets (voffset, soffset and instoffset) into the SDValue[3]
298 // array pointed to by Offsets.
299 void setBufferOffsets(SDValue CombinedOffset, SelectionDAG &DAG,
300 SDValue *Offsets, Align Alignment = Align(4)) const;
301
302 // Convert the i128 that an addrspace(8) pointer is natively represented as
303 // into the v4i32 that all the buffer intrinsics expect to receive. We can't
304 // add register classes for i128 on pain of the promotion logic going haywire,
305 // so this slightly ugly hack is what we've got. If passed a non-pointer
306 // argument (as would be seen in older buffer intrinsics), does nothing.
307 SDValue bufferRsrcPtrToVector(SDValue MaybePointer, SelectionDAG &DAG) const;
308
309 // Wrap a 64-bit pointer into a v4i32 (which is how all SelectionDAG code
310 // represents ptr addrspace(8)) using the flags specified in the intrinsic.
311 SDValue lowerPointerAsRsrcIntrin(SDNode *Op, SelectionDAG &DAG) const;
312
313 // Handle 8 bit and 16 bit buffer loads
314 SDValue handleByteShortBufferLoads(SelectionDAG &DAG, EVT LoadVT, SDLoc DL,
317 bool IsTFE = false) const;
318
319 // Handle 8 bit and 16 bit buffer stores
320 SDValue handleByteShortBufferStores(SelectionDAG &DAG, EVT VDataType,
321 SDLoc DL, SDValue Ops[],
322 MemSDNode *M) const;
323
324public:
325 SITargetLowering(const TargetMachine &tm, const GCNSubtarget &STI);
326
327 const GCNSubtarget *getSubtarget() const;
328
330
331 bool isFPExtFoldable(const SelectionDAG &DAG, unsigned Opcode, EVT DestVT,
332 EVT SrcVT) const override;
333
334 bool isFPExtFoldable(const MachineInstr &MI, unsigned Opcode, LLT DestTy,
335 LLT SrcTy) const override;
336
337 bool isShuffleMaskLegal(ArrayRef<int> /*Mask*/, EVT /*VT*/) const override;
338
339 // While address space 7 should never make it to codegen, it still needs to
340 // have a MVT to prevent some analyses that query this function from breaking.
341 // We use the custum MVT::amdgpuBufferFatPointer and
342 // amdgpu::amdgpuBufferStridedPointer for this, though we use v8i32 for the
343 // memory type (which is probably unused).
344 MVT getPointerTy(const DataLayout &DL, unsigned AS) const override;
345 MVT getPointerMemTy(const DataLayout &DL, unsigned AS) const override;
346
348 MachineFunction &MF,
349 unsigned IntrinsicID) const override;
350
353 SelectionDAG &DAG) const override;
354
357 Type *&AccessTy) const override;
358
359 bool isLegalFlatAddressingMode(const AddrMode &AM, unsigned AddrSpace) const;
360 bool isLegalGlobalAddressingMode(const AddrMode &AM) const;
361 bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty,
362 unsigned AS,
363 Instruction *I = nullptr) const override;
364
365 bool canMergeStoresTo(unsigned AS, EVT MemVT,
366 const MachineFunction &MF) const override;
367
369 unsigned Size, unsigned AddrSpace, Align Alignment,
371 unsigned *IsFast = nullptr) const;
372
374 LLT Ty, unsigned AddrSpace, Align Alignment,
376 unsigned *IsFast = nullptr) const override {
377 if (IsFast)
378 *IsFast = 0;
379 return allowsMisalignedMemoryAccessesImpl(Ty.getSizeInBits(), AddrSpace,
380 Alignment, Flags, IsFast);
381 }
382
384 EVT VT, unsigned AS, Align Alignment,
386 unsigned *IsFast = nullptr) const override;
387
388 EVT getOptimalMemOpType(LLVMContext &Context, const MemOp &Op,
389 const AttributeList &FuncAttributes) const override;
390
391 bool isMemOpHasNoClobberedMemOperand(const SDNode *N) const;
392
393 static bool isNonGlobalAddrSpace(unsigned AS);
394
395 bool isFreeAddrSpaceCast(unsigned SrcAS, unsigned DestAS) const override;
396
398 getPreferredVectorAction(MVT VT) const override;
399
401 Type *Ty) const override;
402
403 bool isExtractSubvectorCheap(EVT ResVT, EVT SrcVT,
404 unsigned Index) const override;
405 bool isExtractVecEltCheap(EVT VT, unsigned Index) const override;
406
407 bool isTypeDesirableForOp(unsigned Op, EVT VT) const override;
408
409 bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override;
410
411 unsigned combineRepeatedFPDivisors() const override {
412 // Combine multiple FDIVs with the same divisor into multiple FMULs by the
413 // reciprocal.
414 return 2;
415 }
416
417 bool supportSplitCSR(MachineFunction *MF) const override;
418 void initializeSplitCSR(MachineBasicBlock *Entry) const override;
420 MachineBasicBlock *Entry,
421 const SmallVectorImpl<MachineBasicBlock *> &Exits) const override;
422
424 bool isVarArg,
426 const SDLoc &DL, SelectionDAG &DAG,
427 SmallVectorImpl<SDValue> &InVals) const override;
428
429 bool CanLowerReturn(CallingConv::ID CallConv,
430 MachineFunction &MF, bool isVarArg,
432 LLVMContext &Context, const Type *RetTy) const override;
433
434 SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg,
436 const SmallVectorImpl<SDValue> &OutVals, const SDLoc &DL,
437 SelectionDAG &DAG) const override;
438
440 CallLoweringInfo &CLI,
441 CCState &CCInfo,
442 const SIMachineFunctionInfo &Info,
443 SmallVectorImpl<std::pair<unsigned, SDValue>> &RegsToPass,
444 SmallVectorImpl<SDValue> &MemOpChains,
445 SDValue Chain) const;
446
448 CallingConv::ID CallConv, bool isVarArg,
450 const SDLoc &DL, SelectionDAG &DAG,
451 SmallVectorImpl<SDValue> &InVals, bool isThisReturn,
452 SDValue ThisVal) const;
453
454 bool mayBeEmittedAsTailCall(const CallInst *) const override;
455
457 SDValue Callee, CallingConv::ID CalleeCC, bool isVarArg,
459 const SmallVectorImpl<SDValue> &OutVals,
460 const SmallVectorImpl<ISD::InputArg> &Ins, SelectionDAG &DAG) const;
461
462 SDValue LowerCall(CallLoweringInfo &CLI,
463 SmallVectorImpl<SDValue> &InVals) const override;
464
469
475
476 Register getRegisterByName(const char* RegName, LLT VT,
477 const MachineFunction &MF) const override;
478
480 MachineBasicBlock *BB) const;
481
484 MachineBasicBlock *BB) const;
485
488 MachineBasicBlock *BB) const override;
489
490 bool enableAggressiveFMAFusion(EVT VT) const override;
491 bool enableAggressiveFMAFusion(LLT Ty) const override;
493 EVT VT) const override;
494 MVT getScalarShiftAmountTy(const DataLayout &, EVT) const override;
495 LLT getPreferredShiftAmountTy(LLT Ty) const override;
496
498 EVT VT) const override;
500 const LLT Ty) const override;
501 bool isFMADLegal(const SelectionDAG &DAG, const SDNode *N) const override;
502 bool isFMADLegal(const MachineInstr &MI, const LLT Ty) const override;
503
507 SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override;
509 SelectionDAG &DAG) const override;
510
511 SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override;
512 SDNode *PostISelFolding(MachineSDNode *N, SelectionDAG &DAG) const override;
513 void AddMemOpInit(MachineInstr &MI) const;
515 SDNode *Node) const override;
516
518
520 SDValue Ptr) const;
522 uint32_t RsrcDword1, uint64_t RsrcDword2And3) const;
523 std::pair<unsigned, const TargetRegisterClass *>
525 StringRef Constraint, MVT VT) const override;
526 ConstraintType getConstraintType(StringRef Constraint) const override;
528 std::vector<SDValue> &Ops,
529 SelectionDAG &DAG) const override;
530 bool getAsmOperandConstVal(SDValue Op, uint64_t &Val) const;
531 bool checkAsmConstraintVal(SDValue Op, StringRef Constraint,
532 uint64_t Val) const;
534 uint64_t Val,
535 unsigned MaxSize = 64) const;
536 SDValue copyToM0(SelectionDAG &DAG, SDValue Chain, const SDLoc &DL,
537 SDValue V) const;
538
539 void finalizeLowering(MachineFunction &MF) const override;
540
542 const APInt &DemandedElts,
543 const SelectionDAG &DAG,
544 unsigned Depth = 0) const override;
545 void computeKnownBitsForFrameIndex(int FrameIdx,
546 KnownBits &Known,
547 const MachineFunction &MF) const override;
549 KnownBits &Known,
550 const APInt &DemandedElts,
551 const MachineRegisterInfo &MRI,
552 unsigned Depth = 0) const override;
553
555 Register R,
556 const MachineRegisterInfo &MRI,
557 unsigned Depth = 0) const override;
559 UniformityInfo *UA) const override;
560
561 bool hasMemSDNodeUser(SDNode *N) const;
562
564 SDValue N1) const override;
565
567 Register N1) const override;
568
570 SDNodeFlags UserFlags = {}, unsigned MaxDepth = 5) const;
571 bool isCanonicalized(Register Reg, const MachineFunction &MF,
572 unsigned MaxDepth = 5) const;
573 bool denormalsEnabledForType(const SelectionDAG &DAG, EVT VT) const;
574 bool denormalsEnabledForType(LLT Ty, const MachineFunction &MF) const;
575
576 bool isKnownNeverNaNForTargetNode(SDValue Op, const APInt &DemandedElts,
577 const SelectionDAG &DAG, bool SNaN = false,
578 unsigned Depth = 0) const override;
580 shouldExpandAtomicRMWInIR(const AtomicRMWInst *) const override;
581 AtomicExpansionKind shouldExpandAtomicLoadInIR(LoadInst *LI) const override;
582 AtomicExpansionKind shouldExpandAtomicStoreInIR(StoreInst *SI) const override;
584 shouldExpandAtomicCmpXchgInIR(const AtomicCmpXchgInst *AI) const override;
585
586 void emitExpandAtomicAddrSpacePredicate(Instruction *AI) const;
587 void emitExpandAtomicRMW(AtomicRMWInst *AI) const override;
588 void emitExpandAtomicCmpXchg(AtomicCmpXchgInst *CI) const override;
589 void emitExpandAtomicLoad(LoadInst *LI) const override;
590 void emitExpandAtomicStore(StoreInst *SI) const override;
591
592 LoadInst *
593 lowerIdempotentRMWIntoFencedLoad(AtomicRMWInst *AI) const override;
594
595 const TargetRegisterClass *getRegClassFor(MVT VT,
596 bool isDivergent) const override;
597 bool requiresUniformRegister(MachineFunction &MF,
598 const Value *V) const override;
599 Align getPrefLoopAlignment(MachineLoop *ML) const override;
600 unsigned
601 getMaxPermittedBytesForAlignment(MachineBasicBlock *MBB) const override;
602
603 void allocateHSAUserSGPRs(CCState &CCInfo,
604 MachineFunction &MF,
605 const SIRegisterInfo &TRI,
606 SIMachineFunctionInfo &Info) const;
607
608 void allocatePreloadKernArgSGPRs(CCState &CCInfo,
609 SmallVectorImpl<CCValAssign> &ArgLocs,
610 const SmallVectorImpl<ISD::InputArg> &Ins,
611 MachineFunction &MF,
612 const SIRegisterInfo &TRI,
613 SIMachineFunctionInfo &Info) const;
614
615 void allocateLDSKernelId(CCState &CCInfo, MachineFunction &MF,
616 const SIRegisterInfo &TRI,
617 SIMachineFunctionInfo &Info) const;
618
619 void allocateSystemSGPRs(CCState &CCInfo,
620 MachineFunction &MF,
621 SIMachineFunctionInfo &Info,
622 CallingConv::ID CallConv,
623 bool IsShader) const;
624
625 void allocateSpecialEntryInputVGPRs(CCState &CCInfo,
626 MachineFunction &MF,
627 const SIRegisterInfo &TRI,
628 SIMachineFunctionInfo &Info) const;
630 CCState &CCInfo,
631 MachineFunction &MF,
632 const SIRegisterInfo &TRI,
633 SIMachineFunctionInfo &Info) const;
634
635 void allocateSpecialInputVGPRs(CCState &CCInfo,
636 MachineFunction &MF,
637 const SIRegisterInfo &TRI,
638 SIMachineFunctionInfo &Info) const;
639 void allocateSpecialInputVGPRsFixed(CCState &CCInfo,
640 MachineFunction &MF,
641 const SIRegisterInfo &TRI,
642 SIMachineFunctionInfo &Info) const;
643
645 getTargetMMOFlags(const Instruction &I) const override;
646};
647
648// Returns true if argument is a boolean value which is not serialized into
649// memory or argument and does not require v_cndmask_b32 to be deserialized.
650bool isBoolSGPR(SDValue V);
651
652} // End namespace llvm
653
654#endif
return SDValue()
Interface definition of the TargetLowering class that is common to all AMD GPUs.
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis Results
block Block Frequency Analysis
IRTranslator LLVM IR MI
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
#define RegName(no)
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
Register Reg
Register const TargetRegisterInfo * TRI
Promote Memory to Register
Definition Mem2Reg.cpp:110
StandardInstrumentations SI(Mod->getContext(), Debug, VerifyEach)
Value * LHS
AMDGPUTargetLowering(const TargetMachine &TM, const TargetSubtargetInfo &STI, const AMDGPUSubtarget &AMDGPUSTI)
Class for arbitrary precision integers.
Definition APInt.h:78
Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
CCState - This class holds information needed while lowering arguments and return values.
CCValAssign - Represent assignment of one arg/retval to a location.
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
This class represents a function call, abstracting a target machine's calling convention.
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
FunctionLoweringInfo - This contains information that is global to a function that is used when lower...
A wrapper class for inspecting calls to intrinsic functions.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
This class is used to represent ISD::LOAD nodes.
Machine Value Type.
Representation of each machine instruction.
A description of a memory reference used in the backend.
Flags
Flags values. These may be or'd together.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
An SDNode that represents everything that will be needed to construct a MachineInstr.
This is an abstract virtual class for memory operations.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
This class keeps track of the SPI_SP_INPUT_ADDR config register, which tells the hardware which inter...
bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override
Return true if folding a constant offset with the given GlobalAddress is legal.
bool isTypeDesirableForOp(unsigned Op, EVT VT) const override
Return true if the target has native support for the specified value type and it is 'desirable' to us...
SDNode * PostISelFolding(MachineSDNode *N, SelectionDAG &DAG) const override
Fold the instructions after selecting them.
SDValue splitTernaryVectorOp(SDValue Op, SelectionDAG &DAG) const
MachineSDNode * wrapAddr64Rsrc(SelectionDAG &DAG, const SDLoc &DL, SDValue Ptr) const
bool isFMAFasterThanFMulAndFAdd(const MachineFunction &MF, EVT VT) const override
Return true if an FMA operation is faster than a pair of fmul and fadd instructions.
SDValue lowerGET_ROUNDING(SDValue Op, SelectionDAG &DAG) const
AtomicExpansionKind shouldExpandAtomicRMWInIR(const AtomicRMWInst *) const override
Returns how the IR-level AtomicExpand pass should expand the given AtomicRMW, if at all.
bool requiresUniformRegister(MachineFunction &MF, const Value *V) const override
Allows target to decide about the register class of the specific value that is live outside the defin...
bool isFMADLegal(const SelectionDAG &DAG, const SDNode *N) const override
Returns true if be combined with to form an ISD::FMAD.
AtomicExpansionKind shouldExpandAtomicStoreInIR(StoreInst *SI) const override
Returns how the given (atomic) store should be expanded by the IR-level AtomicExpand pass into.
void bundleInstWithWaitcnt(MachineInstr &MI) const
Insert MI into a BUNDLE with an S_WAITCNT 0 immediately following it.
SDValue lowerROTR(SDValue Op, SelectionDAG &DAG) const
MVT getScalarShiftAmountTy(const DataLayout &, EVT) const override
Return the type to use for a scalar shift opcode, given the shifted amount type.
SDValue LowerCall(CallLoweringInfo &CLI, SmallVectorImpl< SDValue > &InVals) const override
This hook must be implemented to lower calls into the specified DAG.
MVT getPointerTy(const DataLayout &DL, unsigned AS) const override
Map address space 7 to MVT::amdgpuBufferFatPointer because that's its in-memory representation.
bool denormalsEnabledForType(const SelectionDAG &DAG, EVT VT) const
void insertCopiesSplitCSR(MachineBasicBlock *Entry, const SmallVectorImpl< MachineBasicBlock * > &Exits) const override
Insert explicit copies in entry and exit blocks.
EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const override
Return the ValueType of the result of SETCC operations.
SDNode * legalizeTargetIndependentNode(SDNode *Node, SelectionDAG &DAG) const
Legalize target independent instructions (e.g.
bool allowsMisalignedMemoryAccessesImpl(unsigned Size, unsigned AddrSpace, Align Alignment, MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *IsFast=nullptr) const
TargetLoweringBase::LegalizeTypeAction getPreferredVectorAction(MVT VT) const override
Return the preferred vector type legalization action.
SDValue lowerFP_EXTEND(SDValue Op, SelectionDAG &DAG) const
const GCNSubtarget * getSubtarget() const
bool enableAggressiveFMAFusion(EVT VT) const override
Return true if target always benefits from combining into FMA for a given value type.
bool shouldEmitGOTReloc(const GlobalValue *GV) const
void CollectTargetIntrinsicOperands(const CallInst &I, SmallVectorImpl< SDValue > &Ops, SelectionDAG &DAG) const override
SDValue splitUnaryVectorOp(SDValue Op, SelectionDAG &DAG) const
SDValue lowerGET_FPENV(SDValue Op, SelectionDAG &DAG) const
bool isCanonicalized(SelectionDAG &DAG, SDValue Op, SDNodeFlags UserFlags={}, unsigned MaxDepth=5) const
void allocateSpecialInputSGPRs(CCState &CCInfo, MachineFunction &MF, const SIRegisterInfo &TRI, SIMachineFunctionInfo &Info) const
void allocateLDSKernelId(CCState &CCInfo, MachineFunction &MF, const SIRegisterInfo &TRI, SIMachineFunctionInfo &Info) const
SDValue LowerSTACKSAVE(SDValue Op, SelectionDAG &DAG) const
bool isReassocProfitable(SelectionDAG &DAG, SDValue N0, SDValue N1) const override
void allocateHSAUserSGPRs(CCState &CCInfo, MachineFunction &MF, const SIRegisterInfo &TRI, SIMachineFunctionInfo &Info) const
ArrayRef< MCPhysReg > getRoundingControlRegisters() const override
Returns a 0 terminated array of rounding control registers that can be attached into strict FP call.
ConstraintType getConstraintType(StringRef Constraint) const override
Given a constraint, return the type of constraint it is for this target.
SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, const SmallVectorImpl< SDValue > &OutVals, const SDLoc &DL, SelectionDAG &DAG) const override
This hook must be implemented to lower outgoing return values, described by the Outs array,...
const TargetRegisterClass * getRegClassFor(MVT VT, bool isDivergent) const override
Return the register class that should be used for the specified value type.
void AddMemOpInit(MachineInstr &MI) const
MachineMemOperand::Flags getTargetMMOFlags(const Instruction &I) const override
This callback is used to inspect load/store instructions and add target-specific MachineMemOperand fl...
bool isLegalGlobalAddressingMode(const AddrMode &AM) const
void computeKnownBitsForFrameIndex(int FrameIdx, KnownBits &Known, const MachineFunction &MF) const override
Determine which of the bits of FrameIndex FIOp are known to be 0.
bool shouldConvertConstantLoadToIntImm(const APInt &Imm, Type *Ty) const override
Return true if it is beneficial to convert a load of a constant to just the constant itself.
Align getPrefLoopAlignment(MachineLoop *ML) const override
Return the preferred loop alignment.
std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const override
Given a physical register constraint (e.g.
void emitExpandAtomicStore(StoreInst *SI) const override
Perform a atomic store using a target-specific way.
AtomicExpansionKind shouldExpandAtomicLoadInIR(LoadInst *LI) const override
Returns how the given (atomic) load should be expanded by the IR-level AtomicExpand pass.
Align computeKnownAlignForTargetInstr(GISelValueTracking &Analysis, Register R, const MachineRegisterInfo &MRI, unsigned Depth=0) const override
Determine the known alignment for the pointer value R.
bool getAsmOperandConstVal(SDValue Op, uint64_t &Val) const
bool isShuffleMaskLegal(ArrayRef< int >, EVT) const override
Targets can use this to indicate that they only support some VECTOR_SHUFFLE operations,...
void emitExpandAtomicLoad(LoadInst *LI) const override
Perform a atomic load using a target-specific way.
EVT getOptimalMemOpType(LLVMContext &Context, const MemOp &Op, const AttributeList &FuncAttributes) const override
Returns the target specific optimal type for load and store operations as a result of memset,...
void ReplaceNodeResults(SDNode *N, SmallVectorImpl< SDValue > &Results, SelectionDAG &DAG) const override
This callback is invoked when a node result type is illegal for the target, and the operation was reg...
Register getRegisterByName(const char *RegName, LLT VT, const MachineFunction &MF) const override
Return the register ID of the name passed in.
void LowerAsmOperandForConstraint(SDValue Op, StringRef Constraint, std::vector< SDValue > &Ops, SelectionDAG &DAG) const override
Lower the specified operand into the Ops vector.
LLT getPreferredShiftAmountTy(LLT Ty) const override
Return the preferred type to use for a shift opcode, given the shifted amount type is ShiftValueTy.
bool isLegalAddressingMode(const DataLayout &DL, const AddrMode &AM, Type *Ty, unsigned AS, Instruction *I=nullptr) const override
Return true if the addressing mode represented by AM is legal for this target, for a load/store of th...
SDValue lowerSET_FPENV(SDValue Op, SelectionDAG &DAG) const
bool shouldPreservePtrArith(const Function &F, EVT PtrVT) const override
True if target has some particular form of dealing with pointer arithmetic semantics for pointers wit...
void getTgtMemIntrinsic(SmallVectorImpl< IntrinsicInfo > &, const CallBase &, MachineFunction &MF, unsigned IntrinsicID) const override
Given an intrinsic, checks if on the target the intrinsic will need to map to a MemIntrinsicNode (tou...
void computeKnownBitsForTargetNode(const SDValue Op, KnownBits &Known, const APInt &DemandedElts, const SelectionDAG &DAG, unsigned Depth=0) const override
Determine which of the bits specified in Mask are known to be either zero or one and return them in t...
SDValue lowerSET_ROUNDING(SDValue Op, SelectionDAG &DAG) const
void allocateSpecialInputVGPRsFixed(CCState &CCInfo, MachineFunction &MF, const SIRegisterInfo &TRI, SIMachineFunctionInfo &Info) const
Allocate implicit function VGPR arguments in fixed registers.
LoadInst * lowerIdempotentRMWIntoFencedLoad(AtomicRMWInst *AI) const override
On some platforms, an AtomicRMW that never actually modifies the value (such as fetch_add of 0) can b...
MachineBasicBlock * emitGWSMemViolTestLoop(MachineInstr &MI, MachineBasicBlock *BB) const
bool getAddrModeArguments(const IntrinsicInst *I, SmallVectorImpl< Value * > &Ops, Type *&AccessTy) const override
CodeGenPrepare sinks address calculations into the same BB as Load/Store instructions reading the add...
bool checkAsmConstraintValA(SDValue Op, uint64_t Val, unsigned MaxSize=64) const
bool shouldEmitFixup(const GlobalValue *GV) const
MachineBasicBlock * splitKillBlock(MachineInstr &MI, MachineBasicBlock *BB) const
void emitExpandAtomicCmpXchg(AtomicCmpXchgInst *CI) const override
Perform a cmpxchg expansion using a target-specific method.
bool canTransformPtrArithOutOfBounds(const Function &F, EVT PtrVT) const override
True if the target allows transformations of in-bounds pointer arithmetic that cause out-of-bounds in...
bool hasMemSDNodeUser(SDNode *N) const
bool isSDNodeSourceOfDivergence(const SDNode *N, FunctionLoweringInfo *FLI, UniformityInfo *UA) const override
MachineBasicBlock * EmitInstrWithCustomInserter(MachineInstr &MI, MachineBasicBlock *BB) const override
This method should be implemented by targets that mark instructions with the 'usesCustomInserter' fla...
bool isEligibleForTailCallOptimization(SDValue Callee, CallingConv::ID CalleeCC, bool isVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, const SmallVectorImpl< SDValue > &OutVals, const SmallVectorImpl< ISD::InputArg > &Ins, SelectionDAG &DAG) const
bool isMemOpHasNoClobberedMemOperand(const SDNode *N) const
bool isLegalFlatAddressingMode(const AddrMode &AM, unsigned AddrSpace) const
SDValue LowerCallResult(SDValue Chain, SDValue InGlue, CallingConv::ID CallConv, bool isVarArg, const SmallVectorImpl< ISD::InputArg > &Ins, const SDLoc &DL, SelectionDAG &DAG, SmallVectorImpl< SDValue > &InVals, bool isThisReturn, SDValue ThisVal) const
SDValue LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, const SmallVectorImpl< ISD::InputArg > &Ins, const SDLoc &DL, SelectionDAG &DAG, SmallVectorImpl< SDValue > &InVals) const override
This hook must be implemented to lower the incoming (formal) arguments, described by the Ins array,...
SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override
This method will be invoked for all target nodes and for any target-independent nodes that the target...
bool isFPExtFoldable(const SelectionDAG &DAG, unsigned Opcode, EVT DestVT, EVT SrcVT) const override
Return true if an fpext operation input to an Opcode operation is free (for instance,...
void AdjustInstrPostInstrSelection(MachineInstr &MI, SDNode *Node) const override
Assign the register class depending on the number of bits set in the writemask.
MVT getRegisterTypeForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const override
Certain combinations of ABIs, Targets and features require that types are legal for some operations a...
void allocateSpecialInputVGPRs(CCState &CCInfo, MachineFunction &MF, const SIRegisterInfo &TRI, SIMachineFunctionInfo &Info) const
Allocate implicit function VGPR arguments at the end of allocated user arguments.
void finalizeLowering(MachineFunction &MF) const override
Execute target specific actions to finalize target lowering.
static bool isNonGlobalAddrSpace(unsigned AS)
void emitExpandAtomicAddrSpacePredicate(Instruction *AI) const
MachineSDNode * buildRSRC(SelectionDAG &DAG, const SDLoc &DL, SDValue Ptr, uint32_t RsrcDword1, uint64_t RsrcDword2And3) const
Return a resource descriptor with the 'Add TID' bit enabled The TID (Thread ID) is multiplied by the ...
unsigned getNumRegistersForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const override
Certain targets require unusual breakdowns of certain types.
bool mayBeEmittedAsTailCall(const CallInst *) const override
Return true if the target may be able emit the call instruction as a tail call.
void passSpecialInputs(CallLoweringInfo &CLI, CCState &CCInfo, const SIMachineFunctionInfo &Info, SmallVectorImpl< std::pair< unsigned, SDValue > > &RegsToPass, SmallVectorImpl< SDValue > &MemOpChains, SDValue Chain) const
SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override
This callback is invoked for operations that are unsupported by the target, which are registered to u...
bool checkAsmConstraintVal(SDValue Op, StringRef Constraint, uint64_t Val) const
bool isKnownNeverNaNForTargetNode(SDValue Op, const APInt &DemandedElts, const SelectionDAG &DAG, bool SNaN=false, unsigned Depth=0) const override
If SNaN is false,.
void emitExpandAtomicRMW(AtomicRMWInst *AI) const override
Perform a atomicrmw expansion using a target-specific way.
static bool shouldExpandVectorDynExt(unsigned EltSize, unsigned NumElem, bool IsDivergentIdx, const GCNSubtarget *Subtarget)
Check if EXTRACT_VECTOR_ELT/INSERT_VECTOR_ELT (<n x e>, var-idx) should be expanded into a set of cmp...
bool shouldUseLDSConstAddress(const GlobalValue *GV) const
bool supportSplitCSR(MachineFunction *MF) const override
Return true if the target supports that a subset of CSRs for the given machine function is handled ex...
bool isExtractVecEltCheap(EVT VT, unsigned Index) const override
Return true if extraction of a scalar element from the given vector type at the given index is cheap.
SDValue LowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG) const
bool allowsMisalignedMemoryAccesses(LLT Ty, unsigned AddrSpace, Align Alignment, MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *IsFast=nullptr) const override
LLT handling variant.
bool isExtractSubvectorCheap(EVT ResVT, EVT SrcVT, unsigned Index) const override
Return true if EXTRACT_SUBVECTOR is cheap for extracting this result type from this source type with ...
unsigned combineRepeatedFPDivisors() const override
Indicate whether this target prefers to combine FDIVs with the same divisor.
bool canMergeStoresTo(unsigned AS, EVT MemVT, const MachineFunction &MF) const override
Returns if it's reasonable to merge stores to MemVT size.
SDValue lowerPREFETCH(SDValue Op, SelectionDAG &DAG) const
SITargetLowering(const TargetMachine &tm, const GCNSubtarget &STI)
void computeKnownBitsForTargetInstr(GISelValueTracking &Analysis, Register R, KnownBits &Known, const APInt &DemandedElts, const MachineRegisterInfo &MRI, unsigned Depth=0) const override
Determine which of the bits specified in Mask are known to be either zero or one and return them in t...
bool isFreeAddrSpaceCast(unsigned SrcAS, unsigned DestAS) const override
Returns true if a cast from SrcAS to DestAS is "cheap", such that e.g.
bool shouldEmitPCReloc(const GlobalValue *GV) const
AtomicExpansionKind shouldExpandAtomicCmpXchgInIR(const AtomicCmpXchgInst *AI) const override
Returns how the given atomic cmpxchg should be expanded by the IR-level AtomicExpand pass.
void initializeSplitCSR(MachineBasicBlock *Entry) const override
Perform necessary initialization to handle a subset of CSRs explicitly via copies.
void allocateSpecialEntryInputVGPRs(CCState &CCInfo, MachineFunction &MF, const SIRegisterInfo &TRI, SIMachineFunctionInfo &Info) const
void allocatePreloadKernArgSGPRs(CCState &CCInfo, SmallVectorImpl< CCValAssign > &ArgLocs, const SmallVectorImpl< ISD::InputArg > &Ins, MachineFunction &MF, const SIRegisterInfo &TRI, SIMachineFunctionInfo &Info) const
SDValue copyToM0(SelectionDAG &DAG, SDValue Chain, const SDLoc &DL, SDValue V) const
SDValue splitBinaryVectorOp(SDValue Op, SelectionDAG &DAG) const
MachinePointerInfo getKernargSegmentPtrInfo(MachineFunction &MF) const
unsigned getVectorTypeBreakdownForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT, EVT &IntermediateVT, unsigned &NumIntermediates, MVT &RegisterVT) const override
Certain targets such as MIPS require that some types such as vectors are always broken down into scal...
MVT getPointerMemTy(const DataLayout &DL, unsigned AS) const override
Similarly, the in-memory representation of a p7 is {p8, i32}, aka v8i32 when padding is added.
void allocateSystemSGPRs(CCState &CCInfo, MachineFunction &MF, SIMachineFunctionInfo &Info, CallingConv::ID CallConv, bool IsShader) const
bool CanLowerReturn(CallingConv::ID CallConv, MachineFunction &MF, bool isVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, LLVMContext &Context, const Type *RetTy) const override
This hook should be implemented to check whether the return values described by the Outs array can fi...
unsigned getMaxPermittedBytesForAlignment(MachineBasicBlock *MBB) const override
Return the maximum amount of bytes allowed to be emitted when padding for alignment.
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
Represent a constant reference to a string, i.e.
Definition StringRef.h:56
LegalizeTypeAction
This enum indicates whether a types are legal for a target, and if not, what action should be used to...
AtomicExpansionKind
Enum that specifies what an atomic load/AtomicRMWInst is expanded to, if at all.
Primary interface to the complete machine description for the target machine.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:46
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
This is an optimization pass for GlobalISel generic memory operations.
GenericUniformityInfo< SSAContext > UniformityInfo
@ Offset
Definition DWP.cpp:557
bool isBoolSGPR(SDValue V)
DWARFExpression::Operation Op
#define N
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
Extended Value Type.
Definition ValueTypes.h:35
InputArg - This struct carries flags and type information about a single incoming (formal) argument o...
This class contains a discriminated union of information about pointers in memory operands,...
These are IR-level optimization flags that may be propagated to SDNodes.
This represents a list of ValueType's that has been intern'd by a SelectionDAG.
This represents an addressing mode of: BaseGV + BaseOffs + BaseReg + Scale*ScaleReg + ScalableOffset*...