22#include "llvm/IR/IntrinsicsAMDGPU.h"
23#include "llvm/IR/IntrinsicsR600.h"
28#include "R600GenCallingConv.inc"
108 {MVT::f32, MVT::v2f32, MVT::v3f32, MVT::v4f32, MVT::v5f32,
109 MVT::v6f32, MVT::v7f32, MVT::v8f32, MVT::v16f32},
126 if (Subtarget->hasCARRY())
129 if (Subtarget->hasBORROW())
133 if (!Subtarget->hasBFE())
138 if (!Subtarget->hasBFE())
142 if (!Subtarget->hasBFE())
154 {MVT::v2i32, MVT::v2f32, MVT::v4i32, MVT::v4f32},
Custom);
157 {MVT::v2i32, MVT::v2f32, MVT::v4i32, MVT::v4f32},
Custom);
164 if (!Subtarget->hasFMA())
170 if (!Subtarget->hasBFI())
174 if (!Subtarget->hasBCNT(32))
177 if (!Subtarget->hasBCNT(64))
180 if (Subtarget->hasFFBH())
183 if (Subtarget->hasFFBL())
188 if (Subtarget->hasBFE())
194 const MVT ScalarIntVTs[] = { MVT::i32, MVT::i64 };
195 for (
MVT VT : ScalarIntVTs)
214 if (std::next(
I) ==
I->getParent()->end())
216 return std::next(
I)->getOpcode() == R600::RETURN;
227 switch (
MI.getOpcode()) {
231 if (
TII->isLDSRetInstr(
MI.getOpcode())) {
232 int DstIdx =
TII->getOperandIdx(
MI.getOpcode(), R600::OpName::dst);
237 if (!
MRI.use_empty(
MI.getOperand(DstIdx).getReg()) ||
238 MI.getOpcode() == R600::LDS_CMPST_RET)
250 case R600::FABS_R600: {
252 *BB,
I, R600::MOV,
MI.getOperand(0).getReg(),
253 MI.getOperand(1).getReg());
258 case R600::FNEG_R600: {
260 *BB,
I, R600::MOV,
MI.getOperand(0).getReg(),
261 MI.getOperand(1).getReg());
266 case R600::MASK_WRITE: {
267 Register maskedRegister =
MI.getOperand(0).getReg();
274 case R600::MOV_IMM_F32:
275 TII->buildMovImm(*BB,
I,
MI.getOperand(0).getReg(),
MI.getOperand(1)
282 case R600::MOV_IMM_I32:
283 TII->buildMovImm(*BB,
I,
MI.getOperand(0).getReg(),
284 MI.getOperand(1).getImm());
287 case R600::MOV_IMM_GLOBAL_ADDR: {
289 auto MIB =
TII->buildDefaultInstruction(
290 *BB,
MI, R600::MOV,
MI.getOperand(0).getReg(), R600::ALU_LITERAL_X);
291 int Idx =
TII->getOperandIdx(*MIB, R600::OpName::literal);
299 case R600::CONST_COPY: {
301 *BB,
MI, R600::MOV,
MI.getOperand(0).getReg(), R600::ALU_CONST);
302 TII->setImmOperand(*NewMI, R600::OpName::src0_sel,
303 MI.getOperand(1).getImm());
307 case R600::RAT_WRITE_CACHELESS_32_eg:
308 case R600::RAT_WRITE_CACHELESS_64_eg:
309 case R600::RAT_WRITE_CACHELESS_128_eg:
311 .
add(
MI.getOperand(0))
312 .
add(
MI.getOperand(1))
316 case R600::RAT_STORE_TYPED_eg:
318 .
add(
MI.getOperand(0))
319 .
add(
MI.getOperand(1))
320 .
add(
MI.getOperand(2))
326 .
add(
MI.getOperand(0));
329 case R600::BRANCH_COND_f32: {
333 .
add(
MI.getOperand(1))
338 .
add(
MI.getOperand(0))
343 case R600::BRANCH_COND_i32: {
347 .
add(
MI.getOperand(1))
348 .
addImm(R600::PRED_SETNE_INT)
352 .
add(
MI.getOperand(0))
357 case R600::EG_ExportSwz:
358 case R600::R600_ExportSwz: {
360 bool isLastInstructionOfItsType =
true;
361 unsigned InstExportType =
MI.getOperand(1).getImm();
363 EndBlock = BB->
end(); NextExportInst != EndBlock;
364 NextExportInst = std::next(NextExportInst)) {
365 if (NextExportInst->getOpcode() == R600::EG_ExportSwz ||
366 NextExportInst->getOpcode() == R600::R600_ExportSwz) {
367 unsigned CurrentInstExportType = NextExportInst->getOperand(1)
369 if (CurrentInstExportType == InstExportType) {
370 isLastInstructionOfItsType =
false;
376 if (!EOP && !isLastInstructionOfItsType)
378 unsigned CfInst = (
MI.getOpcode() == R600::EG_ExportSwz) ? 84 : 40;
380 .
add(
MI.getOperand(0))
381 .
add(
MI.getOperand(1))
382 .
add(
MI.getOperand(2))
383 .
add(
MI.getOperand(3))
384 .
add(
MI.getOperand(4))
385 .
add(
MI.getOperand(5))
386 .
add(
MI.getOperand(6))
396 MI.eraseFromParent();
407 switch (
Op.getOpcode()) {
417 case ISD::FSIN:
return LowerTrig(
Op, DAG);
419 case ISD::STORE:
return LowerSTORE(
Op, DAG);
422 assert((!Result.getNode() ||
423 Result.getNode()->getNumValues() == 2) &&
424 "Load should return a value and a chain");
428 case ISD::BRCOND:
return LowerBRCOND(
Op, DAG);
431 case ISD::ADDRSPACECAST:
432 return lowerADDRSPACECAST(
Op, DAG);
435 unsigned IntrinsicID =
Op.getConstantOperandVal(1);
436 switch (IntrinsicID) {
437 case Intrinsic::r600_store_swizzle: {
459 unsigned IntrinsicID =
Op.getConstantOperandVal(0);
460 EVT VT =
Op.getValueType();
462 switch (IntrinsicID) {
463 case Intrinsic::r600_tex:
464 case Intrinsic::r600_texc: {
466 switch (IntrinsicID) {
467 case Intrinsic::r600_tex:
470 case Intrinsic::r600_texc:
500 case Intrinsic::r600_dot4: {
522 case Intrinsic::r600_implicitarg_ptr: {
527 case Intrinsic::r600_read_ngroups_x:
528 return LowerImplicitParameter(DAG, VT,
DL, 0);
529 case Intrinsic::r600_read_ngroups_y:
530 return LowerImplicitParameter(DAG, VT,
DL, 1);
531 case Intrinsic::r600_read_ngroups_z:
532 return LowerImplicitParameter(DAG, VT,
DL, 2);
533 case Intrinsic::r600_read_global_size_x:
534 return LowerImplicitParameter(DAG, VT,
DL, 3);
535 case Intrinsic::r600_read_global_size_y:
536 return LowerImplicitParameter(DAG, VT,
DL, 4);
537 case Intrinsic::r600_read_global_size_z:
538 return LowerImplicitParameter(DAG, VT,
DL, 5);
539 case Intrinsic::r600_read_local_size_x:
540 return LowerImplicitParameter(DAG, VT,
DL, 6);
541 case Intrinsic::r600_read_local_size_y:
542 return LowerImplicitParameter(DAG, VT,
DL, 7);
543 case Intrinsic::r600_read_local_size_z:
544 return LowerImplicitParameter(DAG, VT,
DL, 8);
546 case Intrinsic::r600_read_tgid_x:
547 case Intrinsic::amdgcn_workgroup_id_x:
550 case Intrinsic::r600_read_tgid_y:
551 case Intrinsic::amdgcn_workgroup_id_y:
554 case Intrinsic::r600_read_tgid_z:
555 case Intrinsic::amdgcn_workgroup_id_z:
558 case Intrinsic::r600_read_tidig_x:
559 case Intrinsic::amdgcn_workitem_id_x:
562 case Intrinsic::r600_read_tidig_y:
563 case Intrinsic::amdgcn_workitem_id_y:
566 case Intrinsic::r600_read_tidig_z:
567 case Intrinsic::amdgcn_workitem_id_z:
571 case Intrinsic::r600_recipsqrt_ieee:
574 case Intrinsic::r600_recipsqrt_clamped:
590 switch (
N->getOpcode()) {
595 if (
N->getValueType(0) == MVT::i1) {
596 Results.push_back(lowerFP_TO_UINT(
N->getOperand(0), DAG));
604 if (
N->getValueType(0) == MVT::i1) {
605 Results.push_back(lowerFP_TO_SINT(
N->getOperand(0), DAG));
673 return vectorToVerticalVector(DAG, Insert);
684 const GlobalValue *GV = GSD->
getGlobal();
694 EVT VT =
Op.getValueType();
705 switch (
Op.getOpcode()) {
733 unsigned mainop,
unsigned ovf)
const {
735 EVT VT =
Op.getValueType();
772 unsigned DwordOffset)
const {
773 unsigned ByteOffset = DwordOffset * 4;
785bool R600TargetLowering::isZero(
SDValue Op)
const {
787 return Cst->isZero();
789 return CstFP->isZero();
793bool R600TargetLowering::isHWTrueValue(
SDValue Op)
const {
795 return CFP->isExactlyValue(1.0);
800bool R600TargetLowering::isHWFalseValue(
SDValue Op)
const {
802 return CFP->getValueAPF().isZero();
809 EVT VT =
Op.getValueType();
818 if (VT == MVT::f32) {
826 EVT CompareVT =
LHS.getValueType();
840 if (isHWTrueValue(False) && isHWFalseValue(True)) {
856 if (isHWTrueValue(True) && isHWFalseValue(False) &&
857 (CompareVT == VT || VT == MVT::i32)) {
895 if (CompareVT != VT) {
900 True = DAG.
getNode(ISD::BITCAST,
DL, CompareVT, True);
901 False = DAG.
getNode(ISD::BITCAST,
DL, CompareVT, False);
920 return DAG.
getNode(ISD::BITCAST,
DL, VT, SelectNode);
927 if (CompareVT == MVT::f32) {
930 }
else if (CompareVT == MVT::i32) {
951 EVT VT =
Op.getValueType();
953 const R600TargetMachine &TM =
993void R600TargetLowering::getStackAddress(
unsigned StackWidth,
996 unsigned &PtrIncr)
const {
997 switch (StackWidth) {
1008 Channel = ElemIdx % 2;
1027 ||
Store->getValue().getValueType() == MVT::i8);
1031 if (
Store->getMemoryVT() == MVT::i8) {
1034 }
else if (
Store->getMemoryVT() == MVT::i16) {
1047 EVT MemVT =
Store->getMemoryVT();
1064 Chain = Dst.getValue(1);
1084 MaskedValue, ShiftAmt);
1091 DstMask = DAG.
getNOT(
DL, DstMask, MVT::i32);
1120 EVT VT =
Value.getValueType();
1122 EVT PtrVT =
Ptr.getValueType();
1160 if (TruncatingStore) {
1163 if (MemVT == MVT::i8) {
1166 assert(MemVT == MVT::i16);
1194 Op->getVTList(), Args, MemVT,
1214 if (MemVT.
bitsLT(MVT::i32))
1215 return lowerPrivateTruncStore(StoreNode, DAG);
1237 return 512 + 4096 * 2;
1239 return 512 + 4096 * 3;
1241 return 512 + 4096 * 4;
1243 return 512 + 4096 * 5;
1245 return 512 + 4096 * 6;
1247 return 512 + 4096 * 7;
1249 return 512 + 4096 * 8;
1251 return 512 + 4096 * 9;
1253 return 512 + 4096 * 10;
1255 return 512 + 4096 * 11;
1257 return 512 + 4096 * 12;
1259 return 512 + 4096 * 13;
1261 return 512 + 4096 * 14;
1263 return 512 + 4096 * 15;
1274 EVT MemVT =
Load->getMemoryVT();
1333 return lowerPrivateExtLoad(
Op, DAG);
1337 EVT VT =
Op.getValueType();
1351 if (ConstantBlock > -1 &&
1388 assert(!MemVT.
isVector() && (MemVT == MVT::i16 || MemVT == MVT::i8));
1395 SDValue MergedValues[2] = { Res, Chain };
1425 const R600FrameLowering *TFL = Subtarget->getFrameLowering();
1434 SDLoc(
Op),
Op.getValueType());
1438 bool IsVarArg)
const {
1477 for (
unsigned i = 0, e = Ins.size(); i < e; ++i) {
1549 unsigned *IsFast)
const {
1553 if (!VT.
isSimple() || VT == MVT::Other)
1563 return VT.
bitsGT(MVT::i32) && Alignment >=
Align(4);
1575 for (
unsigned i = 0; i < 4; i++)
1579 for (
unsigned i = 0; i < 4; i++) {
1584 RemapSwizzle[i] = 7;
1587 RemapSwizzle[i] = 4;
1588 NewBldVec[i] = DAG.
getUNDEF(MVT::f32);
1589 }
else if (
C->isExactlyValue(1.0)) {
1590 RemapSwizzle[i] = 5;
1591 NewBldVec[i] = DAG.
getUNDEF(MVT::f32);
1598 for (
unsigned j = 0; j < i; j++) {
1599 if (NewBldVec[i] == NewBldVec[j]) {
1601 RemapSwizzle[i] = j;
1619 bool isUnmovable[4] = {
false,
false,
false,
false};
1620 for (
unsigned i = 0; i < 4; i++)
1624 for (
unsigned i = 0; i < 4; i++) {
1625 RemapSwizzle[i] = i;
1629 isUnmovable[Idx] =
true;
1633 for (
unsigned i = 0; i < 4; i++) {
1636 if (isUnmovable[Idx])
1639 std::swap(NewBldVec[Idx], NewBldVec[i]);
1640 std::swap(RemapSwizzle[i], RemapSwizzle[Idx]);
1653 DenseMap<unsigned, unsigned> SwizzleRemap;
1656 for (
unsigned i = 0; i < 4; i++) {
1658 auto It = SwizzleRemap.
find(Idx);
1659 if (It != SwizzleRemap.
end())
1663 SwizzleRemap.
clear();
1665 for (
unsigned i = 0; i < 4; i++) {
1667 auto It = SwizzleRemap.
find(Idx);
1668 if (It != SwizzleRemap.
end())
1693 for (
unsigned i = 0; i < 4; i++) {
1703 EVT NewVT = MVT::v4i32;
1704 unsigned NumElements = 4;
1730 switch (
N->getOpcode()) {
1805 if (Elt <
Ops.size()) {
1808 EVT OpVT =
Ops[0].getValueType();
1826 unsigned Element = Const->getZExtValue();
1835 unsigned Element = Const->getZExtValue();
1836 return DAG.
getNode(ISD::BITCAST,
DL,
N->getVTList(),
1863 if (LHS.getOperand(2).getNode() != True.
getNode() ||
1865 RHS.getNode() != False.
getNode()) {
1904 NewArgs[1] = OptimizeSwizzle(
N->getOperand(1), &NewArgs[4], DAG,
DL);
1933 NewArgs[1] = OptimizeSwizzle(
N->getOperand(1), &NewArgs[2], DAG,
DL);
1952bool R600TargetLowering::FoldOperand(
SDNode *ParentNode,
unsigned SrcIdx,
1957 if (!Src.isMachineOpcode())
1960 switch (Src.getMachineOpcode()) {
1961 case R600::FNEG_R600:
1964 Src = Src.getOperand(0);
1967 case R600::FABS_R600:
1973 case R600::CONST_COPY: {
1975 bool HasDst =
TII->getOperandIdx(Opcode, R600::OpName::dst) > -1;
1985 int SrcIndices[] = {
1986 TII->getOperandIdx(Opcode, R600::OpName::src0),
1987 TII->getOperandIdx(Opcode, R600::OpName::src1),
1988 TII->getOperandIdx(Opcode, R600::OpName::src2),
1989 TII->getOperandIdx(Opcode, R600::OpName::src0_X),
1990 TII->getOperandIdx(Opcode, R600::OpName::src0_Y),
1991 TII->getOperandIdx(Opcode, R600::OpName::src0_Z),
1992 TII->getOperandIdx(Opcode, R600::OpName::src0_W),
1993 TII->getOperandIdx(Opcode, R600::OpName::src1_X),
1994 TII->getOperandIdx(Opcode, R600::OpName::src1_Y),
1995 TII->getOperandIdx(Opcode, R600::OpName::src1_Z),
1996 TII->getOperandIdx(Opcode, R600::OpName::src1_W)
1998 std::vector<unsigned> Consts;
1999 for (
int OtherSrcIdx : SrcIndices) {
2000 int OtherSelIdx =
TII->getSelIdx(Opcode, OtherSrcIdx);
2001 if (OtherSrcIdx < 0 || OtherSelIdx < 0)
2007 if (RegisterSDNode *
Reg =
2009 if (
Reg->getReg() == R600::ALU_CONST) {
2017 if (!
TII->fitsConstReadLimitations(Consts)) {
2025 case R600::MOV_IMM_GLOBAL_ADDR:
2027 if (
Imm->getAsZExtVal())
2029 Imm = Src.getOperand(0);
2030 Src = DAG.
getRegister(R600::ALU_LITERAL_X, MVT::i32);
2032 case R600::MOV_IMM_I32:
2033 case R600::MOV_IMM_F32: {
2034 unsigned ImmReg = R600::ALU_LITERAL_X;
2035 uint64_t ImmValue = 0;
2037 if (Src.getMachineOpcode() == R600::MOV_IMM_F32) {
2040 if (FloatValue == 0.0) {
2041 ImmReg = R600::ZERO;
2042 }
else if (FloatValue == 0.5) {
2043 ImmReg = R600::HALF;
2044 }
else if (FloatValue == 1.0) {
2050 uint64_t
Value = Src.getConstantOperandVal(0);
2052 ImmReg = R600::ZERO;
2053 }
else if (
Value == 1) {
2054 ImmReg = R600::ONE_INT;
2063 if (ImmReg == R600::ALU_LITERAL_X) {
2067 if (
C->getZExtValue())
2082 const R600InstrInfo *
TII = Subtarget->getInstrInfo();
2083 if (!
Node->isMachineOpcode())
2086 unsigned Opcode =
Node->getMachineOpcode();
2089 std::vector<SDValue>
Ops(
Node->op_begin(),
Node->op_end());
2091 if (Opcode == R600::DOT_4) {
2092 int OperandIdx[] = {
2093 TII->getOperandIdx(Opcode, R600::OpName::src0_X),
2094 TII->getOperandIdx(Opcode, R600::OpName::src0_Y),
2095 TII->getOperandIdx(Opcode, R600::OpName::src0_Z),
2096 TII->getOperandIdx(Opcode, R600::OpName::src0_W),
2097 TII->getOperandIdx(Opcode, R600::OpName::src1_X),
2098 TII->getOperandIdx(Opcode, R600::OpName::src1_Y),
2099 TII->getOperandIdx(Opcode, R600::OpName::src1_Z),
2100 TII->getOperandIdx(Opcode, R600::OpName::src1_W)
2103 TII->getOperandIdx(Opcode, R600::OpName::src0_neg_X),
2104 TII->getOperandIdx(Opcode, R600::OpName::src0_neg_Y),
2105 TII->getOperandIdx(Opcode, R600::OpName::src0_neg_Z),
2106 TII->getOperandIdx(Opcode, R600::OpName::src0_neg_W),
2107 TII->getOperandIdx(Opcode, R600::OpName::src1_neg_X),
2108 TII->getOperandIdx(Opcode, R600::OpName::src1_neg_Y),
2109 TII->getOperandIdx(Opcode, R600::OpName::src1_neg_Z),
2110 TII->getOperandIdx(Opcode, R600::OpName::src1_neg_W)
2113 TII->getOperandIdx(Opcode, R600::OpName::src0_abs_X),
2114 TII->getOperandIdx(Opcode, R600::OpName::src0_abs_Y),
2115 TII->getOperandIdx(Opcode, R600::OpName::src0_abs_Z),
2116 TII->getOperandIdx(Opcode, R600::OpName::src0_abs_W),
2117 TII->getOperandIdx(Opcode, R600::OpName::src1_abs_X),
2118 TII->getOperandIdx(Opcode, R600::OpName::src1_abs_Y),
2119 TII->getOperandIdx(Opcode, R600::OpName::src1_abs_Z),
2120 TII->getOperandIdx(Opcode, R600::OpName::src1_abs_W)
2122 for (
unsigned i = 0; i < 8; i++) {
2123 if (OperandIdx[i] < 0)
2128 bool HasDst =
TII->getOperandIdx(Opcode, R600::OpName::dst) > -1;
2129 int SelIdx =
TII->getSelIdx(Opcode, OperandIdx[i]);
2132 SDValue &Sel = (SelIdx > -1) ?
Ops[SelIdx] : FakeOp;
2133 if (FoldOperand(Node, i, Src, Neg, Abs, Sel, FakeOp, DAG))
2136 }
else if (Opcode == R600::REG_SEQUENCE) {
2137 for (
unsigned i = 1, e =
Node->getNumOperands(); i < e; i += 2) {
2139 if (FoldOperand(Node, i, Src, FakeOp, FakeOp, FakeOp, FakeOp, DAG))
2143 if (!
TII->hasInstrModifiers(Opcode))
2145 int OperandIdx[] = {
2146 TII->getOperandIdx(Opcode, R600::OpName::src0),
2147 TII->getOperandIdx(Opcode, R600::OpName::src1),
2148 TII->getOperandIdx(Opcode, R600::OpName::src2)
2151 TII->getOperandIdx(Opcode, R600::OpName::src0_neg),
2152 TII->getOperandIdx(Opcode, R600::OpName::src1_neg),
2153 TII->getOperandIdx(Opcode, R600::OpName::src2_neg)
2156 TII->getOperandIdx(Opcode, R600::OpName::src0_abs),
2157 TII->getOperandIdx(Opcode, R600::OpName::src1_abs),
2160 for (
unsigned i = 0; i < 3; i++) {
2161 if (OperandIdx[i] < 0)
2166 SDValue &Abs = (AbsIdx[i] > -1) ?
Ops[AbsIdx[i] - 1] : FakeAbs;
2167 bool HasDst =
TII->getOperandIdx(Opcode, R600::OpName::dst) > -1;
2168 int SelIdx =
TII->getSelIdx(Opcode, OperandIdx[i]);
2169 int ImmIdx =
TII->getOperandIdx(Opcode, R600::OpName::literal);
2174 SDValue &Sel = (SelIdx > -1) ?
Ops[SelIdx] : FakeOp;
2176 if (FoldOperand(Node, i, Src, Neg, Abs, Sel, Imm, DAG))
2185R600TargetLowering::shouldExpandAtomicRMWInIR(
AtomicRMWInst *RMW)
const {
2200 unsigned ValSize =
DL.getTypeSizeInBits(RMW->
getType());
2201 if (ValSize == 32 || ValSize == 64)
2207 unsigned Size = IntTy->getBitWidth();
unsigned const MachineRegisterInfo * MRI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis Results
Interfaces for producing common pass manager configurations.
const HexagonInstrInfo * TII
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
static bool isUndef(const MachineInstr &MI)
Promote Memory to Register
static bool isEOP(MachineBasicBlock::iterator I)
static SDValue ReorganizeVector(SelectionDAG &DAG, SDValue VectorEntry, DenseMap< unsigned, unsigned > &RemapSwizzle)
static int ConstantAddressBlock(unsigned AddressSpace)
static SDValue CompactSwizzlableVector(SelectionDAG &DAG, SDValue VectorEntry, DenseMap< unsigned, unsigned > &RemapSwizzle)
R600 DAG Lowering interface definition.
Provides R600 specific target descriptions.
AMDGPU R600 specific subclass of TargetSubtarget.
The AMDGPU TargetMachine interface definition for hw codegen targets.
const SmallVectorImpl< MachineOperand > & Cond
static Type * getValueType(Value *V)
Returns the type of the given value/instruction V.
static std::optional< unsigned > getOpcode(ArrayRef< VPValue * > Values)
Returns the opcode of Values or ~0 if they do not all agree.
unsigned getStackWidth(const MachineFunction &MF) const
SDValue combineFMinMaxLegacy(const SDLoc &DL, EVT VT, SDValue LHS, SDValue RHS, SDValue True, SDValue False, SDValue CC, DAGCombinerInfo &DCI) const
Generate Min/Max node.
void analyzeFormalArgumentsCompute(CCState &State, const SmallVectorImpl< ISD::InputArg > &Ins) const
The SelectionDAGBuilder will automatically promote function arguments with illegal types.
SDValue LowerSDIVREM(SDValue Op, SelectionDAG &DAG) const
SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override
This callback is invoked for operations that are unsupported by the target, which are registered to u...
uint32_t getImplicitParameterOffset(const MachineFunction &MF, const ImplicitParameter Param) const
Helper function that returns the byte offset of the given type of implicit parameter.
virtual SDValue LowerGlobalAddress(AMDGPUMachineFunction *MFI, SDValue Op, SelectionDAG &DAG) const
SDValue CreateLiveInRegisterRaw(SelectionDAG &DAG, const TargetRegisterClass *RC, Register Reg, EVT VT) const
AMDGPUTargetLowering(const TargetMachine &TM, const AMDGPUSubtarget &STI)
void ReplaceNodeResults(SDNode *N, SmallVectorImpl< SDValue > &Results, SelectionDAG &DAG) const override
This callback is invoked when a node result type is illegal for the target, and the operation was reg...
SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override
This method will be invoked for all target nodes and for any target-independent nodes that the target...
void LowerUDIVREM64(SDValue Op, SelectionDAG &DAG, SmallVectorImpl< SDValue > &Results) const
LLVM_ABI float convertToFloat() const
Converts this APFloat to host float value.
APInt bitcastToAPInt() const
uint64_t getZExtValue() const
Get zero extended value.
unsigned getSrcAddressSpace() const
unsigned getDestAddressSpace() const
an instruction that atomically reads a memory location, combines it with another value,...
@ UIncWrap
Increment one up to a maximum value.
@ FMin
*p = minnum(old, v) minnum matches the behavior of llvm.minnum.
@ FMax
*p = maxnum(old, v) maxnum matches the behavior of llvm.maxnum.
@ UDecWrap
Decrement one until a minimum value or zero.
BinOp getOperation() const
CCState - This class holds information needed while lowering arguments and return values.
LLVM_ABI void AnalyzeFormalArguments(const SmallVectorImpl< ISD::InputArg > &Ins, CCAssignFn Fn)
AnalyzeFormalArguments - Analyze an array of argument values, incorporating info about the formals in...
CCValAssign - Represent assignment of one arg/retval to a location.
Register getLocReg() const
int64_t getLocMemOffset() const
const APFloat & getValueAPF() const
static LLVM_ABI ConstantPointerNull * get(PointerType *T)
Static factory methods - Return objects of the specified value.
uint64_t getZExtValue() const
A parsed version of the target data layout string in and methods for querying it.
iterator find(const_arg_type_t< KeyT > Val)
const DataLayout & getDataLayout() const
Get the data layout of the module this function belongs to.
LLVM_ABI unsigned getAddressSpace() const
const GlobalValue * getGlobal() const
LLVM_ABI const Function * getFunction() const
Return the function this instruction belongs to.
This is an important class for using LLVM in a threaded context.
bool isIndexed() const
Return true if this is a pre/post inc/dec load/store.
This class is used to represent ISD::LOAD nodes.
const SDValue & getBasePtr() const
ISD::LoadExtType getExtensionType() const
Return whether this is a plain node, or one of the varieties of value-extending loads.
static auto integer_valuetypes()
LLVM_ABI DebugLoc findDebugLoc(instr_iterator MBBI)
Find the next valid DebugLoc starting at MBBI, skipping any debug instructions.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
MachineInstrBundleIterator< MachineInstr > iterator
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
Register addLiveIn(MCRegister PReg, const TargetRegisterClass *RC)
addLiveIn - Add the specified physical register as a live-in value and create a corresponding virtual...
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
Representation of each machine instruction.
Flags
Flags values. These may be or'd together.
@ MODereferenceable
The memory access is dereferenceable (i.e., doesn't trap).
@ MONonTemporal
The memory access is non-temporal.
@ MOInvariant
The memory access always returns the same value (or traps).
Flags getFlags() const
Return the raw flags of the source value,.
const Value * getValue() const
Return the base address of the memory access.
MachineOperand class - Representation of each machine instruction operand.
const GlobalValue * getGlobal() const
unsigned getTargetFlags() const
int64_t getOffset() const
Return the offset from the symbol in this operand.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
An SDNode that represents everything that will be needed to construct a MachineInstr.
unsigned getAddressSpace() const
Return the address space for the associated pointer.
AAMDNodes getAAInfo() const
Returns the AA info that describes the dereference.
MachineMemOperand * getMemOperand() const
Return a MachineMemOperand object describing the memory reference performed by operation.
const MachinePointerInfo & getPointerInfo() const
const SDValue & getChain() const
EVT getMemoryVT() const
Return the type of the in-memory value.
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
StackOffset getFrameIndexReference(const MachineFunction &MF, int FI, Register &FrameReg) const override
const R600InstrInfo * getInstrInfo() const override
SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override
This method will be invoked for all target nodes and for any target-independent nodes that the target...
MachineBasicBlock * EmitInstrWithCustomInserter(MachineInstr &MI, MachineBasicBlock *BB) const override
This method should be implemented by targets that mark instructions with the 'usesCustomInserter' fla...
bool canMergeStoresTo(unsigned AS, EVT MemVT, const MachineFunction &MF) const override
Returns if it's reasonable to merge stores to MemVT size.
R600TargetLowering(const TargetMachine &TM, const R600Subtarget &STI)
EVT getSetCCResultType(const DataLayout &DL, LLVMContext &, EVT VT) const override
Return the ValueType of the result of SETCC operations.
void ReplaceNodeResults(SDNode *N, SmallVectorImpl< SDValue > &Results, SelectionDAG &DAG) const override
This callback is invoked when a node result type is illegal for the target, and the operation was reg...
bool allowsMisalignedMemoryAccesses(EVT VT, unsigned AS, Align Alignment, MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *IsFast=nullptr) const override
Determine if the target supports unaligned memory accesses.
CCAssignFn * CCAssignFnForCall(CallingConv::ID CC, bool IsVarArg) const
SDValue LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, const SmallVectorImpl< ISD::InputArg > &Ins, const SDLoc &DL, SelectionDAG &DAG, SmallVectorImpl< SDValue > &InVals) const override
XXX Only kernel functions are supported, so we can assume for now that every function is a kernel fun...
SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override
This callback is invoked for operations that are unsupported by the target, which are registered to u...
Wrapper class representing virtual and physical registers.
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
uint64_t getAsZExtVal() const
Helper method returns the zero-extended integer value of a ConstantSDNode.
unsigned getMachineOpcode() const
This may only be called if isMachineOpcode returns true.
const SDValue & getOperand(unsigned Num) const
uint64_t getConstantOperandVal(unsigned Num) const
Helper method returns the integer value of a ConstantSDNode operand.
EVT getValueType(unsigned ResNo) const
Return the type of a specified result.
op_iterator op_end() const
op_iterator op_begin() const
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
const SDValue & getOperand(unsigned i) const
uint64_t getConstantOperandVal(unsigned i) const
unsigned getOpcode() const
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
LLVM_ABI SDValue getExtLoad(ISD::LoadExtType ExtType, const SDLoc &dl, EVT VT, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, EVT MemVT, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
SDValue getTargetGlobalAddress(const GlobalValue *GV, const SDLoc &DL, EVT VT, int64_t offset=0, unsigned TargetFlags=0)
LLVM_ABI SDValue getMergeValues(ArrayRef< SDValue > Ops, const SDLoc &dl)
Create a MERGE_VALUES node from the given operands.
LLVM_ABI SDVTList getVTList(EVT VT)
Return an SDVTList that represents the list of values specified.
LLVM_ABI SDValue getAllOnesConstant(const SDLoc &DL, EVT VT, bool IsTarget=false, bool IsOpaque=false)
LLVM_ABI MachineSDNode * getMachineNode(unsigned Opcode, const SDLoc &dl, EVT VT)
These are used for target selectors to create a new node with specified return type(s),...
LLVM_ABI SDValue getConstantFP(double Val, const SDLoc &DL, EVT VT, bool isTarget=false)
Create a ConstantFPSDNode wrapping a constant value.
LLVM_ABI SDValue getRegister(Register Reg, EVT VT)
LLVM_ABI SDValue getLoad(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr)
Loads are not normal binary operators: their result type is not determined by their operands,...
LLVM_ABI SDValue getMemIntrinsicNode(unsigned Opcode, const SDLoc &dl, SDVTList VTList, ArrayRef< SDValue > Ops, EVT MemVT, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags Flags=MachineMemOperand::MOLoad|MachineMemOperand::MOStore, LocationSize Size=LocationSize::precise(0), const AAMDNodes &AAInfo=AAMDNodes())
Creates a MemIntrinsicNode that may produce a result and takes a list of operands.
LLVM_ABI SDValue getNOT(const SDLoc &DL, SDValue Val, EVT VT)
Create a bitwise NOT operation as (XOR Val, -1).
SDValue getUNDEF(EVT VT)
Return an UNDEF node. UNDEF does not have a useful SDLoc.
SDValue getBuildVector(EVT VT, const SDLoc &DL, ArrayRef< SDValue > Ops)
Return an ISD::BUILD_VECTOR node.
SDValue getCopyFromReg(SDValue Chain, const SDLoc &dl, Register Reg, EVT VT)
LLVM_ABI SDValue getZeroExtendInReg(SDValue Op, const SDLoc &DL, EVT VT)
Return the expression required to zero extend the Op value assuming it was the smaller SrcTy value.
const DataLayout & getDataLayout() const
LLVM_ABI SDValue getConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
Create a ConstantSDNode wrapping a constant value.
LLVM_ABI SDValue getTruncStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, EVT SVT, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI SDValue getStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
Helper function to build ISD::STORE nodes.
LLVM_ABI SDValue getSignedConstant(int64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
SDValue getSelectCC(const SDLoc &DL, SDValue LHS, SDValue RHS, SDValue True, SDValue False, ISD::CondCode Cond, SDNodeFlags Flags=SDNodeFlags())
Helper function to make it easier to build SelectCC's if you just have an ISD::CondCode instead of an...
LLVM_ABI SDValue getIntPtrConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI SDValue getValueType(EVT)
LLVM_ABI SDValue getNode(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDUse > Ops)
Gets or creates the specified node.
SDValue getTargetConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isOpaque=false)
LLVM_ABI SDValue getVectorIdxConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI void ReplaceAllUsesOfValueWith(SDValue From, SDValue To)
Replace any uses of From with To, leaving uses of other values produced by From.getNode() alone.
MachineFunction & getMachineFunction() const
LLVM_ABI SDValue getCondCode(ISD::CondCode Cond)
LLVMContext * getContext() const
SDValue getEntryNode() const
Return the token chain corresponding to the entry of the function.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
This class is used to represent ISD::STORE nodes.
const SDValue & getBasePtr() const
const SDValue & getValue() const
bool isTruncatingStore() const
Return true if the op does a truncation before store.
void setBooleanVectorContents(BooleanContent Ty)
Specify how the target extends the result of a vector boolean value from a vector of i1 to a wider ty...
void setOperationAction(unsigned Op, MVT VT, LegalizeAction Action)
Indicate that the specified operation does not work with the specified type and indicate what to do a...
const TargetMachine & getTargetMachine() const
void setHasExtractBitsInsn(bool hasExtractInsn=true)
Tells the code generator that the target has BitExtract instructions.
void setBooleanContents(BooleanContent Ty)
Specify how the target extends the result of integer and floating point boolean values from i1 to a w...
void computeRegisterProperties(const TargetRegisterInfo *TRI)
Once all of the register classes are added, this allows us to compute derived properties we expose.
void addRegisterClass(MVT VT, const TargetRegisterClass *RC)
Add the specified register class as an available regclass for the specified value type.
bool isCondCodeLegal(ISD::CondCode CC, MVT VT) const
Return true if the specified condition code is legal for a comparison of the specified types on this ...
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
bool isOperationLegal(unsigned Op, EVT VT) const
Return true if the specified operation is legal on this target.
void setTruncStoreAction(MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified truncating store does not work with the specified type and indicate what ...
@ ZeroOrNegativeOneBooleanContent
AtomicExpansionKind
Enum that specifies what an atomic load/AtomicRMWInst is expanded to, if at all.
void setCondCodeAction(ArrayRef< ISD::CondCode > CCs, MVT VT, LegalizeAction Action)
Indicate that the specified condition code is or isn't supported on the target and indicate what to d...
void setTargetDAGCombine(ArrayRef< ISD::NodeType > NTs)
Targets should invoke this method for each target independent node that they want to provide a custom...
void setLoadExtAction(unsigned ExtType, MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified load with extension does not work with the specified type and indicate wh...
void setSchedulingPreference(Sched::Preference Pref)
Specify the target scheduling preference.
SDValue scalarizeVectorStore(StoreSDNode *ST, SelectionDAG &DAG) const
SDValue expandUnalignedStore(StoreSDNode *ST, SelectionDAG &DAG) const
Expands an unaligned store to 2 half-size stores for integer values, and possibly more for vectors.
bool expandFP_TO_SINT(SDNode *N, SDValue &Result, SelectionDAG &DAG) const
Expand float(f32) to SINT(i64) conversion.
std::pair< SDValue, SDValue > scalarizeVectorLoad(LoadSDNode *LD, SelectionDAG &DAG) const
Turn load of vector type into a load of the individual elements.
virtual MachineBasicBlock * EmitInstrWithCustomInserter(MachineInstr &MI, MachineBasicBlock *MBB) const
This method should be implemented by targets that mark instructions with the 'usesCustomInserter' fla...
void expandShiftParts(SDNode *N, SDValue &Lo, SDValue &Hi, SelectionDAG &DAG) const
Expand shift-by-parts.
Primary interface to the complete machine description for the target machine.
Type * getType() const
All values are typed, get the type of this value.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ LOCAL_ADDRESS
Address space for local memory.
@ PARAM_I_ADDRESS
Address space for indirect addressable parameter memory (VTX1).
@ CONSTANT_ADDRESS
Address space for constant memory (VTX2).
@ FLAT_ADDRESS
Address space for flat memory.
@ GLOBAL_ADDRESS
Address space for global memory (RAT0, VTX0).
@ PRIVATE_ADDRESS
Address space for private memory.
@ BUILD_VERTICAL_VECTOR
This node is for VLIW targets and it is used to represent a vector that is stored in consecutive regi...
@ CONST_DATA_PTR
Pointer to the start of the shader's constant data.
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
constexpr char Args[]
Key for Kernel::Metadata::mArgs.
LLVM_READNONE constexpr bool isShader(CallingConv::ID CC)
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ AMDGPU_CS
Used for Mesa/AMDPAL compute shaders.
@ AMDGPU_VS
Used for Mesa vertex shaders, or AMDPAL last shader stage before rasterization (vertex shader if tess...
@ AMDGPU_KERNEL
Used for AMDGPU code object kernels.
@ AMDGPU_HS
Used for Mesa/AMDPAL hull shaders (= tessellation control shaders).
@ AMDGPU_GS
Used for Mesa/AMDPAL geometry shaders.
@ AMDGPU_PS
Used for Mesa/AMDPAL pixel shaders.
@ Cold
Attempts to make code in the caller as efficient as possible under the assumption that the call is no...
@ SPIR_KERNEL
Used for SPIR kernel functions.
@ Fast
Attempts to make calls as fast as possible (e.g.
@ AMDGPU_ES
Used for AMDPAL shader stage before geometry shader if geometry is in use.
@ AMDGPU_LS
Used for AMDPAL vertex shader if tessellation is in use.
@ C
The default llvm calling convention, compatible with C.
bool isNON_EXTLoad(const SDNode *N)
Returns true if the specified node is a non-extending load.
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
@ ADDC
Carry-setting nodes for multiple precision addition and subtraction.
@ FMAD
FMAD - Perform a * b + c, while getting the same result as the separately rounded operations.
@ ADD
Simple integer binary arithmetic operators.
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
@ INTRINSIC_VOID
OUTCHAIN = INTRINSIC_VOID(INCHAIN, INTRINSICID, arg1, arg2, ...) This node represents a target intrin...
@ FADD
Simple binary floating point operators.
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
@ SDIVREM
SDIVREM/UDIVREM - Divide two integers and produce both a quotient and remainder result.
@ SIGN_EXTEND
Conversion operators.
@ CTTZ_ZERO_UNDEF
Bit counting operators with an undefined result for zero inputs.
@ IS_FPCLASS
Performs a check of floating point class property, defined by IEEE-754.
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
@ SHL
Shift and rotation operations.
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
@ AND
Bitwise operators - logical and, logical or, logical xor.
@ INTRINSIC_WO_CHAIN
RESULT = INTRINSIC_WO_CHAIN(INTRINSICID, arg1, arg2, ...) This node represents a target intrinsic fun...
@ ADDE
Carry-using nodes for multiple precision addition and subtraction.
@ INSERT_VECTOR_ELT
INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element at IDX replaced with VAL.
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
@ SHL_PARTS
SHL_PARTS/SRA_PARTS/SRL_PARTS - These operators are used for expanded integer shift operations.
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
LLVM_ABI CondCode getSetCCInverse(CondCode Operation, EVT Type)
Return the operation corresponding to !(X op Y), where 'op' is a valid SetCC operation.
LLVM_ABI CondCode getSetCCSwappedOperands(CondCode Operation)
Return the operation corresponding to (Y op X) when given the operation for (X op Y).
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
LoadExtType
LoadExtType enum - This enum defines the three variants of LOADEXT (load with extension).
int getLDSNoRetOp(uint16_t Opcode)
@ Kill
The last use of a register.
NodeAddr< NodeBase * > Node
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
FunctionAddr VTableAddr Value
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
LLVM_ABI bool isNullConstant(SDValue V)
Returns true if V is a constant integer zero.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
bool CCAssignFn(unsigned ValNo, MVT ValVT, MVT LocVT, CCValAssign::LocInfo LocInfo, ISD::ArgFlagsTy ArgFlags, Type *OrigTy, CCState &State)
CCAssignFn - This function assigns a location for Val, updating State to reflect the change.
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
LLVM_ABI bool isAllOnesConstant(SDValue V)
Returns true if V is an integer constant with all bits set.
LLVM_ABI void reportFatalUsageError(Error Err)
Report a fatal error that does not indicate a bug in LLVM.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
This struct is a compact representation of a valid (non-zero power of two) alignment.
EVT changeVectorElementTypeToInteger() const
Return a vector with the same number of elements as this vector, but with the element type converted ...
TypeSize getStoreSize() const
Return the number of bytes overwritten by a store of the specified value type.
bool isSimple() const
Test if the given EVT is simple (as opposed to being extended).
bool bitsGT(EVT VT) const
Return true if this has more bits than VT.
bool bitsLT(EVT VT) const
Return true if this has less bits than VT.
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
uint64_t getScalarSizeInBits() const
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
bool isVector() const
Return true if this is a vector value type.
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
bool bitsGE(EVT VT) const
Return true if this has no less bits than VT.
EVT getVectorElementType() const
Given a vector type, return the type of each element.
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
bool bitsLE(EVT VT) const
Return true if this has no more bits than VT.
This class contains a discriminated union of information about pointers in memory operands,...
bool isBeforeLegalizeOps() const