65 SP::I0, SP::I1, SP::I2, SP::I3, SP::I4, SP::I5
73 ValNo, ValVT, State.AllocateStack(8,
Align(4)), LocVT, LocInfo));
82 ValNo, ValVT, State.AllocateStack(4,
Align(4)), LocVT, LocInfo));
91 SP::I0, SP::I1, SP::I2, SP::I3, SP::I4, SP::I5
113 assert((LocVT == MVT::f32 || LocVT == MVT::f128
115 "Can't handle non-64 bits locations");
118 unsigned size = (LocVT == MVT::f128) ? 16 : 8;
121 unsigned Offset = State.AllocateStack(
size, alignment);
124 if (LocVT == MVT::i64 &&
Offset < 6*8)
127 else if (LocVT == MVT::f64 &&
Offset < 16*8)
130 else if (LocVT == MVT::f32 &&
Offset < 16*8)
133 else if (LocVT == MVT::f128 &&
Offset < 16*8)
151 if (LocVT == MVT::f32)
165 unsigned Offset = State.AllocateStack(4,
Align(4));
167 if (LocVT == MVT::f32 &&
Offset < 16*8) {
174 if (LocVT == MVT::i32 &&
Offset < 6*8) {
226#include "SparcGenCallingConv.inc"
232 static_assert(SP::I0 + 7 == SP::I7 && SP::O0 + 7 == SP::O7,
234 if (
Reg >= SP::I0 &&
Reg <= SP::I7)
235 return Reg - SP::I0 + SP::O0;
242 const Type *RetTy)
const {
244 CCState CCInfo(CallConv, isVarArg, MF, RVLocs, Context);
245 return CCInfo.
CheckReturn(Outs, Subtarget->is64Bit() ? RetCC_Sparc64
255 if (Subtarget->is64Bit())
284 for (
unsigned i = 0, realRVLocIdx = 0;
286 ++i, ++realRVLocIdx) {
290 SDValue Arg = OutVals[realRVLocIdx];
318 unsigned RetAddrOffset = 8;
340 return DAG.
getNode(SPISD::RET_GLUE,
DL, MVT::Other, RetOps);
369 for (
unsigned i = 0; i != RVLocs.
size(); ++i) {
398 if (i+1 < RVLocs.
size() && RVLocs[i+1].getLocReg() == VA.
getLocReg()) {
419 return DAG.
getNode(SPISD::RET_GLUE,
DL, MVT::Other, RetOps);
426 if (Subtarget->is64Bit())
455 for (
unsigned i = 0, e = ArgLocs.
size(); i != e; ++i, ++InIdx) {
459 if (Ins[InIdx].Flags.isSRet()) {
476 Register VRegHi = RegInfo.createVirtualRegister(&SP::IntRegsRegClass);
491 &SP::IntRegsRegClass);
504 Register VReg = RegInfo.createVirtualRegister(&SP::IntRegsRegClass);
510 else if (VA.
getLocVT() != MVT::i32) {
573 unsigned ArgIndex = Ins[InIdx].OrigArgIndex;
574 assert(Ins[InIdx].PartOffset == 0);
575 while (i + 1 != e && Ins[InIdx + 1].OrigArgIndex == ArgIndex) {
577 unsigned PartOffset = Ins[InIdx + 1].PartOffset;
602 SP::I0, SP::I1, SP::I2, SP::I3, SP::I4, SP::I5
605 const MCPhysReg *CurArgReg = ArgRegs+NumAllocated, *ArgRegEnd = ArgRegs+6;
607 if (NumAllocated == 6)
611 ArgOffset = 68+4*NumAllocated;
617 std::vector<SDValue> OutChains;
619 for (; CurArgReg != ArgRegEnd; ++CurArgReg) {
620 Register VReg = RegInfo.createVirtualRegister(&SP::IntRegsRegClass);
633 if (!OutChains.empty()) {
634 OutChains.push_back(Chain);
656 const unsigned ArgArea = 128;
669 if (VA.getValVT() == MVT::i32 && VA.needsCustom())
675 switch (VA.getLocInfo()) {
700 unsigned Offset = VA.getLocMemOffset() + ArgArea;
701 unsigned ValSize = VA.getValVT().getSizeInBits() / 8;
728 Subtarget->getStackPointerBias());
734 for (; ArgOffset < 6*8; ArgOffset += 8) {
744 if (!OutChains.
empty())
758 return TRI->isReservedReg(MF, r);
762 return TRI->isReservedReg(MF, r);
770 F, (
"SPARC doesn't support"
771 " function calls if any of the argument registers is reserved.")});
777 if (Subtarget->is64Bit())
785 return Call->hasFnAttr(Attribute::ReturnsTwice);
794 const char *CalleeName =
E->getSymbol();
808 auto &Outs = CLI.
Outs;
812 if (Caller.getFnAttribute(
"disable-tail-calls").getValueAsString() ==
"true")
818 unsigned StackSizeLimit = Subtarget->is64Bit() ? 48 : 0;
824 if (!Outs.empty() && Caller.hasStructRetAttr() != Outs[0].Flags.isSRet())
829 for (
auto &Arg : Outs)
830 if (Arg.Flags.isByVal())
867 ArgsSize = (ArgsSize+7) & ~7;
873 for (
unsigned i = 0, e = Outs.
size(); i != e; ++i) {
875 if (!Flags.isByVal())
879 unsigned Size = Flags.getByValSize();
880 Align Alignment = Flags.getNonZeroByValAlign();
887 Chain = DAG.
getMemcpy(Chain, dl, FIPtr, Arg, SizeNode, Alignment,
900 assert(!isTailCall || ArgsSize == 0);
909 bool hasStructRetAttr =
false;
910 unsigned SRetArgSize = 0;
912 for (
unsigned i = 0, realArgIdx = 0, byvalArgIdx = 0, e = ArgLocs.
size();
916 SDValue Arg = OutVals[realArgIdx];
921 if (Flags.isByVal()) {
922 Arg = ByValArgs[byvalArgIdx++];
948 if (Flags.isSRet()) {
960 hasStructRetAttr =
true;
962 assert(Outs[realArgIdx].OrigArgIndex == 0);
1034 unsigned ArgIndex = Outs[realArgIdx].OrigArgIndex;
1035 assert(Outs[realArgIdx].PartOffset == 0);
1038 if (i + 1 != e && Outs[realArgIdx + 1].OrigArgIndex == ArgIndex) {
1039 Type *OrigArgType = CLI.
Args[ArgIndex].Ty;
1047 SlotVT = Outs[realArgIdx].VT;
1053 DAG.
getStore(Chain, dl, Arg, SpillSlot,
1057 while (i + 1 != e && Outs[realArgIdx + 1].OrigArgIndex == ArgIndex) {
1058 SDValue PartValue = OutVals[realArgIdx + 1];
1059 unsigned PartOffset = Outs[realArgIdx + 1].PartOffset;
1067 "Not enough space for argument part!");
1100 if (!MemOpChains.
empty())
1108 for (
const auto &[OrigReg,
N] : RegsToPass) {
1127 Ops.push_back(Chain);
1128 Ops.push_back(Callee);
1129 if (hasStructRetAttr)
1131 for (
const auto &[OrigReg,
N] : RegsToPass) {
1140 ?
TRI->getRTCallPreservedMask(CallConv)
1146 assert(Mask &&
"Missing call preserved mask for calling convention");
1150 Ops.push_back(InGlue);
1154 return DAG.
getNode(SPISD::TAIL_CALL, dl, MVT::Other,
Ops);
1157 Chain = DAG.
getNode(SPISD::CALL, dl, NodeTys,
Ops);
1171 for (
unsigned i = 0; i != RVLocs.
size(); ++i) {
1172 assert(RVLocs[i].isRegLoc() &&
"Can only return in registers!");
1173 if (RVLocs[i].getLocVT() == MVT::v2i32) {
1176 Chain, dl,
toCallerWindow(RVLocs[i++].getLocReg()), MVT::i32, InGlue);
1177 Chain =
Lo.getValue(1);
1178 InGlue =
Lo.getValue(2);
1182 Chain, dl,
toCallerWindow(RVLocs[i].getLocReg()), MVT::i32, InGlue);
1183 Chain =
Hi.getValue(1);
1184 InGlue =
Hi.getValue(2);
1191 RVLocs[i].getValVT(), InGlue)
1206 .
Case(
"i0", SP::I0).
Case(
"i1", SP::I1).
Case(
"i2", SP::I2).
Case(
"i3", SP::I3)
1207 .
Case(
"i4", SP::I4).
Case(
"i5", SP::I5).
Case(
"i6", SP::I6).
Case(
"i7", SP::I7)
1208 .
Case(
"o0", SP::O0).
Case(
"o1", SP::O1).
Case(
"o2", SP::O2).
Case(
"o3", SP::O3)
1209 .
Case(
"o4", SP::O4).
Case(
"o5", SP::O5).
Case(
"o6", SP::O6).
Case(
"o7", SP::O7)
1210 .
Case(
"l0", SP::L0).
Case(
"l1", SP::L1).
Case(
"l2", SP::L2).
Case(
"l3", SP::L3)
1211 .
Case(
"l4", SP::L4).
Case(
"l5", SP::L5).
Case(
"l6", SP::L6).
Case(
"l7", SP::L7)
1212 .
Case(
"g0", SP::G0).
Case(
"g1", SP::G1).
Case(
"g2", SP::G2).
Case(
"g3", SP::G3)
1213 .
Case(
"g4", SP::G4).
Case(
"g5", SP::G5).
Case(
"g6", SP::G6).
Case(
"g7", SP::G7)
1220 if (!
TRI->isReservedReg(MF, Reg))
1237 MVT ValTy = VA.getLocVT();
1240 if (!VA.isRegLoc() || (ValTy != MVT::f64 && ValTy != MVT::f128))
1243 if (!Outs[VA.getValNo()].Flags.isVarArg())
1248 Register firstReg = (ValTy == MVT::f64) ? SP::D0 : SP::Q0;
1249 unsigned argSize = (ValTy == MVT::f64) ? 8 : 16;
1250 unsigned Offset = argSize * (VA.getLocReg() - firstReg);
1251 assert(
Offset < 16*8 &&
"Offset out of range, bad register enum?");
1255 unsigned IReg = SP::I0 +
Offset/8;
1256 if (ValTy == MVT::f64)
1261 assert(ValTy == MVT::f128 &&
"Unexpected type!");
1270 VA.getLocVT(), VA.getLocInfo());
1298 unsigned StackReserved = 6 * 8u;
1299 unsigned ArgsSize = std::max<unsigned>(StackReserved, CCInfo.
getStackSize());
1302 ArgsSize =
alignTo(ArgsSize, 16);
1326 for (
unsigned i = 0, e = ArgLocs.
size(); i != e; ++i) {
1382 RegsToPass.
push_back(std::make_pair(HiReg, Hi64));
1383 RegsToPass.
push_back(std::make_pair(LoReg, Lo64));
1395 if (i+1 < ArgLocs.
size() && ArgLocs[i+1].isRegLoc() &&
1396 ArgLocs[i+1].getLocReg() == VA.
getLocReg()) {
1408 RegsToPass.
push_back(std::make_pair(Reg, Arg));
1419 Subtarget->getStackPointerBias() +
1427 if (!MemOpChains.
empty())
1435 for (
const auto &[Reg,
N] : RegsToPass) {
1452 Ops.push_back(Chain);
1453 Ops.push_back(Callee);
1454 for (
const auto &[Reg,
N] : RegsToPass)
1460 ((hasReturnsTwice) ?
TRI->getRTCallPreservedMask(CLI.
CallConv)
1467 assert(Mask &&
"Missing call preserved mask for calling convention");
1473 Ops.push_back(InGlue);
1478 return DAG.
getNode(SPISD::TAIL_CALL,
DL, MVT::Other,
Ops);
1498 if (CLI.
Ins.size() == 1 && CLI.
Ins[0].VT == MVT::f32 && !CLI.
CB)
1499 CLI.
Ins[0].Flags.setInReg();
1504 for (
unsigned i = 0; i != RVLocs.
size(); ++i) {
1649 if (!Subtarget->useSoftFloat()) {
1654 if (Subtarget->is64Bit()) {
1731 if (Subtarget->is64Bit()) {
1793 if (Subtarget->isVIS3()) {
1798 if (Subtarget->is64Bit()) {
1820 if (Subtarget->isV9()) {
1823 if (Subtarget->is64Bit())
1827 }
else if (Subtarget->hasLeonCasa())
1842 if (Subtarget->is64Bit()) {
1849 if (!Subtarget->isV9()) {
1891 if (Subtarget->useSoftMulDiv()) {
1899 if (Subtarget->is64Bit()) {
1933 if (Subtarget->isV9() && Subtarget->hasHardQuad()) {
1941 if (Subtarget->hasHardQuad()) {
1949 if (Subtarget->isV9()) {
1972 if (Subtarget->fixAllFDIVSQRT()) {
1979 if (Subtarget->hasNoFMULS()) {
1984 if (!Subtarget->is64Bit())
1987 if (Subtarget->hasLeonCycleCounter())
1990 if (Subtarget->isVIS3()) {
2002 }
else if (Subtarget->usePopc()) {
2043 return Subtarget->useSoftFloat();
2059 const APInt &DemandedElts,
2061 unsigned Depth)
const {
2065 switch (
Op.getOpcode()) {
2067 case SPISD::SELECT_ICC:
2068 case SPISD::SELECT_XCC:
2069 case SPISD::SELECT_FCC:
2084 (((
LHS.getOpcode() == SPISD::SELECT_ICC ||
2085 LHS.getOpcode() == SPISD::SELECT_XCC) &&
2086 LHS.getOperand(3).getOpcode() == SPISD::CMPICC) ||
2087 (
LHS.getOpcode() == SPISD::SELECT_FCC &&
2088 (
LHS.getOperand(3).getOpcode() == SPISD::CMPFCC ||
2089 LHS.getOperand(3).getOpcode() == SPISD::CMPFCC_V9))) &&
2092 SPCC =
LHS.getConstantOperandVal(2);
2104 GA->getValueType(0),
2105 GA->getOffset(), TF);
2109 CP->getAlign(), CP->getOffset(), TF);
2119 ES->getValueType(0), TF);
2127 unsigned HiTF,
unsigned LoTF,
2130 EVT VT =
Op.getValueType();
2150 Idx = DAG.
getNode(SPISD::Lo,
DL,
Op.getValueType(),
2154 Idx =
makeHiLoPair(
Op, ELF::R_SPARC_GOT22, ELF::R_SPARC_GOT10, DAG);
2173 return makeHiLoPair(
Op, ELF::R_SPARC_HI22, ELF::R_SPARC_LO10, DAG);
2179 L44 = DAG.
getNode(SPISD::Lo,
DL, VT, L44);
2223 : ELF::R_SPARC_TLS_LDM_HI22);
2226 : ELF::R_SPARC_TLS_LDM_LO10);
2229 : ELF::R_SPARC_TLS_LDM_ADD);
2232 : ELF::R_SPARC_TLS_LDM_CALL);
2249 const uint32_t *Mask = Subtarget->getRegisterInfo()->getCallPreservedMask(
2251 assert(Mask &&
"Missing call preserved mask for calling convention");
2258 Chain = DAG.
getNode(SPISD::TLS_CALL,
DL, NodeTys,
Ops);
2274 return DAG.
getNode(SPISD::TLS_ADD,
DL, PtrVT, Ret, HiLo,
2279 unsigned ldTF = ((PtrVT == MVT::i64) ? ELF::R_SPARC_TLS_IE_LDX
2280 : ELF::R_SPARC_TLS_IE_LD);
2290 ELF::R_SPARC_TLS_IE_LO10, DAG);
2295 return DAG.
getNode(SPISD::TLS_ADD,
DL, PtrVT,
2326 Args.emplace_back(Arg, ArgTy);
2332 RTLIB::Libcall LibFunc,
2333 unsigned numArgs)
const {
2335 if (LibFuncImpl == RTLIB::Unsupported)
2345 Type *RetTyABI = RetTy;
2354 if (!Subtarget->is64Bit()) {
2355 Entry.IsSRet =
true;
2356 Entry.IndirectType = RetTy;
2358 Entry.IsReturned =
false;
2359 Args.push_back(Entry);
2363 assert(
Op->getNumOperands() >= numArgs &&
"Not enough operands!");
2364 for (
unsigned i = 0, e = numArgs; i != e; ++i) {
2376 if (RetTyABI == RetTy)
2392 const char *
LibCall =
nullptr;
2393 bool is64Bit = Subtarget->is64Bit();
2433 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2440 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2445 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2450 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2455 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2461 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2466 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2473 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2480 return DAG.
getNode(SPISD::CMPICC,
DL, MVT::Glue, Result, RHS);
2489 if (
Op.getOperand(0).getValueType() == MVT::f64)
2492 if (
Op.getOperand(0).getValueType() == MVT::f32)
2503 if (
Op.getOperand(0).getValueType() != MVT::f128)
2506 if (
Op.getValueType() == MVT::f64)
2507 return TLI.
LowerF128Op(
Op, DAG, RTLIB::FPROUND_F128_F64, 1);
2508 if (
Op.getValueType() == MVT::f32)
2509 return TLI.
LowerF128Op(
Op, DAG, RTLIB::FPROUND_F128_F32, 1);
2519 EVT VT =
Op.getValueType();
2520 assert(VT == MVT::i32 || VT == MVT::i64);
2523 if (
Op.getOperand(0).getValueType() == MVT::f128
2525 RTLIB::Libcall LibFunc =
2526 VT == MVT::i32 ? RTLIB::FPTOSINT_F128_I32 : RTLIB::FPTOSINT_F128_I64;
2536 Op = DAG.
getNode(SPISD::FTOI, dl, MVT::f32,
Op.getOperand(0));
2538 Op = DAG.
getNode(SPISD::FTOX, dl, MVT::f64,
Op.getOperand(0));
2547 EVT OpVT =
Op.getOperand(0).getValueType();
2548 assert(OpVT == MVT::i32 || (OpVT == MVT::i64));
2550 EVT floatVT = (OpVT == MVT::i32) ? MVT::f32 : MVT::f64;
2553 if (
Op.getValueType() == MVT::f128
2555 RTLIB::Libcall LibFunc =
2556 OpVT == MVT::i32 ? RTLIB::SINTTOFP_I32_F128 : RTLIB::SINTTOFP_I64_F128;
2566 unsigned opcode = (OpVT == MVT::i32)? SPISD::ITOF : SPISD::XTOF;
2567 return DAG.
getNode(opcode, dl,
Op.getValueType(), Tmp);
2573 EVT VT =
Op.getValueType();
2577 if (
Op.getOperand(0).getValueType() != MVT::f128 ||
2581 assert(VT == MVT::i32 || VT == MVT::i64);
2585 VT == MVT::i32 ? RTLIB::FPTOUINT_F128_I32 : RTLIB::FPTOUINT_F128_I64, 1);
2591 EVT OpVT =
Op.getOperand(0).getValueType();
2592 assert(OpVT == MVT::i32 || OpVT == MVT::i64);
2596 if (
Op.getValueType() != MVT::f128 || (hasHardQuad && TLI.
isTypeLegal(OpVT)))
2600 OpVT == MVT::i32 ? RTLIB::UINTTOFP_I32_F128
2601 : RTLIB::UINTTOFP_I64_F128,
2623 if (
LHS.getValueType().isInteger()) {
2626 if (
is64Bit && isV9 &&
LHS.getValueType() == MVT::i64 &&
2628 return DAG.
getNode(SPISD::BR_REG, dl, MVT::Other, Chain, Dest,
2632 CompareFlag = DAG.
getNode(SPISD::CMPICC, dl, MVT::Glue,
LHS,
RHS);
2636 Opc =
LHS.getValueType() == MVT::i32 ? SPISD::BPICC : SPISD::BPXCC;
2641 if (!hasHardQuad &&
LHS.getValueType() == MVT::f128) {
2644 Opc = isV9 ? SPISD::BPICC : SPISD::BRICC;
2646 unsigned CmpOpc = isV9 ? SPISD::CMPFCC_V9 : SPISD::CMPFCC;
2649 Opc = isV9 ? SPISD::BRFCC_V9 : SPISD::BRFCC;
2652 return DAG.
getNode(
Opc, dl, MVT::Other, Chain, Dest,
2673 if (
LHS.getValueType().isInteger()) {
2680 EVT ValType = TrueVal.getValueType();
2681 bool IsEligibleType = ValType.isScalarInteger() || ValType == MVT::f32 ||
2682 ValType == MVT::f64 ||
2683 (ValType == MVT::f128 && hasHardQuad);
2684 if (
is64Bit && isV9 &&
LHS.getValueType() == MVT::i64 &&
2687 SPISD::SELECT_REG, dl, TrueVal.getValueType(), TrueVal, FalseVal,
2690 CompareFlag = DAG.
getNode(SPISD::CMPICC, dl, MVT::Glue,
LHS,
RHS);
2691 Opc =
LHS.getValueType() == MVT::i32 ?
2692 SPISD::SELECT_ICC : SPISD::SELECT_XCC;
2695 if (!hasHardQuad &&
LHS.getValueType() == MVT::f128) {
2698 Opc = SPISD::SELECT_ICC;
2700 unsigned CmpOpc = isV9 ? SPISD::CMPFCC_V9 : SPISD::CMPFCC;
2702 Opc = SPISD::SELECT_FCC;
2706 return DAG.
getNode(
Opc, dl, TrueVal.getValueType(), TrueVal, FalseVal,
2732 EVT VT =
Node->getValueType(0);
2757 EVT VT =
Op->getValueType(0);
2766 OffsetToStackStart += 16 * (Subtarget.is64Bit() ? 8 : 4);
2768 if (!Subtarget.is64Bit())
2769 OffsetToStackStart += 4;
2783 EVT VT =
Size->getValueType(0);
2786 unsigned SPReg = SP::O6;
2791 unsigned regSpillArea;
2792 if (Subtarget->is64Bit()) {
2832 bool IsOveraligned = MaybeAlignment.has_value();
2857 bool AlwaysFlush =
false) {
2861 EVT VT =
Op.getValueType();
2863 unsigned FrameReg = SP::I6;
2874 unsigned Offset = (Subtarget->is64Bit()) ? (stackBias + 112) : 56;
2881 if (Subtarget->is64Bit())
2891 uint64_t depth =
Op.getConstantOperandVal(0);
2904 EVT VT =
Op.getValueType();
2906 uint64_t depth =
Op.getConstantOperandVal(0);
2919 unsigned Offset = (Subtarget->is64Bit()) ? 120 : 60;
2949 Lo32 = DAG.
getNode(opcode, dl, MVT::f32, Lo32);
2951 Hi32 = DAG.
getNode(opcode, dl, MVT::f32, Hi32);
3009 if (MemVT == MVT::f128)
3058 if (MemVT == MVT::f128)
3061 if (MemVT == MVT::i64) {
3076 &&
"invalid opcode");
3080 if (
Op.getValueType() == MVT::f64)
3082 if (
Op.getValueType() != MVT::f128)
3098 Lo64 = DAG.
getNode(
Op.getOpcode(), dl, MVT::f64, Lo64);
3103 Hi64 = DAG.
getNode(
Op.getOpcode(), dl, MVT::f64, Hi64);
3129 unsigned IntNo =
Op.getConstantOperandVal(0);
3132 case Intrinsic::thread_pointer: {
3142 bool hasHardQuad = Subtarget->hasHardQuad();
3143 bool isV9 = Subtarget->isV9();
3144 bool is64Bit = Subtarget->is64Bit();
3146 switch (
Op.getOpcode()) {
3201 APInt V =
C->getValueAPF().bitcastToAPInt();
3215 Src.getSimpleValueType() == MVT::f64)
3223 switch (
N->getOpcode()) {
3235 switch (
MI.getOpcode()) {
3237 case SP::SELECT_CC_Int_ICC:
3238 case SP::SELECT_CC_FP_ICC:
3239 case SP::SELECT_CC_DFP_ICC:
3240 case SP::SELECT_CC_QFP_ICC:
3241 if (Subtarget->isV9())
3244 case SP::SELECT_CC_Int_XCC:
3245 case SP::SELECT_CC_FP_XCC:
3246 case SP::SELECT_CC_DFP_XCC:
3247 case SP::SELECT_CC_QFP_XCC:
3249 case SP::SELECT_CC_Int_FCC:
3250 case SP::SELECT_CC_FP_FCC:
3251 case SP::SELECT_CC_DFP_FCC:
3252 case SP::SELECT_CC_QFP_FCC:
3253 if (Subtarget->isV9())
3261 unsigned BROpcode)
const {
3284 F->insert(It, IfFalseMBB);
3285 F->insert(It, SinkMBB);
3305 MI.getOperand(0).getReg())
3311 MI.eraseFromParent();
3323 if (Constraint.
size() == 1) {
3324 switch (Constraint[0]) {
3340 const char *constraint)
const {
3342 Value *CallOperandVal =
info.CallOperandVal;
3345 if (!CallOperandVal)
3349 switch (*constraint) {
3371 if (Constraint.
size() > 1)
3374 char ConstraintLetter = Constraint[0];
3375 switch (ConstraintLetter) {
3388 if (Result.getNode()) {
3389 Ops.push_back(Result);
3395std::pair<unsigned, const TargetRegisterClass *>
3399 if (Constraint.
empty())
3400 return std::make_pair(0U,
nullptr);
3402 if (Constraint.
size() == 1) {
3403 switch (Constraint[0]) {
3405 if (VT == MVT::v2i32)
3406 return std::make_pair(0U, &SP::IntPairRegClass);
3407 else if (Subtarget->is64Bit())
3408 return std::make_pair(0U, &SP::I64RegsRegClass);
3410 return std::make_pair(0U, &SP::IntRegsRegClass);
3412 if (VT == MVT::f32 || VT == MVT::i32)
3413 return std::make_pair(0U, &SP::FPRegsRegClass);
3414 else if (VT == MVT::f64 || VT == MVT::i64)
3415 return std::make_pair(0U, &SP::LowDFPRegsRegClass);
3416 else if (VT == MVT::f128)
3417 return std::make_pair(0U, &SP::LowQFPRegsRegClass);
3419 return std::make_pair(0U,
nullptr);
3421 if (VT == MVT::f32 || VT == MVT::i32)
3422 return std::make_pair(0U, &SP::FPRegsRegClass);
3423 else if (VT == MVT::f64 || VT == MVT::i64 )
3424 return std::make_pair(0U, &SP::DFPRegsRegClass);
3425 else if (VT == MVT::f128)
3426 return std::make_pair(0U, &SP::QFPRegsRegClass);
3428 return std::make_pair(0U,
nullptr);
3432 if (Constraint.
front() !=
'{')
3433 return std::make_pair(0U,
nullptr);
3435 assert(Constraint.
back() ==
'}' &&
"Not a brace enclosed constraint?");
3438 return std::make_pair(0U,
nullptr);
3440 unsigned long long RegNo;
3449 return std::make_pair(0U,
nullptr);
3450 const char RegTypes[] = {
'g',
'o',
'l',
'i'};
3451 char RegType = RegTypes[RegNo / 8];
3452 char RegIndex =
'0' + (RegNo % 8);
3453 char Tmp[] = {
'{', RegType, RegIndex,
'}', 0};
3458 if (VT != MVT::f32 && VT != MVT::Other &&
RegName[0] ==
'f' &&
3460 if (VT == MVT::f64 && (RegNo % 2 == 0)) {
3463 }
else if (VT == MVT::f128 && (RegNo % 4 == 0)) {
3467 return std::make_pair(0U,
nullptr);
3473 if (!ResultPair.second)
3474 return std::make_pair(0U,
nullptr);
3477 if (Subtarget->is64Bit() && VT == MVT::i64) {
3478 assert(ResultPair.second == &SP::IntRegsRegClass &&
3479 "Unexpected register class");
3480 return std::make_pair(ResultPair.first, &SP::I64RegsRegClass);
3498 RTLIB::Libcall libCall = RTLIB::UNKNOWN_LIBCALL;
3500 switch (
N->getOpcode()) {
3502 llvm_unreachable(
"Do not know how to custom type legalize this operation!");
3507 if (
N->getOperand(0).getValueType() != MVT::f128
3508 ||
N->getValueType(0) != MVT::i64)
3511 ? RTLIB::FPTOSINT_F128_I64
3512 : RTLIB::FPTOUINT_F128_I64);
3517 assert(Subtarget->hasLeonCycleCounter());
3523 Results.push_back(
N->getOperand(0));
3529 if (
N->getValueType(0) != MVT::f128
3530 ||
N->getOperand(0).getValueType() != MVT::i64)
3534 ? RTLIB::SINTTOFP_I64_F128
3535 : RTLIB::UINTTOFP_I64_F128);
3562 if (!Subtarget->getTargetTriple().isOSLinux())
3568 if (Subtarget->isVIS3())
3569 return VT == MVT::f32 || VT == MVT::f64;
3574 bool ForCodeSize)
const {
3575 if (VT != MVT::f32 && VT != MVT::f64)
3577 if (Subtarget->isVIS() && Imm.isZero())
3579 if (Subtarget->isVIS3())
3580 return Imm.isExactlyValue(+0.5) || Imm.isExactlyValue(-0.5) ||
3581 Imm.getExactLog2Abs() == -1;
3590 if (Subtarget->is64Bit() && Subtarget->usePopc())
3598 return Subtarget->isUA2007() && !Subtarget->useSoftFloat();
3603 assert(
MI.getOpcode() == SP::SUBCCrr ||
MI.getOpcode() == SP::SUBCCri);
3605 if (!
Node->hasAnyUseOfValue(0))
3606 MI.getOperand(0).setReg(SP::G0);
3612 bool HasStoreSemantics =
3630 return Builder.CreateFence(Ord);
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static SDValue LowerVASTART(SDValue Op, SelectionDAG &DAG)
static SDValue LowerSTORE(SDValue Op, SelectionDAG &DAG, const ARMSubtarget *Subtarget)
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
Function Alias Analysis Results
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
static LPCC::CondCode IntCondCCodeToICC(SDValue CC, const SDLoc &DL, SDValue &RHS, SelectionDAG &DAG)
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static CodeModel::Model getCodeModel(const PPCSubtarget &S, const TargetMachine &TM, const MachineOperand &MO)
static constexpr MCPhysReg SPReg
static SDValue LowerFP_TO_UINT(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI, bool hasHardQuad)
static bool CC_Sparc_Assign_Ret_Split_64(unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static SDValue LowerUINT_TO_FP(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI, bool hasHardQuad)
static bool CC_Sparc_Assign_Split_64(unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static SDValue getFRAMEADDR(uint64_t depth, SDValue Op, SelectionDAG &DAG, const SparcSubtarget *Subtarget, bool AlwaysFlush=false)
static unsigned toCallerWindow(unsigned Reg)
static SDValue LowerSTACKADDRESS(SDValue Op, SelectionDAG &DAG, const SparcSubtarget &Subtarget)
static SDValue LowerF128Store(SDValue Op, SelectionDAG &DAG)
static SPCC::CondCodes intCondCCodeToRcond(ISD::CondCode CC)
intCondCCodeToRcond - Convert a DAG integer condition code to a SPARC rcond condition.
static SDValue LowerLOAD(SDValue Op, SelectionDAG &DAG)
static void fixupVariableFloatArgs(SmallVectorImpl< CCValAssign > &ArgLocs, ArrayRef< ISD::OutputArg > Outs)
static SDValue LowerFP_TO_SINT(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI, bool hasHardQuad)
static SPCC::CondCodes FPCondCCodeToFCC(ISD::CondCode CC)
FPCondCCodeToFCC - Convert a DAG floatingp oint condition code to a SPARC FCC condition.
static bool isAnyArgRegReserved(const SparcRegisterInfo *TRI, const MachineFunction &MF)
static SDValue getFLUSHW(SDValue Op, SelectionDAG &DAG)
static bool hasReturnsTwiceAttr(SelectionDAG &DAG, SDValue Callee, const CallBase *Call)
static SDValue LowerDYNAMIC_STACKALLOC(SDValue Op, SelectionDAG &DAG, const SparcSubtarget *Subtarget)
static SDValue LowerFRAMEADDR(SDValue Op, SelectionDAG &DAG, const SparcSubtarget *Subtarget)
static SDValue LowerF128_FPROUND(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI)
static SDValue LowerF64Op(SDValue SrcReg64, const SDLoc &dl, SelectionDAG &DAG, unsigned opcode)
static bool RetCC_Sparc64_Full(unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static SDValue LowerBR_CC(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI, bool hasHardQuad, bool isV9, bool is64Bit)
static void emitReservedArgRegCallError(const MachineFunction &MF)
static SDValue LowerATOMIC_LOAD_STORE(SDValue Op, SelectionDAG &DAG)
static bool RetCC_Sparc64_Half(unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static SDValue LowerSELECT_CC(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI, bool hasHardQuad, bool isV9, bool is64Bit)
static SDValue LowerF128_FPEXTEND(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI)
static SDValue LowerFNEGorFABS(SDValue Op, SelectionDAG &DAG, bool isV9)
static SDValue LowerVAARG(SDValue Op, SelectionDAG &DAG)
static bool CC_Sparc64_Half(unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static bool CC_Sparc64_Full(unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static bool CC_Sparc_Assign_SRet(unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static bool Analyze_CC_Sparc64_Half(bool IsReturn, unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static SDValue LowerF128Load(SDValue Op, SelectionDAG &DAG)
static SDValue LowerRETURNADDR(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI, const SparcSubtarget *Subtarget)
static void LookThroughSetCC(SDValue &LHS, SDValue &RHS, ISD::CondCode CC, unsigned &SPCC)
static bool Analyze_CC_Sparc64_Full(bool IsReturn, unsigned &ValNo, MVT &ValVT, MVT &LocVT, CCValAssign::LocInfo &LocInfo, ISD::ArgFlagsTy &ArgFlags, CCState &State)
static SDValue LowerSINT_TO_FP(SDValue Op, SelectionDAG &DAG, const SparcTargetLowering &TLI, bool hasHardQuad)
This file implements the StringSwitch template, which mimics a switch() statement whose cases are str...
This file describes how to lower LLVM code to machine code.
static bool is64Bit(const char *name)
Class for arbitrary precision integers.
This class represents an incoming formal argument to a Function.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
an instruction that atomically reads a memory location, combines it with another value,...
BinOp getOperation() const
LLVM Basic Block Representation.
CCState - This class holds information needed while lowering arguments and return values.
unsigned getFirstUnallocated(ArrayRef< MCPhysReg > Regs) const
getFirstUnallocated - Return the index of the first unallocated register in the set,...
LLVM_ABI void AnalyzeCallResult(const SmallVectorImpl< ISD::InputArg > &Ins, CCAssignFn Fn)
AnalyzeCallResult - Analyze the return values of a call, incorporating info about the passed values i...
LLVM_ABI bool CheckReturn(const SmallVectorImpl< ISD::OutputArg > &Outs, CCAssignFn Fn)
CheckReturn - Analyze the return values of a function, returning true if the return can be performed ...
LLVM_ABI void AnalyzeReturn(const SmallVectorImpl< ISD::OutputArg > &Outs, CCAssignFn Fn)
AnalyzeReturn - Analyze the returned values of a return, incorporating info about the result values i...
LLVM_ABI void AnalyzeCallOperands(const SmallVectorImpl< ISD::OutputArg > &Outs, CCAssignFn Fn)
AnalyzeCallOperands - Analyze the outgoing arguments to a call, incorporating info about the passed v...
uint64_t getStackSize() const
Returns the size of the currently allocated portion of the stack.
LLVM_ABI void AnalyzeFormalArguments(const SmallVectorImpl< ISD::InputArg > &Ins, CCAssignFn Fn)
AnalyzeFormalArguments - Analyze an array of argument values, incorporating info about the formals in...
CCValAssign - Represent assignment of one arg/retval to a location.
Register getLocReg() const
LocInfo getLocInfo() const
static CCValAssign getReg(unsigned ValNo, MVT ValVT, MCRegister Reg, MVT LocVT, LocInfo HTP, bool IsCustom=false)
static CCValAssign getCustomReg(unsigned ValNo, MVT ValVT, MCRegister Reg, MVT LocVT, LocInfo HTP)
static CCValAssign getMem(unsigned ValNo, MVT ValVT, int64_t Offset, MVT LocVT, LocInfo HTP, bool IsCustom=false)
int64_t getLocMemOffset() const
static CCValAssign getCustomMem(unsigned ValNo, MVT ValVT, int64_t Offset, MVT LocVT, LocInfo HTP)
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
This is the shared class of boolean and integer constants.
A parsed version of the target data layout string in and methods for querying it.
bool isLittleEndian() const
Layout endianness...
LLVM_ABI TypeSize getTypeAllocSize(Type *Ty) const
Returns the offset in bytes between successive objects of the specified type, including alignment pad...
Diagnostic information for unsupported feature in backend.
const Function & getFunction() const
bool hasStructRetAttr() const
Determine if the function returns a structure through first or second pointer argument.
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
const GlobalValue * getGlobal() const
Module * getParent()
Get the module that this global value is contained inside of...
Common base class shared among various IRBuilders.
This is an important class for using LLVM in a threaded context.
LLVM_ABI CallingConv::ID getLibcallImplCallingConv(RTLIB::LibcallImpl Call) const
Get the CallingConv that should be used for the specified libcall.
LLVM_ABI RTLIB::LibcallImpl getLibcallImpl(RTLIB::Libcall Call) const
Return the lowering's selection of implementation call for Call.
This class is used to represent ISD::LOAD nodes.
const SDValue & getBasePtr() const
const SDValue & getOffset() const
ISD::LoadExtType getExtensionType() const
Return whether this is a plain node, or one of the varieties of value-extending loads.
Wrapper class representing physical registers. Should be passed by value.
static auto integer_fixedlen_vector_valuetypes()
static auto integer_valuetypes()
TypeSize getSizeInBits() const
Returns the size of the specified MVT in bits.
static MVT getIntegerVT(unsigned BitWidth)
static auto fp_valuetypes()
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
const BasicBlock * getBasicBlock() const
Return the LLVM basic block that this instance corresponded to originally.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
LLVM_ABI int CreateFixedObject(uint64_t Size, int64_t SPOffset, bool IsImmutable, bool isAliased=false)
Create a new object at a fixed location on the stack.
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
void setFrameAddressIsTaken(bool T)
void setHasTailCall(bool V=true)
void setReturnAddressIsTaken(bool s)
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
Register addLiveIn(MCRegister PReg, const TargetRegisterClass *RC)
addLiveIn - Add the specified physical register as a live-in value and create a corresponding virtual...
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addReg(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
Representation of each machine instruction.
Flags getFlags() const
Return the raw flags of the source value,.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
void addLiveIn(MCRegister Reg, Register vreg=Register())
addLiveIn - Add the specified register as a live-in.
Align getBaseAlign() const
Returns alignment and volatility of the memory access.
AAMDNodes getAAInfo() const
Returns the AA info that describes the dereference.
MachineMemOperand * getMemOperand() const
Return a MachineMemOperand object describing the memory reference performed by operation.
const MachinePointerInfo & getPointerInfo() const
const SDValue & getChain() const
EVT getMemoryVT() const
Return the type of the in-memory value.
A Module instance is used to store all the information related to an LLVM module.
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
Wrapper class representing virtual and physical registers.
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
unsigned getOpcode() const
Return the SelectionDAG opcode value for this node.
EVT getValueType(unsigned ResNo) const
Return the type of a specified result.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
const SDValue & getOperand(unsigned i) const
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
LLVM_ABI SDValue getExtLoad(ISD::LoadExtType ExtType, const SDLoc &dl, EVT VT, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, EVT MemVT, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
SDValue getTargetGlobalAddress(const GlobalValue *GV, const SDLoc &DL, EVT VT, int64_t offset=0, unsigned TargetFlags=0)
const SDValue & getRoot() const
Return the root tag of the SelectionDAG.
SDValue getCopyToReg(SDValue Chain, const SDLoc &dl, Register Reg, SDValue N)
LLVM_ABI SDValue getMergeValues(ArrayRef< SDValue > Ops, const SDLoc &dl)
Create a MERGE_VALUES node from the given operands.
LLVM_ABI SDVTList getVTList(EVT VT)
Return an SDVTList that represents the list of values specified.
LLVM_ABI MachineSDNode * getMachineNode(unsigned Opcode, const SDLoc &dl, EVT VT)
These are used for target selectors to create a new node with specified return type(s),...
LLVM_ABI SDValue getRegister(Register Reg, EVT VT)
LLVM_ABI SDValue getLoad(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr)
Loads are not normal binary operators: their result type is not determined by their operands,...
LLVM_ABI SDValue getMemcpy(SDValue Chain, const SDLoc &dl, SDValue Dst, SDValue Src, SDValue Size, Align Alignment, bool isVol, bool AlwaysInline, const CallInst *CI, std::optional< bool > OverrideTailCall, MachinePointerInfo DstPtrInfo, MachinePointerInfo SrcPtrInfo, const AAMDNodes &AAInfo=AAMDNodes(), BatchAAResults *BatchAA=nullptr)
SDValue getCALLSEQ_END(SDValue Chain, SDValue Op1, SDValue Op2, SDValue InGlue, const SDLoc &DL)
Return a new CALLSEQ_END node, which always must have a glue result (to ensure it's not CSE'd).
SDValue getBuildVector(EVT VT, const SDLoc &DL, ArrayRef< SDValue > Ops)
Return an ISD::BUILD_VECTOR node.
SDValue getCopyFromReg(SDValue Chain, const SDLoc &dl, Register Reg, EVT VT)
const DataLayout & getDataLayout() const
LLVM_ABI SDValue getConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
Create a ConstantSDNode wrapping a constant value.
LLVM_ABI SDValue getMemBasePlusOffset(SDValue Base, TypeSize Offset, const SDLoc &DL, const SDNodeFlags Flags=SDNodeFlags())
Returns sum of the base pointer and offset.
SDValue getSignedTargetConstant(int64_t Val, const SDLoc &DL, EVT VT, bool isOpaque=false)
LLVM_ABI SDValue getStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
Helper function to build ISD::STORE nodes.
LLVM_ABI SDValue getSignedConstant(int64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
SDValue getCALLSEQ_START(SDValue Chain, uint64_t InSize, uint64_t OutSize, const SDLoc &DL)
Return a new CALLSEQ_START node, that starts new call frame, in which InSize bytes are set up inside ...
LLVM_ABI SDValue getTargetExtractSubreg(int SRIdx, const SDLoc &DL, EVT VT, SDValue Operand)
A convenience function for creating TargetInstrInfo::EXTRACT_SUBREG nodes.
LLVM_ABI SDValue getExternalSymbol(const char *Sym, EVT VT)
const TargetMachine & getTarget() const
const LibcallLoweringInfo & getLibcalls() const
LLVM_ABI SDValue getIntPtrConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI SDValue getValueType(EVT)
LLVM_ABI SDValue getNode(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDUse > Ops)
Gets or creates the specified node.
SDValue getTargetConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isOpaque=false)
SDValue getTargetBlockAddress(const BlockAddress *BA, EVT VT, int64_t Offset=0, unsigned TargetFlags=0)
MachineFunction & getMachineFunction() const
LLVM_ABI SDValue getFrameIndex(int FI, EVT VT, bool isTarget=false)
LLVM_ABI KnownBits computeKnownBits(SDValue Op, unsigned Depth=0) const
Determine which bits of Op are known to be either zero or one and return them in Known.
LLVM_ABI SDValue getRegisterMask(const uint32_t *RegMask)
LLVMContext * getContext() const
LLVM_ABI SDValue getTargetExternalSymbol(const char *Sym, EVT VT, unsigned TargetFlags=0)
LLVM_ABI SDValue CreateStackTemporary(TypeSize Bytes, Align Alignment)
Create a stack temporary based on the size in bytes and the alignment.
SDValue getTargetConstantPool(const Constant *C, EVT VT, MaybeAlign Align=std::nullopt, int Offset=0, unsigned TargetFlags=0)
LLVM_ABI SDValue getTargetInsertSubreg(int SRIdx, const SDLoc &DL, EVT VT, SDValue Operand, SDValue Subreg)
A convenience function for creating TargetInstrInfo::INSERT_SUBREG nodes.
SDValue getEntryNode() const
Return the token chain corresponding to the entry of the function.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Register getSRetReturnReg() const
int getVarArgsFrameOffset() const
void setVarArgsFrameOffset(int Offset)
void setSRetReturnReg(Register Reg)
int64_t getStackPointerBias() const
The 64-bit ABI uses biased stack and frame pointers, so the stack frame of the current function is th...
void ReplaceNodeResults(SDNode *N, SmallVectorImpl< SDValue > &Results, SelectionDAG &DAG) const override
This callback is invoked when a node result type is illegal for the target, and the operation was reg...
SDValue withTargetFlags(SDValue Op, unsigned TF, SelectionDAG &DAG) const
bool CanLowerReturn(CallingConv::ID CallConv, MachineFunction &MF, bool isVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, LLVMContext &Context, const Type *RetTy) const override
This hook should be implemented to check whether the return values described by the Outs array can fi...
bool useSoftFloat() const override
SDValue bitcastConstantFPToInt(ConstantFPSDNode *C, const SDLoc &DL, SelectionDAG &DAG) const
MachineBasicBlock * expandSelectCC(MachineInstr &MI, MachineBasicBlock *BB, unsigned BROpcode) const
bool isFPImmLegal(const APFloat &Imm, EVT VT, bool ForCodeSize) const override
Returns true if the target can instruction select the specified FP immediate natively.
ConstraintWeight getSingleConstraintMatchWeight(AsmOperandInfo &info, const char *constraint) const override
Examine constraint string and operand type and determine a weight value.
std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const override
Given a physical register constraint (e.g.
bool isCtlzFast() const override
Return true if ctlz instruction is fast.
SDValue LowerFormalArguments(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, const SmallVectorImpl< ISD::InputArg > &Ins, const SDLoc &dl, SelectionDAG &DAG, SmallVectorImpl< SDValue > &InVals) const override
This hook must be implemented to lower the incoming (formal) arguments, described by the Ins array,...
ConstraintType getConstraintType(StringRef Constraint) const override
getConstraintType - Given a constraint letter, return the type of constraint it is for this target.
SDValue LowerFormalArguments_32(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, const SmallVectorImpl< ISD::InputArg > &Ins, const SDLoc &dl, SelectionDAG &DAG, SmallVectorImpl< SDValue > &InVals) const
LowerFormalArguments32 - V8 uses a very simple ABI, where all values are passed in either one or two ...
bool isCheapToSpeculateCtlz(Type *Ty) const override
Return true if it is cheap to speculate a call to intrinsic ctlz.
SDValue LowerCall(TargetLowering::CallLoweringInfo &CLI, SmallVectorImpl< SDValue > &InVals) const override
This hook must be implemented to lower calls into the specified DAG.
bool isCheapToSpeculateCttz(Type *Ty) const override
Return true if it is cheap to speculate a call to intrinsic cttz.
bool IsEligibleForTailCallOptimization(CCState &CCInfo, CallLoweringInfo &CLI, MachineFunction &MF) const
IsEligibleForTailCallOptimization - Check whether the call is eligible for tail call optimization.
bool isOffsetFoldingLegal(const GlobalAddressSDNode *GA) const override
Return true if folding a constant offset with the given GlobalAddress is legal.
bool isFNegFree(EVT VT) const override
Return true if an fneg operation is free to the point where it is never worthwhile to replace it with...
SDValue LowerF128_LibCallArg(SDValue Chain, ArgListTy &Args, SDValue Arg, const SDLoc &DL, SelectionDAG &DAG) const
SDValue makeHiLoPair(SDValue Op, unsigned HiTF, unsigned LoTF, SelectionDAG &DAG) const
SDValue LowerBlockAddress(SDValue Op, SelectionDAG &DAG) const
SDValue LowerGlobalAddress(SDValue Op, SelectionDAG &DAG) const
Instruction * emitLeadingFence(IRBuilderBase &Builder, Instruction *Inst, AtomicOrdering Ord) const override
Inserts in the IR a target-specific intrinsic specifying a fence.
void AdjustInstrPostInstrSelection(MachineInstr &MI, SDNode *Node) const override
This method should be implemented by targets that mark instructions with the 'hasPostISelHook' flag.
void computeKnownBitsForTargetNode(const SDValue Op, KnownBits &Known, const APInt &DemandedElts, const SelectionDAG &DAG, unsigned Depth=0) const override
computeKnownBitsForTargetNode - Determine which of the bits specified in Mask are known to be either ...
SDValue LowerCall_64(TargetLowering::CallLoweringInfo &CLI, SmallVectorImpl< SDValue > &InVals) const
bool isFMAFasterThanFMulAndFAdd(const MachineFunction &MF, EVT VT) const override
Return true if an FMA operation is faster than a pair of fmul and fadd instructions.
SDValue LowerReturn(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, const SmallVectorImpl< SDValue > &OutVals, const SDLoc &dl, SelectionDAG &DAG) const override
This hook must be implemented to lower outgoing return values, described by the Outs array,...
Instruction * emitTrailingFence(IRBuilderBase &Builder, Instruction *Inst, AtomicOrdering Ord) const override
SDValue makeAddress(SDValue Op, SelectionDAG &DAG) const
MachineBasicBlock * EmitInstrWithCustomInserter(MachineInstr &MI, MachineBasicBlock *MBB) const override
This method should be implemented by targets that mark instructions with the 'usesCustomInserter' fla...
SDValue LowerINTRINSIC_WO_CHAIN(SDValue Op, SelectionDAG &DAG) const
SDValue LowerReturn_32(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, const SmallVectorImpl< SDValue > &OutVals, const SDLoc &DL, SelectionDAG &DAG) const
SDValue PerformDAGCombine(SDNode *N, DAGCombinerInfo &DCI) const override
This method will be invoked for all target nodes and for any target-independent nodes that the target...
SDValue PerformBITCASTCombine(SDNode *N, DAGCombinerInfo &DCI) const
SDValue LowerReturn_64(SDValue Chain, CallingConv::ID CallConv, bool IsVarArg, const SmallVectorImpl< ISD::OutputArg > &Outs, const SmallVectorImpl< SDValue > &OutVals, const SDLoc &DL, SelectionDAG &DAG) const
SDValue LowerF128Op(SDValue Op, SelectionDAG &DAG, RTLIB::Libcall LibFunc, unsigned numArgs) const
SDValue LowerGlobalTLSAddress(SDValue Op, SelectionDAG &DAG) const
EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const override
getSetCCResultType - Return the ISD::SETCC ValueType
SDValue LowerCall_32(TargetLowering::CallLoweringInfo &CLI, SmallVectorImpl< SDValue > &InVals) const
bool useLoadStackGuardNode(const Module &M) const override
Override to support customized stack guard loading.
AtomicExpansionKind shouldExpandAtomicRMWInIR(const AtomicRMWInst *AI) const override
Returns how the IR-level AtomicExpand pass should expand the given AtomicRMW, if at all.
SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const override
This callback is invoked for operations that are unsupported by the target, which are registered to u...
SDValue LowerConstantPool(SDValue Op, SelectionDAG &DAG) const
SDValue LowerFormalArguments_64(SDValue Chain, CallingConv::ID CallConv, bool isVarArg, const SmallVectorImpl< ISD::InputArg > &Ins, const SDLoc &dl, SelectionDAG &DAG, SmallVectorImpl< SDValue > &InVals) const
SparcTargetLowering(const TargetMachine &TM, const SparcSubtarget &STI)
void LowerAsmOperandForConstraint(SDValue Op, StringRef Constraint, std::vector< SDValue > &Ops, SelectionDAG &DAG) const override
LowerAsmOperandForConstraint - Lower the specified operand into the Ops vector.
Register getRegisterByName(const char *RegName, LLT VT, const MachineFunction &MF) const override
Return the register ID of the name passed in.
SDValue LowerF128Compare(SDValue LHS, SDValue RHS, unsigned &SPCC, const SDLoc &DL, SelectionDAG &DAG) const
StackOffset holds a fixed and a scalable offset in bytes.
This class is used to represent ISD::STORE nodes.
const SDValue & getBasePtr() const
const SDValue & getOffset() const
const SDValue & getValue() const
StringRef - Represent a constant reference to a string, i.e.
constexpr bool empty() const
empty - Check if the string is empty.
char back() const
back - Get the last character in the string.
constexpr size_t size() const
size - Get the string size.
char front() const
front - Get the first character in the string.
constexpr const char * data() const
data - Get a pointer to the start of the string (which may not be null terminated).
A switch()-like statement whose cases are string literals.
StringSwitch & Case(StringLiteral S, T Value)
TargetInstrInfo - Interface to description of machine instruction set.
void setBooleanVectorContents(BooleanContent Ty)
Specify how the target extends the result of a vector boolean value from a vector of i1 to a wider ty...
void setOperationAction(unsigned Op, MVT VT, LegalizeAction Action)
Indicate that the specified operation does not work with the specified type and indicate what to do a...
bool PredictableSelectIsExpensive
Tells the code generator that select is more expensive than a branch if the branch is usually predict...
EVT getValueType(const DataLayout &DL, Type *Ty, bool AllowUnknown=false) const
Return the EVT corresponding to this LLVM type.
virtual const TargetRegisterClass * getRegClassFor(MVT VT, bool isDivergent=false) const
Return the register class that should be used for the specified value type.
MVT getVectorIdxTy(const DataLayout &DL) const
Returns the type to be used for the index operand of: ISD::INSERT_VECTOR_ELT, ISD::EXTRACT_VECTOR_ELT...
const TargetMachine & getTargetMachine() const
virtual unsigned getNumRegistersForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const
Certain targets require unusual breakdowns of certain types.
virtual MVT getRegisterTypeForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const
Certain combinations of ABIs, Targets and features require that types are legal for some operations a...
bool isJumpExpensive() const
Return true if Flow Control is an expensive operation that should be avoided.
void setMaxAtomicSizeInBitsSupported(unsigned SizeInBits)
Set the maximum atomic operation size supported by the backend.
void setMinFunctionAlignment(Align Alignment)
Set the target's minimum function alignment.
void setBooleanContents(BooleanContent Ty)
Specify how the target extends the result of integer and floating point boolean values from i1 to a w...
void computeRegisterProperties(const TargetRegisterInfo *TRI)
Once all of the register classes are added, this allows us to compute derived properties we expose.
void addRegisterClass(MVT VT, const TargetRegisterClass *RC)
Add the specified register class as an available regclass for the specified value type.
bool isTypeLegal(EVT VT) const
Return true if the target has native support for the specified value type.
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
void setTruncStoreAction(MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified truncating store does not work with the specified type and indicate what ...
@ ZeroOrOneBooleanContent
void setMinCmpXchgSizeInBits(unsigned SizeInBits)
Sets the minimum cmpxchg or ll/sc size supported by the backend.
void setStackPointerRegisterToSaveRestore(Register R)
If set to a physical register, this specifies the register that llvm.savestack/llvm....
AtomicExpansionKind
Enum that specifies what an atomic load/AtomicRMWInst is expanded to, if at all.
void setTargetDAGCombine(ArrayRef< ISD::NodeType > NTs)
Targets should invoke this method for each target independent node that they want to provide a custom...
void setLoadExtAction(unsigned ExtType, MVT ValVT, MVT MemVT, LegalizeAction Action)
Indicate that the specified load with extension does not work with the specified type and indicate wh...
std::vector< ArgListEntry > ArgListTy
void setJumpIsExpensive(bool isExpensive=true)
Tells the code generator not to expand logic operations on comparison predicates into separate sequen...
virtual ConstraintType getConstraintType(StringRef Constraint) const
Given a constraint, return the type of constraint it is for this target.
virtual SDValue LowerToTLSEmulatedModel(const GlobalAddressSDNode *GA, SelectionDAG &DAG) const
Lower TLS global address SDNode for target independent emulated TLS model.
std::pair< SDValue, SDValue > LowerCallTo(CallLoweringInfo &CLI) const
This function lowers an abstract call to a function into an actual call.
bool isPositionIndependent() const
virtual ConstraintWeight getSingleConstraintMatchWeight(AsmOperandInfo &info, const char *constraint) const
Examine constraint string and operand type and determine a weight value.
virtual std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const
Given a physical register constraint (e.g.
TargetLowering(const TargetLowering &)=delete
virtual bool useLoadStackGuardNode(const Module &M) const
If this function returns true, SelectionDAGBuilder emits a LOAD_STACK_GUARD node when it is lowering ...
virtual void LowerAsmOperandForConstraint(SDValue Op, StringRef Constraint, std::vector< SDValue > &Ops, SelectionDAG &DAG) const
Lower the specified operand into the Ops vector.
Primary interface to the complete machine description for the target machine.
TLSModel::Model getTLSModel(const GlobalValue *GV) const
Returns the TLS model which should be used for the given global variable.
bool useEmulatedTLS() const
Returns true if this target uses emulated TLS.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
static constexpr TypeSize getFixed(ScalarTy ExactSize)
The instances of the Type class are immutable: once they are created, they are never changed.
static LLVM_ABI IntegerType * getInt32Ty(LLVMContext &C)
static LLVM_ABI Type * getVoidTy(LLVMContext &C)
bool isFP128Ty() const
Return true if this is 'fp128'.
LLVM_ABI TypeSize getPrimitiveSizeInBits() const LLVM_READONLY
Return the basic size of this type if it is a primitive type.
LLVMContext & getContext() const
Return the LLVMContext in which this type was uniqued.
LLVM Value Representation.
Type * getType() const
All values are typed, get the type of this value.
self_iterator getIterator()
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ C
The default llvm calling convention, compatible with C.
@ SETCC
SetCC operator - This evaluates to a true value iff the condition is true.
@ STACKRESTORE
STACKRESTORE has two operands, an input chain and a pointer to restore to it returns an output chain.
@ STACKSAVE
STACKSAVE - STACKSAVE has one operand, an input chain.
@ SMUL_LOHI
SMUL_LOHI/UMUL_LOHI - Multiply two integers of type iN, producing a signed/unsigned value of type i[2...
@ STACKADDRESS
STACKADDRESS - Represents the llvm.stackaddress intrinsic.
@ BSWAP
Byte Swap and Counting operators.
@ VAEND
VAEND, VASTART - VAEND and VASTART have three operands: an input chain, pointer, and a SRCVALUE.
@ ATOMIC_STORE
OUTCHAIN = ATOMIC_STORE(INCHAIN, val, ptr) This corresponds to "store atomic" instruction.
@ ADDC
Carry-setting nodes for multiple precision addition and subtraction.
@ ADD
Simple integer binary arithmetic operators.
@ LOAD
LOAD and STORE have token chains as their first operand, then the same operands as an LLVM load/store...
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
@ FADD
Simple binary floating point operators.
@ ATOMIC_FENCE
OUTCHAIN = ATOMIC_FENCE(INCHAIN, ordering, scope) This corresponds to the fence instruction.
@ SDIVREM
SDIVREM/UDIVREM - Divide two integers and produce both a quotient and remainder result.
@ FP16_TO_FP
FP16_TO_FP, FP_TO_FP16 - These operators are used to perform promotions and truncation for half-preci...
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
@ BUILD_PAIR
BUILD_PAIR - This is the opposite of EXTRACT_ELEMENT in some ways.
@ BUILTIN_OP_END
BUILTIN_OP_END - This must be the last enum value in this list.
@ SIGN_EXTEND
Conversion operators.
@ CTTZ_ZERO_UNDEF
Bit counting operators with an undefined result for zero inputs.
@ FSINCOS
FSINCOS - Compute both fsin and fcos as a single operation.
@ FNEG
Perform various unary floating-point operations inspired by libm.
@ BR_CC
BR_CC - Conditional branch.
@ BRIND
BRIND - Indirect branch.
@ BR_JT
BR_JT - Jumptable branch.
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
@ ATOMIC_LOAD
Val, OUTCHAIN = ATOMIC_LOAD(INCHAIN, ptr) This corresponds to "load atomic" instruction.
@ UNDEF
UNDEF - An undefined node.
@ VACOPY
VACOPY - VACOPY has 5 operands: an input chain, a destination pointer, a source pointer,...
@ CopyFromReg
CopyFromReg - This node indicates that the input value is a virtual or physical register that is defi...
@ MULHU
MULHU/MULHS - Multiply high - Multiply two integers of type iN, producing an unsigned/signed value of...
@ SHL
Shift and rotation operations.
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
@ DEBUGTRAP
DEBUGTRAP - Trap intended to get the attention of a debugger.
@ SELECT_CC
Select with condition operator - This selects between a true value and a false value (ops #2 and #3) ...
@ ATOMIC_CMP_SWAP
Val, OUTCHAIN = ATOMIC_CMP_SWAP(INCHAIN, ptr, cmp, swap) For double-word atomic operations: ValLo,...
@ DYNAMIC_STACKALLOC
DYNAMIC_STACKALLOC - Allocate some number of bytes on the stack aligned to a specified boundary.
@ SIGN_EXTEND_INREG
SIGN_EXTEND_INREG - This operator atomically performs a SHL/SRA pair to sign extend a small value in ...
@ FP_EXTEND
X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
@ FRAMEADDR
FRAMEADDR, RETURNADDR - These nodes represent llvm.frameaddress and llvm.returnaddress on the DAG.
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
@ READCYCLECOUNTER
READCYCLECOUNTER - This corresponds to the readcyclecounter intrinsic.
@ AND
Bitwise operators - logical and, logical or, logical xor.
@ TRAP
TRAP - Trapping instruction.
@ INTRINSIC_WO_CHAIN
RESULT = INTRINSIC_WO_CHAIN(INTRINSICID, arg1, arg2, ...) This node represents a target intrinsic fun...
@ ADDE
Carry-using nodes for multiple precision addition and subtraction.
@ INSERT_VECTOR_ELT
INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element at IDX replaced with VAL.
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
@ ATOMIC_SWAP
Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt) Val, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN,...
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
@ VAARG
VAARG - VAARG has four operands: an input chain, a pointer, a SRCVALUE, and the alignment.
@ BRCOND
BRCOND - Conditional branch.
@ SHL_PARTS
SHL_PARTS/SRA_PARTS/SRL_PARTS - These operators are used for expanded integer shift operations.
@ AssertSext
AssertSext, AssertZext - These nodes record if a register contains a value that has already been zero...
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
bool isUnsignedIntSetCC(CondCode Code)
Return true if this is a setcc instruction that performs an unsigned comparison when used with intege...
This is an optimization pass for GlobalISel generic memory operations.
auto size(R &&Range, std::enable_if_t< std::is_base_of< std::random_access_iterator_tag, typename std::iterator_traits< decltype(Range.begin())>::iterator_category >::value, void > *=nullptr)
Get the size of a range.
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
LLVM_ABI bool isNullConstant(SDValue V)
Returns true if V is a constant integer zero.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
bool isStrongerThanMonotonic(AtomicOrdering AO)
std::string utostr(uint64_t X, bool isNeg=false)
bool isReleaseOrStronger(AtomicOrdering AO)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
AtomicOrdering
Atomic ordering for LLVM's memory model.
uint16_t MCPhysReg
An unsigned integer type large enough to represent all physical registers, but not necessarily virtua...
uint64_t alignTo(uint64_t Size, Align A)
Returns a multiple of A needed to store Size bytes.
DWARFExpression::Operation Op
bool isAcquireOrStronger(AtomicOrdering AO)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
LLVM_ABI bool isOneConstant(SDValue V)
Returns true if V is a constant integer one.
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
LLVM_ABI bool getAsUnsignedInteger(StringRef Str, unsigned Radix, unsigned long long &Result)
Helper functions for StringRef::getAsInteger.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
This struct is a compact representation of a valid (non-zero power of two) alignment.
EVT changeVectorElementTypeToInteger() const
Return a vector with the same number of elements as this vector, but with the element type converted ...
TypeSize getStoreSize() const
Return the number of bytes overwritten by a store of the specified value type.
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
static EVT getIntegerVT(LLVMContext &Context, unsigned BitWidth)
Returns the EVT that represents an integer with the given number of bits.
uint64_t getFixedSizeInBits() const
Return the size of the specified fixed width value type in bits.
bool isVector() const
Return true if this is a vector value type.
LLVM_ABI Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
Incoming for lane maks phi as machine instruction, incoming register Reg and incoming block Block are...
void resetAll()
Resets the known state of all bits.
KnownBits intersectWith(const KnownBits &RHS) const
Returns KnownBits information that is known to be true for both this and RHS.
This class contains a discriminated union of information about pointers in memory operands,...
MachinePointerInfo getWithOffset(int64_t O) const
static LLVM_ABI MachinePointerInfo getGOT(MachineFunction &MF)
Return a MachinePointerInfo record that refers to a GOT entry.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
This represents a list of ValueType's that has been intern'd by a SelectionDAG.
This contains information for each constraint that we are lowering.
This structure contains all information that is necessary for lowering calls.
SmallVector< ISD::InputArg, 32 > Ins
CallLoweringInfo & setDebugLoc(const SDLoc &dl)
SmallVector< ISD::OutputArg, 32 > Outs
SmallVector< SDValue, 32 > OutVals
CallLoweringInfo & setChain(SDValue InChain)
CallLoweringInfo & setCallee(CallingConv::ID CC, Type *ResultType, SDValue Target, ArgListTy &&ArgsList, AttributeSet ResultAttrs={})