79#include "llvm/IR/IntrinsicsAArch64.h"
80#include "llvm/IR/IntrinsicsAMDGPU.h"
81#include "llvm/IR/IntrinsicsWebAssembly.h"
114#define DEBUG_TYPE "isel"
122 cl::desc(
"Insert the experimental `assertalign` node."),
127 cl::desc(
"Generate low-precision inline sequences "
128 "for some float libcalls"),
134 cl::desc(
"Set the case probability threshold for peeling the case from a "
135 "switch statement. A value greater than 100 will void this "
155 const SDValue *Parts,
unsigned NumParts,
158 std::optional<CallingConv::ID> CC);
167 unsigned NumParts,
MVT PartVT,
EVT ValueVT,
const Value *V,
169 std::optional<CallingConv::ID> CC = std::nullopt,
170 std::optional<ISD::NodeType> AssertOp = std::nullopt) {
174 PartVT, ValueVT, CC))
181 assert(NumParts > 0 &&
"No parts to assemble!");
192 unsigned RoundBits = PartBits * RoundParts;
193 EVT RoundVT = RoundBits == ValueBits ?
199 if (RoundParts > 2) {
203 PartVT, HalfVT, V, InChain);
214 if (RoundParts < NumParts) {
216 unsigned OddParts = NumParts - RoundParts;
219 OddVT, V, InChain, CC);
235 assert(ValueVT ==
EVT(MVT::ppcf128) && PartVT == MVT::f64 &&
246 !PartVT.
isVector() &&
"Unexpected split");
258 if (PartEVT == ValueVT)
262 ValueVT.
bitsLT(PartEVT)) {
275 if (ValueVT.
bitsLT(PartEVT)) {
280 Val = DAG.
getNode(*AssertOp,
DL, PartEVT, Val,
295 llvm::Attribute::StrictFP)) {
297 DAG.
getVTList(ValueVT, MVT::Other), InChain, Val,
309 if (PartEVT == MVT::x86mmx && ValueVT.
isInteger() &&
310 ValueVT.
bitsLT(PartEVT)) {
319 const Twine &ErrMsg) {
322 return Ctx.emitError(ErrMsg);
325 if (CI->isInlineAsm()) {
327 *CI, ErrMsg +
", possible invalid constraint for vector type"));
330 return Ctx.emitError(
I, ErrMsg);
339 const SDValue *Parts,
unsigned NumParts,
342 std::optional<CallingConv::ID> CallConv) {
344 assert(NumParts > 0 &&
"No parts to assemble!");
345 const bool IsABIRegCopy = CallConv.has_value();
354 unsigned NumIntermediates;
359 *DAG.
getContext(), *CallConv, ValueVT, IntermediateVT,
360 NumIntermediates, RegisterVT);
364 NumIntermediates, RegisterVT);
367 assert(NumRegs == NumParts &&
"Part count doesn't match vector breakdown!");
369 assert(RegisterVT == PartVT &&
"Part type doesn't match vector breakdown!");
372 "Part type sizes don't match!");
376 if (NumIntermediates == NumParts) {
379 for (
unsigned i = 0; i != NumParts; ++i)
381 V, InChain, CallConv);
382 }
else if (NumParts > 0) {
385 assert(NumParts % NumIntermediates == 0 &&
386 "Must expand into a divisible number of parts!");
387 unsigned Factor = NumParts / NumIntermediates;
388 for (
unsigned i = 0; i != NumIntermediates; ++i)
390 IntermediateVT, V, InChain, CallConv);
405 DL, BuiltVectorTy,
Ops);
411 if (PartEVT == ValueVT)
427 "Cannot narrow, it would be a lossy transformation");
433 if (PartEVT == ValueVT)
458 }
else if (ValueVT.
bitsLT(PartEVT)) {
467 *DAG.
getContext(), V,
"non-trivial scalar-to-vector conversion");
498 std::optional<CallingConv::ID> CallConv);
505 unsigned NumParts,
MVT PartVT,
const Value *V,
506 std::optional<CallingConv::ID> CallConv = std::nullopt,
520 unsigned OrigNumParts = NumParts;
522 "Copying to an illegal type!");
528 EVT PartEVT = PartVT;
529 if (PartEVT == ValueVT) {
530 assert(NumParts == 1 &&
"No-op copy with multiple parts!");
539 assert(NumParts == 1 &&
"Do not know what to promote to!");
550 "Unknown mismatch!");
552 Val = DAG.
getNode(ExtendKind,
DL, ValueVT, Val);
553 if (PartVT == MVT::x86mmx)
558 assert(NumParts == 1 && PartEVT != ValueVT);
564 "Unknown mismatch!");
567 if (PartVT == MVT::x86mmx)
574 "Failed to tile the value with PartVT!");
577 if (PartEVT != ValueVT) {
579 "scalar-to-vector conversion failed");
588 if (NumParts & (NumParts - 1)) {
591 "Do not know what to expand to!");
593 unsigned RoundBits = RoundParts * PartBits;
594 unsigned OddParts = NumParts - RoundParts;
603 std::reverse(Parts + RoundParts, Parts + NumParts);
605 NumParts = RoundParts;
617 for (
unsigned StepSize = NumParts; StepSize > 1; StepSize /= 2) {
618 for (
unsigned i = 0; i < NumParts; i += StepSize) {
619 unsigned ThisBits = StepSize * PartBits / 2;
622 SDValue &Part1 = Parts[i+StepSize/2];
629 if (ThisBits == PartBits && ThisVT != PartVT) {
637 std::reverse(Parts, Parts + OrigNumParts);
659 if (ValueEVT == MVT::bf16 && PartEVT == MVT::f16) {
661 "Cannot widen to illegal type");
665 }
else if (PartEVT != ValueEVT) {
680 Ops.append((PartNumElts - ValueNumElts).getFixedValue(), EltUndef);
691 std::optional<CallingConv::ID> CallConv) {
695 const bool IsABIRegCopy = CallConv.has_value();
698 EVT PartEVT = PartVT;
699 if (PartEVT == ValueVT) {
745 "lossy conversion of vector to scalar type");
760 unsigned NumIntermediates;
764 *DAG.
getContext(), *CallConv, ValueVT, IntermediateVT, NumIntermediates,
769 NumIntermediates, RegisterVT);
772 assert(NumRegs == NumParts &&
"Part count doesn't match vector breakdown!");
774 assert(RegisterVT == PartVT &&
"Part type doesn't match vector breakdown!");
777 "Mixing scalable and fixed vectors when copying in parts");
779 std::optional<ElementCount> DestEltCnt;
789 if (ValueVT == BuiltVectorTy) {
813 for (
unsigned i = 0; i != NumIntermediates; ++i) {
828 if (NumParts == NumIntermediates) {
831 for (
unsigned i = 0; i != NumParts; ++i)
833 }
else if (NumParts > 0) {
836 assert(NumIntermediates != 0 &&
"division by zero");
837 assert(NumParts % NumIntermediates == 0 &&
838 "Must expand into a divisible number of parts!");
839 unsigned Factor = NumParts / NumIntermediates;
840 for (
unsigned i = 0; i != NumIntermediates; ++i)
848 if (
I.hasOperandBundlesOtherThan(AllowedBundles)) {
852 for (
unsigned i = 0, e =
I.getNumOperandBundles(); i != e; ++i) {
855 OS << LS << U.getTagName();
858 Twine(
"cannot lower ", Name)
864 EVT valuevt, std::optional<CallingConv::ID> CC)
870 std::optional<CallingConv::ID> CC) {
884 for (
unsigned i = 0; i != NumRegs; ++i)
885 Regs.push_back(Reg + i);
886 RegVTs.push_back(RegisterVT);
888 Reg = Reg.id() + NumRegs;
915 for (
unsigned i = 0; i != NumRegs; ++i) {
921 *Glue =
P.getValue(2);
924 Chain =
P.getValue(1);
952 EVT FromVT(MVT::Other);
956 }
else if (NumSignBits > 1) {
964 assert(FromVT != MVT::Other);
970 RegisterVT, ValueVT, V, Chain,
CallConv);
986 unsigned NumRegs =
Regs.size();
1001 NumParts, RegisterVT, V,
CallConv, ExtendKind);
1007 for (
unsigned i = 0; i != NumRegs; ++i) {
1019 if (NumRegs == 1 || Glue)
1030 Chain = Chains[NumRegs-1];
1036 unsigned MatchingIdx,
const SDLoc &dl,
1038 std::vector<SDValue> &
Ops)
const {
1043 Flag.setMatchingOp(MatchingIdx);
1044 else if (!
Regs.empty() &&
Regs.front().isVirtual()) {
1052 Flag.setRegClass(RC->
getID());
1063 "No 1:1 mapping from clobbers to regs?");
1066 for (
unsigned I = 0, E =
ValueVTs.size();
I != E; ++
I) {
1071 "If we clobbered the stack pointer, MFI should know about it.");
1080 for (
unsigned i = 0; i != NumRegs; ++i) {
1081 assert(Reg <
Regs.size() &&
"Mismatch in # registers expected");
1093 unsigned RegCount = std::get<0>(CountAndVT);
1094 MVT RegisterVT = std::get<1>(CountAndVT);
1112 SL->init(
DAG.getTargetLoweringInfo(), TM,
DAG.getDataLayout());
1114 *
DAG.getMachineFunction().getFunction().getParent());
1119 UnusedArgNodeMap.clear();
1121 PendingExports.clear();
1122 PendingConstrainedFP.clear();
1123 PendingConstrainedFPStrict.clear();
1131 DanglingDebugInfoMap.clear();
1138 if (Pending.
empty())
1144 unsigned i = 0, e = Pending.
size();
1145 for (; i != e; ++i) {
1147 if (Pending[i].
getNode()->getOperand(0) == Root)
1155 if (Pending.
size() == 1)
1182 if (!PendingConstrainedFPStrict.empty()) {
1183 assert(PendingConstrainedFP.empty());
1184 updateRoot(PendingConstrainedFPStrict);
1197 if (!PendingConstrainedFP.empty()) {
1198 assert(PendingConstrainedFPStrict.empty());
1199 updateRoot(PendingConstrainedFP);
1203 return DAG.getRoot();
1211 PendingConstrainedFP.size() +
1212 PendingConstrainedFPStrict.size());
1214 PendingConstrainedFP.end());
1215 PendingLoads.append(PendingConstrainedFPStrict.begin(),
1216 PendingConstrainedFPStrict.end());
1217 PendingConstrainedFP.clear();
1218 PendingConstrainedFPStrict.clear();
1225 PendingExports.append(PendingConstrainedFPStrict.begin(),
1226 PendingConstrainedFPStrict.end());
1227 PendingConstrainedFPStrict.clear();
1228 return updateRoot(PendingExports);
1235 assert(Variable &&
"Missing variable");
1242 <<
"dbg_declare: Dropping debug info (bad/undef/unused-arg address)\n");
1258 if (IsParameter && FINode) {
1260 SDV =
DAG.getFrameIndexDbgValue(Variable,
Expression, FINode->getIndex(),
1261 true,
DL, SDNodeOrder);
1266 FuncArgumentDbgValueKind::Declare,
N);
1269 SDV =
DAG.getDbgValue(Variable,
Expression,
N.getNode(),
N.getResNo(),
1270 true,
DL, SDNodeOrder);
1272 DAG.AddDbgValue(SDV, IsParameter);
1277 FuncArgumentDbgValueKind::Declare,
N)) {
1279 <<
" (could not emit func-arg dbg_value)\n");
1290 for (
auto It = FnVarLocs->locs_begin(&
I), End = FnVarLocs->locs_end(&
I);
1292 auto *Var = FnVarLocs->getDILocalVariable(It->VariableID);
1294 if (It->Values.isKillLocation(It->Expr)) {
1300 It->Values.hasArgList())) {
1303 FnVarLocs->getDILocalVariable(It->VariableID),
1304 It->Expr, Vals.
size() > 1, It->DL, SDNodeOrder);
1317 bool SkipDbgVariableRecords =
DAG.getFunctionVarLocs();
1320 for (
DbgRecord &DR :
I.getDbgRecordRange()) {
1322 assert(DLR->getLabel() &&
"Missing label");
1324 DAG.getDbgLabel(DLR->getLabel(), DLR->getDebugLoc(), SDNodeOrder);
1325 DAG.AddDbgLabel(SDV);
1329 if (SkipDbgVariableRecords)
1337 if (
FuncInfo.PreprocessedDVRDeclares.contains(&DVR))
1339 LLVM_DEBUG(
dbgs() <<
"SelectionDAG visiting dbg_declare: " << DVR
1348 if (Values.
empty()) {
1365 SDNodeOrder, IsVariadic)) {
1376 if (
I.isTerminator()) {
1377 HandlePHINodesInSuccessorBlocks(
I.getParent());
1384 bool NodeInserted =
false;
1385 std::unique_ptr<SelectionDAG::DAGNodeInsertedListener> InsertedListener;
1386 MDNode *PCSectionsMD =
I.getMetadata(LLVMContext::MD_pcsections);
1387 MDNode *MMRA =
I.getMetadata(LLVMContext::MD_mmra);
1388 if (PCSectionsMD || MMRA) {
1389 InsertedListener = std::make_unique<SelectionDAG::DAGNodeInsertedListener>(
1390 DAG, [&](
SDNode *) { NodeInserted =
true; });
1400 if (PCSectionsMD || MMRA) {
1401 auto It = NodeMap.find(&
I);
1402 if (It != NodeMap.end()) {
1404 DAG.addPCSections(It->second.getNode(), PCSectionsMD);
1406 DAG.addMMRAMetadata(It->second.getNode(), MMRA);
1407 }
else if (NodeInserted) {
1410 errs() <<
"warning: loosing !pcsections and/or !mmra metadata ["
1411 <<
I.getModule()->getName() <<
"]\n";
1420void SelectionDAGBuilder::visitPHI(
const PHINode &) {
1430#define HANDLE_INST(NUM, OPCODE, CLASS) \
1431 case Instruction::OPCODE: visit##OPCODE((const CLASS&)I); break;
1432#include "llvm/IR/Instruction.def"
1444 for (
const Value *V : Values) {
1469 DanglingDebugInfoMap[Values[0]].emplace_back(Var, Expr,
DL, Order);
1474 auto isMatchingDbgValue = [&](DanglingDebugInfo &DDI) {
1475 DIVariable *DanglingVariable = DDI.getVariable();
1477 if (DanglingVariable == Variable && Expr->
fragmentsOverlap(DanglingExpr)) {
1479 << printDDI(
nullptr, DDI) <<
"\n");
1485 for (
auto &DDIMI : DanglingDebugInfoMap) {
1486 DanglingDebugInfoVector &DDIV = DDIMI.second;
1490 for (
auto &DDI : DDIV)
1491 if (isMatchingDbgValue(DDI))
1494 erase_if(DDIV, isMatchingDbgValue);
1502 auto DanglingDbgInfoIt = DanglingDebugInfoMap.find(V);
1503 if (DanglingDbgInfoIt == DanglingDebugInfoMap.end())
1506 DanglingDebugInfoVector &DDIV = DanglingDbgInfoIt->second;
1507 for (
auto &DDI : DDIV) {
1509 unsigned DbgSDNodeOrder = DDI.getSDNodeOrder();
1512 assert(Variable->isValidLocationForIntrinsic(
DL) &&
1513 "Expected inlined-at fields to agree");
1523 if (!EmitFuncArgumentDbgValue(V, Variable, Expr,
DL,
1524 FuncArgumentDbgValueKind::Value, Val)) {
1526 << printDDI(V, DDI) <<
"\n");
1533 <<
"changing SDNodeOrder from " << DbgSDNodeOrder <<
" to "
1534 << ValSDNodeOrder <<
"\n");
1535 SDV = getDbgValue(Val, Variable, Expr,
DL,
1536 std::max(DbgSDNodeOrder, ValSDNodeOrder));
1537 DAG.AddDbgValue(SDV,
false);
1541 <<
" in EmitFuncArgumentDbgValue\n");
1543 LLVM_DEBUG(
dbgs() <<
"Dropping debug info for " << printDDI(V, DDI)
1547 DAG.getConstantDbgValue(Variable, Expr,
Poison,
DL, DbgSDNodeOrder);
1548 DAG.AddDbgValue(SDV,
false);
1555 DanglingDebugInfo &DDI) {
1560 const Value *OrigV = V;
1564 unsigned SDOrder = DDI.getSDNodeOrder();
1568 bool StackValue =
true;
1593 if (!AdditionalValues.
empty())
1603 dbgs() <<
"Salvaged debug location info for:\n " << *Var <<
"\n"
1604 << *OrigV <<
"\nBy stripping back to:\n " << *V <<
"\n");
1612 assert(OrigV &&
"V shouldn't be null");
1614 auto *SDV =
DAG.getConstantDbgValue(Var, Expr,
Poison,
DL, SDNodeOrder);
1615 DAG.AddDbgValue(SDV,
false);
1617 << printDDI(OrigV, DDI) <<
"\n");
1634 unsigned Order,
bool IsVariadic) {
1639 if (visitEntryValueDbgValue(Values, Var, Expr, DbgLoc))
1644 for (
const Value *V : Values) {
1654 if (CE->getOpcode() == Instruction::IntToPtr) {
1673 N = UnusedArgNodeMap[V];
1678 EmitFuncArgumentDbgValue(V, Var, Expr, DbgLoc,
1679 FuncArgumentDbgValueKind::Value,
N))
1706 bool IsParamOfFunc =
1714 auto VMI =
FuncInfo.ValueMap.find(V);
1715 if (VMI !=
FuncInfo.ValueMap.end()) {
1720 V->getType(), std::nullopt);
1726 unsigned BitsToDescribe = 0;
1728 BitsToDescribe = *VarSize;
1730 BitsToDescribe = Fragment->SizeInBits;
1733 if (
Offset >= BitsToDescribe)
1736 unsigned RegisterSize = RegAndSize.second;
1737 unsigned FragmentSize = (
Offset + RegisterSize > BitsToDescribe)
1738 ? BitsToDescribe -
Offset
1741 Expr,
Offset, FragmentSize);
1745 Var, *FragmentExpr, RegAndSize.first,
false, DbgLoc, Order);
1746 DAG.AddDbgValue(SDV,
false);
1762 DAG.getDbgValueList(Var, Expr, LocationOps, Dependencies,
1763 false, DbgLoc, Order, IsVariadic);
1764 DAG.AddDbgValue(SDV,
false);
1770 for (
auto &Pair : DanglingDebugInfoMap)
1771 for (
auto &DDI : Pair.second)
1782 if (It !=
FuncInfo.ValueMap.end()) {
1786 DAG.getDataLayout(), InReg, Ty,
1803 if (
N.getNode())
return N;
1863 return DAG.getSplatBuildVector(
1866 return DAG.getConstant(*CI,
DL, VT);
1878 getValue(CPA->getAddrDiscriminator()),
1879 getValue(CPA->getDiscriminator()));
1895 visit(CE->getOpcode(), *CE);
1897 assert(N1.
getNode() &&
"visit didn't populate the NodeMap!");
1903 for (
const Use &U :
C->operands()) {
1909 for (
unsigned i = 0, e = Val->
getNumValues(); i != e; ++i)
1910 Constants.push_back(
SDValue(Val, i));
1919 for (
uint64_t i = 0, e = CDS->getNumElements(); i != e; ++i) {
1923 for (
unsigned i = 0, e = Val->
getNumValues(); i != e; ++i)
1932 if (
C->getType()->isStructTy() ||
C->getType()->isArrayTy()) {
1934 "Unknown struct or array constant!");
1938 unsigned NumElts = ValueVTs.
size();
1942 for (
unsigned i = 0; i != NumElts; ++i) {
1943 EVT EltVT = ValueVTs[i];
1945 Constants[i] =
DAG.getUNDEF(EltVT);
1956 return DAG.getBlockAddress(BA, VT);
1959 return getValue(Equiv->getGlobalValue());
1964 if (VT == MVT::aarch64svcount) {
1965 assert(
C->isNullValue() &&
"Can only zero this target type!");
1971 assert(
C->isNullValue() &&
"Can only zero this target type!");
1988 for (
unsigned i = 0; i != NumElements; ++i)
2016 return DAG.getFrameIndex(
2024 Inst->getType(), std::nullopt);
2038void SelectionDAGBuilder::visitCatchPad(
const CatchPadInst &
I) {
2051 if (IsMSVCCXX || IsCoreCLR)
2057 MachineBasicBlock *TargetMBB =
FuncInfo.getMBB(
I.getSuccessor());
2058 FuncInfo.MBB->addSuccessor(TargetMBB);
2065 if (TargetMBB != NextBlock(
FuncInfo.MBB) ||
2074 DAG.getMachineFunction().setHasEHContTarget(
true);
2080 Value *ParentPad =
I.getCatchSwitchParentPad();
2083 SuccessorColor = &
FuncInfo.Fn->getEntryBlock();
2086 assert(SuccessorColor &&
"No parent funclet for catchret!");
2087 MachineBasicBlock *SuccessorColorMBB =
FuncInfo.getMBB(SuccessorColor);
2088 assert(SuccessorColorMBB &&
"No MBB for SuccessorColor!");
2093 DAG.getBasicBlock(SuccessorColorMBB));
2097void SelectionDAGBuilder::visitCleanupPad(
const CleanupPadInst &CPI) {
2103 FuncInfo.MBB->setIsEHFuncletEntry();
2104 FuncInfo.MBB->setIsCleanupFuncletEntry();
2133 UnwindDests.emplace_back(FuncInfo.
getMBB(EHPadBB), Prob);
2139 UnwindDests.emplace_back(FuncInfo.
getMBB(EHPadBB), Prob);
2140 UnwindDests.back().first->setIsEHScopeEntry();
2143 UnwindDests.back().first->setIsEHFuncletEntry();
2147 for (
const BasicBlock *CatchPadBB : CatchSwitch->handlers()) {
2148 UnwindDests.emplace_back(FuncInfo.
getMBB(CatchPadBB), Prob);
2150 if (IsMSVCCXX || IsCoreCLR)
2151 UnwindDests.back().first->setIsEHFuncletEntry();
2153 UnwindDests.back().first->setIsEHScopeEntry();
2155 NewEHPadBB = CatchSwitch->getUnwindDest();
2161 if (BPI && NewEHPadBB)
2163 EHPadBB = NewEHPadBB;
2170 auto UnwindDest =
I.getUnwindDest();
2171 BranchProbabilityInfo *BPI =
FuncInfo.BPI;
2172 BranchProbability UnwindDestProb =
2177 for (
auto &UnwindDest : UnwindDests) {
2178 UnwindDest.first->setIsEHPad();
2179 addSuccessorWithProb(
FuncInfo.MBB, UnwindDest.first, UnwindDest.second);
2181 FuncInfo.MBB->normalizeSuccProbs();
2184 MachineBasicBlock *CleanupPadMBB =
2185 FuncInfo.getMBB(
I.getCleanupPad()->getParent());
2191void SelectionDAGBuilder::visitCatchSwitch(
const CatchSwitchInst &CSI) {
2195void SelectionDAGBuilder::visitRet(
const ReturnInst &
I) {
2196 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
2197 auto &
DL =
DAG.getDataLayout();
2209 if (
I.getParent()->getTerminatingDeoptimizeCall()) {
2226 SmallVector<uint64_t, 4>
Offsets;
2229 unsigned NumValues = ValueVTs.
size();
2232 Align BaseAlign =
DL.getPrefTypeAlign(
I.getOperand(0)->getType());
2233 for (
unsigned i = 0; i != NumValues; ++i) {
2240 if (MemVTs[i] != ValueVTs[i])
2242 Chains[i] =
DAG.getStore(
2250 MVT::Other, Chains);
2251 }
else if (
I.getNumOperands() != 0) {
2254 unsigned NumValues =
Types.size();
2258 const Function *
F =
I.getParent()->getParent();
2261 I.getOperand(0)->getType(),
F->getCallingConv(),
2265 if (
F->getAttributes().hasRetAttr(Attribute::SExt))
2267 else if (
F->getAttributes().hasRetAttr(Attribute::ZExt))
2270 LLVMContext &
Context =
F->getContext();
2271 bool RetInReg =
F->getAttributes().hasRetAttr(Attribute::InReg);
2273 for (
unsigned j = 0;
j != NumValues; ++
j) {
2286 &Parts[0], NumParts, PartVT, &
I, CC, ExtendKind);
2289 ISD::ArgFlagsTy
Flags = ISD::ArgFlagsTy();
2293 if (
I.getOperand(0)->getType()->isPointerTy()) {
2295 Flags.setPointerAddrSpace(
2299 if (NeedsRegBlock) {
2300 Flags.setInConsecutiveRegs();
2301 if (j == NumValues - 1)
2302 Flags.setInConsecutiveRegsLast();
2310 else if (
F->getAttributes().hasRetAttr(Attribute::NoExt))
2313 for (
unsigned i = 0; i < NumParts; ++i) {
2316 VT, Types[j], 0, 0));
2326 const Function *
F =
I.getParent()->getParent();
2328 F->getAttributes().hasAttrSomewhere(Attribute::SwiftError)) {
2330 ISD::ArgFlagsTy
Flags = ISD::ArgFlagsTy();
2331 Flags.setSwiftError();
2343 bool isVarArg =
DAG.getMachineFunction().getFunction().isVarArg();
2345 DAG.getMachineFunction().getFunction().getCallingConv();
2346 Chain =
DAG.getTargetLoweringInfo().LowerReturn(
2351 "LowerReturn didn't return a valid chain!");
2362 if (V->getType()->isEmptyTy())
2366 if (VMI !=
FuncInfo.ValueMap.end()) {
2368 "Unused value assigned virtual registers!");
2381 if (
FuncInfo.isExportedInst(V))
return;
2393 if (VI->getParent() == FromBB)
2419 const BasicBlock *SrcBB = Src->getBasicBlock();
2420 const BasicBlock *DstBB = Dst->getBasicBlock();
2424 auto SuccSize = std::max<uint32_t>(
succ_size(SrcBB), 1);
2434 Src->addSuccessorWithoutProb(Dst);
2437 Prob = getEdgeProbability(Src, Dst);
2438 Src->addSuccessor(Dst, Prob);
2444 return I->getParent() == BB;
2468 if (CurBB == SwitchBB ||
2474 InvertCond ? IC->getInversePredicate() : IC->getPredicate();
2479 InvertCond ? FC->getInversePredicate() : FC->getPredicate();
2481 if (FC->hasNoNaNs() ||
2489 CaseBlock CB(Condition, BOp->getOperand(0), BOp->getOperand(1),
nullptr,
2491 SL->SwitchCases.push_back(CB);
2500 SL->SwitchCases.push_back(CB);
2508 unsigned Depth = 0) {
2517 if (Necessary !=
nullptr) {
2520 if (Necessary->contains(
I))
2548 if (BPI !=
nullptr) {
2554 std::optional<bool> Likely;
2557 else if (BPI->
isEdgeHot(
I.getParent(), IfFalse))
2561 if (
Opc == (*Likely ? Instruction::And : Instruction::Or))
2573 if (CostThresh <= 0)
2594 Value *BrCond =
I.getCondition();
2595 auto ShouldCountInsn = [&RhsDeps, &BrCond](
const Instruction *Ins) {
2596 for (
const auto *U : Ins->users()) {
2599 if (UIns != BrCond && !RhsDeps.
contains(UIns))
2612 for (
unsigned PruneIters = 0; PruneIters < MaxPruneIters; ++PruneIters) {
2614 for (
const auto &InsPair : RhsDeps) {
2615 if (!ShouldCountInsn(InsPair.first)) {
2616 ToDrop = InsPair.first;
2620 if (ToDrop ==
nullptr)
2622 RhsDeps.erase(ToDrop);
2625 for (
const auto &InsPair : RhsDeps) {
2630 CostOfIncluding +=
TTI->getInstructionCost(
2633 if (CostOfIncluding > CostThresh)
2659 const Value *BOpOp0, *BOpOp1;
2673 if (BOpc == Instruction::And)
2674 BOpc = Instruction::Or;
2675 else if (BOpc == Instruction::Or)
2676 BOpc = Instruction::And;
2682 bool BOpIsInOrAndTree = BOpc && BOpc ==
Opc && BOp->
hasOneUse();
2687 TProb, FProb, InvertCond);
2697 if (
Opc == Instruction::Or) {
2718 auto NewTrueProb = TProb / 2;
2719 auto NewFalseProb = TProb / 2 + FProb;
2722 NewFalseProb, InvertCond);
2729 Probs[1], InvertCond);
2731 assert(
Opc == Instruction::And &&
"Unknown merge op!");
2751 auto NewTrueProb = TProb + FProb / 2;
2752 auto NewFalseProb = FProb / 2;
2755 NewFalseProb, InvertCond);
2762 Probs[1], InvertCond);
2771 if (Cases.size() != 2)
return true;
2775 if ((Cases[0].CmpLHS == Cases[1].CmpLHS &&
2776 Cases[0].CmpRHS == Cases[1].CmpRHS) ||
2777 (Cases[0].CmpRHS == Cases[1].CmpLHS &&
2778 Cases[0].CmpLHS == Cases[1].CmpRHS)) {
2784 if (Cases[0].CmpRHS == Cases[1].CmpRHS &&
2785 Cases[0].CC == Cases[1].CC &&
2788 if (Cases[0].CC ==
ISD::SETEQ && Cases[0].TrueBB == Cases[1].ThisBB)
2790 if (Cases[0].CC ==
ISD::SETNE && Cases[0].FalseBB == Cases[1].ThisBB)
2797void SelectionDAGBuilder::visitUncondBr(
const UncondBrInst &
I) {
2807 if (Succ0MBB != NextBlock(BrMBB) ||
2816void SelectionDAGBuilder::visitCondBr(
const CondBrInst &
I) {
2817 MachineBasicBlock *BrMBB =
FuncInfo.MBB;
2819 MachineBasicBlock *Succ0MBB =
FuncInfo.getMBB(
I.getSuccessor(0));
2823 const Value *CondVal =
I.getCondition();
2824 MachineBasicBlock *Succ1MBB =
FuncInfo.getMBB(
I.getSuccessor(1));
2843 bool IsUnpredictable =
I.hasMetadata(LLVMContext::MD_unpredictable);
2845 if (!
DAG.getTargetLoweringInfo().isJumpExpensive() && BOp &&
2848 const Value *BOp0, *BOp1;
2851 Opcode = Instruction::And;
2853 Opcode = Instruction::Or;
2860 DAG.getTargetLoweringInfo().getJumpConditionMergingParams(
2861 Opcode, BOp0, BOp1))) {
2863 getEdgeProbability(BrMBB, Succ0MBB),
2864 getEdgeProbability(BrMBB, Succ1MBB),
2869 assert(
SL->SwitchCases[0].ThisBB == BrMBB &&
"Unexpected lowering!");
2873 for (
unsigned i = 1, e =
SL->SwitchCases.size(); i != e; ++i) {
2880 SL->SwitchCases.erase(
SL->SwitchCases.begin());
2886 for (
unsigned i = 1, e =
SL->SwitchCases.size(); i != e; ++i)
2887 FuncInfo.MF->erase(
SL->SwitchCases[i].ThisBB);
2889 SL->SwitchCases.clear();
2895 nullptr, Succ0MBB, Succ1MBB, BrMBB,
getCurSDLoc(),
2916 if (CB.
TrueBB != NextBlock(SwitchBB)) {
2923 auto &TLI =
DAG.getTargetLoweringInfo();
2947 Cond =
DAG.getSetCC(dl, MVT::i1, CondLHS, CondRHS, CB.
CC);
2959 Cond =
DAG.getSetCC(dl, MVT::i1, CmpOp,
DAG.getConstant(
High, dl, VT),
2963 VT, CmpOp,
DAG.getConstant(
Low, dl, VT));
2964 Cond =
DAG.getSetCC(dl, MVT::i1, SUB,
2979 if (CB.
TrueBB == NextBlock(SwitchBB)) {
2995 BrCond =
DAG.getNode(
ISD::BR, dl, MVT::Other, BrCond,
2998 DAG.setRoot(BrCond);
3004 assert(JT.
SL &&
"Should set SDLoc for SelectionDAG!");
3005 assert(JT.
Reg &&
"Should lower JT Header first!");
3006 EVT PTy =
DAG.getTargetLoweringInfo().getJumpTableRegTy(
DAG.getDataLayout());
3010 Index.getValue(1), Table, Index);
3011 DAG.setRoot(BrJumpTable);
3019 assert(JT.
SL &&
"Should set SDLoc for SelectionDAG!");
3026 DAG.getConstant(JTH.
First, dl, VT));
3041 JT.
Reg = JumpTableReg;
3049 Sub.getValueType()),
3053 MVT::Other, CopyTo, CMP,
3057 if (JT.
MBB != NextBlock(SwitchBB))
3058 BrCond =
DAG.getNode(
ISD::BR, dl, MVT::Other, BrCond,
3059 DAG.getBasicBlock(JT.
MBB));
3061 DAG.setRoot(BrCond);
3064 if (JT.
MBB != NextBlock(SwitchBB))
3066 DAG.getBasicBlock(JT.
MBB)));
3068 DAG.setRoot(CopyTo);
3092 if (PtrTy != PtrMemTy)
3108 auto &
DL =
DAG.getDataLayout();
3117 SDValue StackSlotPtr =
DAG.getFrameIndex(FI, PtrTy);
3124 PtrMemTy, dl,
DAG.getEntryNode(), StackSlotPtr,
3138 assert(GuardCheckFn &&
"Guard check function is null");
3149 Entry.IsInReg =
true;
3150 Args.push_back(Entry);
3156 getValue(GuardCheckFn), std::move(Args));
3158 std::pair<SDValue, SDValue> Result = TLI.
LowerCallTo(CLI);
3159 DAG.setRoot(Result.second);
3171 Guard =
DAG.getLoad(PtrMemTy, dl, Chain, GuardPtr,
3177 Guard =
DAG.getPOISON(PtrMemTy);
3219 auto &
DL =
DAG.getDataLayout();
3227 SDValue StackSlotPtr =
DAG.getFrameIndex(FI, PtrTy);
3233 PtrMemTy, dl,
DAG.getEntryNode(), StackSlotPtr,
3248 if (GuardCheckFn->hasParamAttribute(0, Attribute::AttrKind::InReg))
3249 Entry.IsInReg =
true;
3250 Args.push_back(Entry);
3256 getValue(GuardCheckFn), std::move(Args));
3262 Chain = TLI.
makeLibCall(
DAG, RTLIB::STACKPROTECTOR_CHECK_FAIL, MVT::isVoid,
3285 DAG.getNode(
ISD::SUB, dl, VT, SwitchOp,
DAG.getConstant(
B.First, dl, VT));
3289 bool UsePtrType =
false;
3313 if (!
B.FallthroughUnreachable)
3314 addSuccessorWithProb(SwitchBB,
B.Default,
B.DefaultProb);
3315 addSuccessorWithProb(SwitchBB,
MBB,
B.Prob);
3319 if (!
B.FallthroughUnreachable) {
3328 DAG.getBasicBlock(
B.Default));
3332 if (
MBB != NextBlock(SwitchBB))
3350 if (PopCount == 1) {
3357 }
else if (PopCount == BB.
Range) {
3365 DAG.getConstant(1, dl, VT), ShiftOp);
3369 VT, SwitchVal,
DAG.getConstant(
B.Mask, dl, VT));
3376 addSuccessorWithProb(SwitchBB,
B.TargetBB,
B.ExtraProb);
3378 addSuccessorWithProb(SwitchBB, NextMBB, BranchProbToNext);
3386 Cmp,
DAG.getBasicBlock(
B.TargetBB));
3389 if (NextMBB != NextBlock(SwitchBB))
3390 BrAnd =
DAG.getNode(
ISD::BR, dl, MVT::Other, BrAnd,
3391 DAG.getBasicBlock(NextMBB));
3396void SelectionDAGBuilder::visitInvoke(
const InvokeInst &
I) {
3414 const Value *Callee(
I.getCalledOperand());
3417 visitInlineAsm(
I, EHPadBB);
3422 case Intrinsic::donothing:
3424 case Intrinsic::seh_try_begin:
3425 case Intrinsic::seh_scope_begin:
3426 case Intrinsic::seh_try_end:
3427 case Intrinsic::seh_scope_end:
3433 case Intrinsic::experimental_patchpoint_void:
3434 case Intrinsic::experimental_patchpoint:
3435 visitPatchpoint(
I, EHPadBB);
3437 case Intrinsic::experimental_gc_statepoint:
3443 case Intrinsic::wasm_throw: {
3445 std::array<SDValue, 4>
Ops = {
3456 case Intrinsic::wasm_rethrow: {
3457 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
3458 std::array<SDValue, 2>
Ops = {
3467 }
else if (
I.hasDeoptState()) {
3488 BranchProbabilityInfo *BPI =
FuncInfo.BPI;
3489 BranchProbability EHPadBBProb =
3495 addSuccessorWithProb(InvokeMBB, Return);
3496 for (
auto &UnwindDest : UnwindDests) {
3497 UnwindDest.first->setIsEHPad();
3498 addSuccessorWithProb(InvokeMBB, UnwindDest.first, UnwindDest.second);
3504 DAG.getBasicBlock(Return)));
3513void SelectionDAGBuilder::visitCallBrIntrinsic(
const CallBrInst &
I) {
3516 DAG.getTargetLoweringInfo().getTgtMemIntrinsic(
3517 Infos,
I,
DAG.getMachineFunction(),
I.getIntrinsicID());
3518 assert(Infos.
empty() &&
"Intrinsic touches memory");
3521 auto [HasChain, OnlyLoad] = getTargetIntrinsicCallProperties(
I);
3524 getTargetIntrinsicOperands(
I, HasChain, OnlyLoad);
3525 SDVTList VTs = getTargetIntrinsicVTList(
I, HasChain);
3529 getTargetNonMemIntrinsicNode(*
I.getType(), HasChain,
Ops, VTs);
3530 Result = handleTargetIntrinsicRet(
I, HasChain, OnlyLoad, Result);
3535void SelectionDAGBuilder::visitCallBr(
const CallBrInst &
I) {
3536 MachineBasicBlock *CallBrMBB =
FuncInfo.MBB;
3538 if (
I.isInlineAsm()) {
3545 assert(!
I.hasOperandBundles() &&
3546 "Can't have operand bundles for intrinsics");
3547 visitCallBrIntrinsic(
I);
3552 SmallPtrSet<BasicBlock *, 8> Dests;
3553 Dests.
insert(
I.getDefaultDest());
3563 if (
I.isInlineAsm()) {
3564 for (BasicBlock *Dest :
I.getIndirectDests()) {
3566 Target->setIsInlineAsmBrIndirectTarget();
3572 Target->setLabelMustBeEmitted();
3574 if (Dests.
insert(Dest).second)
3583 DAG.getBasicBlock(Return)));
3586void SelectionDAGBuilder::visitResume(
const ResumeInst &RI) {
3587 llvm_unreachable(
"SelectionDAGBuilder shouldn't visit resume instructions!");
3590void SelectionDAGBuilder::visitLandingPad(
const LandingPadInst &LP) {
3592 "Call to landingpad not in landing pad!");
3596 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
3612 assert(ValueVTs.
size() == 2 &&
"Only two-valued landingpads are supported");
3617 if (
FuncInfo.ExceptionPointerVirtReg) {
3618 Ops[0] =
DAG.getZExtOrTrunc(
3619 DAG.getCopyFromReg(
DAG.getEntryNode(), dl,
3626 Ops[1] =
DAG.getZExtOrTrunc(
3627 DAG.getCopyFromReg(
DAG.getEntryNode(), dl,
3634 DAG.getVTList(ValueVTs),
Ops);
3642 if (JTB.first.HeaderBB ==
First)
3643 JTB.first.HeaderBB =
Last;
3656 for (
unsigned i = 0, e =
I.getNumSuccessors(); i != e; ++i) {
3658 bool Inserted =
Done.insert(BB).second;
3663 addSuccessorWithProb(IndirectBrMBB, Succ);
3673 if (!
I.shouldLowerToTrap(
DAG.getTarget().Options.TrapUnreachable,
3674 DAG.getTarget().Options.NoTrapAfterNoreturn))
3680void SelectionDAGBuilder::visitUnary(
const User &
I,
unsigned Opcode) {
3683 Flags.copyFMF(*FPOp);
3691void SelectionDAGBuilder::visitBinary(
const User &
I,
unsigned Opcode) {
3694 Flags.setNoSignedWrap(OFBinOp->hasNoSignedWrap());
3695 Flags.setNoUnsignedWrap(OFBinOp->hasNoUnsignedWrap());
3698 Flags.setExact(ExactOp->isExact());
3700 Flags.setDisjoint(DisjointOp->isDisjoint());
3702 Flags.copyFMF(*FPOp);
3711void SelectionDAGBuilder::visitShift(
const User &
I,
unsigned Opcode) {
3715 EVT ShiftTy =
DAG.getTargetLoweringInfo().getShiftAmountTy(
3720 if (!
I.getType()->isVectorTy() && Op2.
getValueType() != ShiftTy) {
3722 "Unexpected shift type");
3732 if (
const OverflowingBinaryOperator *OFBinOp =
3734 nuw = OFBinOp->hasNoUnsignedWrap();
3735 nsw = OFBinOp->hasNoSignedWrap();
3737 if (
const PossiblyExactOperator *ExactOp =
3739 exact = ExactOp->isExact();
3742 Flags.setExact(exact);
3743 Flags.setNoSignedWrap(nsw);
3744 Flags.setNoUnsignedWrap(nuw);
3750void SelectionDAGBuilder::visitSDiv(
const User &
I) {
3761void SelectionDAGBuilder::visitICmp(
const ICmpInst &
I) {
3767 auto &TLI =
DAG.getTargetLoweringInfo();
3780 Flags.setSameSign(
I.hasSameSign());
3782 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
3788void SelectionDAGBuilder::visitFCmp(
const FCmpInst &
I) {
3795 if (FPMO->hasNoNaNs() ||
3796 (
DAG.isKnownNeverNaN(Op1) &&
DAG.isKnownNeverNaN(Op2)))
3800 Flags.copyFMF(*FPMO);
3802 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
3812 return isa<SelectInst>(V);
3816void SelectionDAGBuilder::visitSelect(
const User &
I) {
3820 unsigned NumValues = ValueVTs.
size();
3821 if (NumValues == 0)
return;
3831 bool IsUnaryAbs =
false;
3832 bool Negate =
false;
3836 Flags.copyFMF(*FPOp);
3838 Flags.setUnpredictable(
3843 EVT VT = ValueVTs[0];
3844 LLVMContext &Ctx = *
DAG.getContext();
3845 auto &TLI =
DAG.getTargetLoweringInfo();
3855 bool UseScalarMinMax = VT.
isVector() &&
3864 switch (SPR.Flavor) {
3873 switch (SPR.NaNBehavior) {
3878 Flags.setNoSignedZeros(
true);
3892 switch (SPR.NaNBehavior) {
3897 Flags.setNoSignedZeros(
true);
3939 for (
unsigned i = 0; i != NumValues; ++i) {
3945 Values[i] =
DAG.getNegative(Values[i], dl, VT);
3948 for (
unsigned i = 0; i != NumValues; ++i) {
3952 Values[i] =
DAG.getNode(
3959 DAG.getVTList(ValueVTs), Values));
3962void SelectionDAGBuilder::visitTrunc(
const User &
I) {
3965 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
3969 Flags.setNoSignedWrap(Trunc->hasNoSignedWrap());
3970 Flags.setNoUnsignedWrap(Trunc->hasNoUnsignedWrap());
3976void SelectionDAGBuilder::visitZExt(
const User &
I) {
3980 auto &TLI =
DAG.getTargetLoweringInfo();
3985 Flags.setNonNeg(PNI->hasNonNeg());
3990 if (
Flags.hasNonNeg() &&
3999void SelectionDAGBuilder::visitSExt(
const User &
I) {
4003 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4008void SelectionDAGBuilder::visitFPTrunc(
const User &
I) {
4014 Flags.copyFMF(*FPOp);
4015 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4018 DAG.getTargetConstant(
4023void SelectionDAGBuilder::visitFPExt(
const User &
I) {
4026 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4030 Flags.copyFMF(*FPOp);
4034void SelectionDAGBuilder::visitFPToUI(
const User &
I) {
4037 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4042void SelectionDAGBuilder::visitFPToSI(
const User &
I) {
4045 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4050void SelectionDAGBuilder::visitUIToFP(
const User &
I) {
4053 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4057 Flags.setNonNeg(PNI->hasNonNeg());
4062void SelectionDAGBuilder::visitSIToFP(
const User &
I) {
4065 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4070void SelectionDAGBuilder::visitPtrToAddr(
const User &
I) {
4073 const auto &TLI =
DAG.getTargetLoweringInfo();
4081void SelectionDAGBuilder::visitPtrToInt(
const User &
I) {
4085 auto &TLI =
DAG.getTargetLoweringInfo();
4086 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4095void SelectionDAGBuilder::visitIntToPtr(
const User &
I) {
4099 auto &TLI =
DAG.getTargetLoweringInfo();
4107void SelectionDAGBuilder::visitBitCast(
const User &
I) {
4110 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
4115 if (DestVT !=
N.getValueType())
4123 setValue(&
I,
DAG.getConstant(
C->getValue(), dl, DestVT,
false,
4129void SelectionDAGBuilder::visitAddrSpaceCast(
const User &
I) {
4130 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4131 const Value *SV =
I.getOperand(0);
4136 unsigned DestAS =
I.getType()->getPointerAddressSpace();
4138 if (!TM.isNoopAddrSpaceCast(SrcAS, DestAS))
4144void SelectionDAGBuilder::visitInsertElement(
const User &
I) {
4145 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4152 InVec, InVal, InIdx));
4155void SelectionDAGBuilder::visitExtractElement(
const User &
I) {
4156 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4165void SelectionDAGBuilder::visitShuffleVector(
const User &
I) {
4170 Mask = SVI->getShuffleMask();
4174 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4182 DAG.getVectorIdxConstant(0,
DL));
4193 unsigned MaskNumElts =
Mask.size();
4195 if (SrcNumElts == MaskNumElts) {
4201 if (SrcNumElts < MaskNumElts) {
4205 if (MaskNumElts % SrcNumElts == 0) {
4209 unsigned NumConcat = MaskNumElts / SrcNumElts;
4210 bool IsConcat =
true;
4211 SmallVector<int, 8> ConcatSrcs(NumConcat, -1);
4212 for (
unsigned i = 0; i != MaskNumElts; ++i) {
4218 if ((Idx % SrcNumElts != (i % SrcNumElts)) ||
4219 (ConcatSrcs[i / SrcNumElts] >= 0 &&
4220 ConcatSrcs[i / SrcNumElts] != (
int)(Idx / SrcNumElts))) {
4225 ConcatSrcs[i / SrcNumElts] = Idx / SrcNumElts;
4232 for (
auto Src : ConcatSrcs) {
4245 unsigned PaddedMaskNumElts =
alignTo(MaskNumElts, SrcNumElts);
4246 unsigned NumConcat = PaddedMaskNumElts / SrcNumElts;
4262 SmallVector<int, 8> MappedOps(PaddedMaskNumElts, -1);
4263 for (
unsigned i = 0; i != MaskNumElts; ++i) {
4265 if (Idx >= (
int)SrcNumElts)
4266 Idx -= SrcNumElts - PaddedMaskNumElts;
4274 if (MaskNumElts != PaddedMaskNumElts)
4276 DAG.getVectorIdxConstant(0,
DL));
4282 assert(SrcNumElts > MaskNumElts);
4286 int StartIdx[2] = {-1, -1};
4287 bool CanExtract =
true;
4288 for (
int Idx : Mask) {
4293 if (Idx >= (
int)SrcNumElts) {
4301 int NewStartIdx =
alignDown(Idx, MaskNumElts);
4302 if (NewStartIdx + MaskNumElts > SrcNumElts ||
4303 (StartIdx[Input] >= 0 && StartIdx[Input] != NewStartIdx))
4307 StartIdx[Input] = NewStartIdx;
4310 if (StartIdx[0] < 0 && StartIdx[1] < 0) {
4316 for (
unsigned Input = 0; Input < 2; ++Input) {
4317 SDValue &Src = Input == 0 ? Src1 : Src2;
4318 if (StartIdx[Input] < 0)
4319 Src =
DAG.getUNDEF(VT);
4322 DAG.getVectorIdxConstant(StartIdx[Input],
DL));
4327 SmallVector<int, 8> MappedOps(Mask);
4328 for (
int &Idx : MappedOps) {
4329 if (Idx >= (
int)SrcNumElts)
4330 Idx -= SrcNumElts + StartIdx[1] - MaskNumElts;
4335 setValue(&
I,
DAG.getVectorShuffle(VT,
DL, Src1, Src2, MappedOps));
4344 for (
int Idx : Mask) {
4348 Res =
DAG.getUNDEF(EltVT);
4350 SDValue &Src = Idx < (int)SrcNumElts ? Src1 : Src2;
4351 if (Idx >= (
int)SrcNumElts) Idx -= SrcNumElts;
4354 DAG.getVectorIdxConstant(Idx,
DL));
4364 ArrayRef<unsigned> Indices =
I.getIndices();
4365 const Value *Op0 =
I.getOperand(0);
4367 Type *AggTy =
I.getType();
4374 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4380 unsigned NumAggValues = AggValueVTs.
size();
4381 unsigned NumValValues = ValValueVTs.
size();
4385 if (!NumAggValues) {
4393 for (; i != LinearIndex; ++i)
4394 Values[i] = IntoUndef ?
DAG.getUNDEF(AggValueVTs[i]) :
4399 for (; i != LinearIndex + NumValValues; ++i)
4400 Values[i] = FromUndef ?
DAG.getUNDEF(AggValueVTs[i]) :
4404 for (; i != NumAggValues; ++i)
4405 Values[i] = IntoUndef ?
DAG.getUNDEF(AggValueVTs[i]) :
4409 DAG.getVTList(AggValueVTs), Values));
4413 ArrayRef<unsigned> Indices =
I.getIndices();
4414 const Value *Op0 =
I.getOperand(0);
4416 Type *ValTy =
I.getType();
4421 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4425 unsigned NumValValues = ValValueVTs.
size();
4428 if (!NumValValues) {
4437 for (
unsigned i = LinearIndex; i != LinearIndex + NumValValues; ++i)
4438 Values[i - LinearIndex] =
4444 DAG.getVTList(ValValueVTs), Values));
4447void SelectionDAGBuilder::visitGetElementPtr(
const User &
I) {
4448 Value *Op0 =
I.getOperand(0);
4454 auto &TLI =
DAG.getTargetLoweringInfo();
4459 bool IsVectorGEP =
I.getType()->isVectorTy();
4460 ElementCount VectorElementCount =
4466 const Value *Idx = GTI.getOperand();
4467 if (StructType *StTy = GTI.getStructTypeOrNull()) {
4472 DAG.getDataLayout().getStructLayout(StTy)->getElementOffset(
Field);
4482 N =
DAG.getMemBasePlusOffset(
4483 N,
DAG.getConstant(
Offset, dl,
N.getValueType()), dl, Flags);
4489 unsigned IdxSize =
DAG.getDataLayout().getIndexSizeInBits(AS);
4491 TypeSize ElementSize =
4492 GTI.getSequentialElementStride(
DAG.getDataLayout());
4497 bool ElementScalable = ElementSize.
isScalable();
4503 C =
C->getSplatValue();
4506 if (CI && CI->isZero())
4508 if (CI && !ElementScalable) {
4509 APInt Offs = ElementMul * CI->getValue().sextOrTrunc(IdxSize);
4512 if (
N.getValueType().isVector())
4513 OffsVal =
DAG.getConstant(
4516 OffsVal =
DAG.getConstant(Offs, dl, IdxTy);
4523 Flags.setNoUnsignedWrap(
true);
4526 OffsVal =
DAG.getSExtOrTrunc(OffsVal, dl,
N.getValueType());
4528 N =
DAG.getMemBasePlusOffset(
N, OffsVal, dl, Flags);
4536 if (
N.getValueType().isVector()) {
4538 VectorElementCount);
4539 IdxN =
DAG.getSplat(VT, dl, IdxN);
4543 N =
DAG.getSplat(VT, dl,
N);
4549 IdxN =
DAG.getSExtOrTrunc(IdxN, dl,
N.getValueType());
4551 SDNodeFlags ScaleFlags;
4560 if (ElementScalable) {
4561 EVT VScaleTy =
N.getValueType().getScalarType();
4564 DAG.getConstant(ElementMul.getZExtValue(), dl, VScaleTy));
4565 if (
N.getValueType().isVector())
4566 VScale =
DAG.getSplatVector(
N.getValueType(), dl, VScale);
4567 IdxN =
DAG.getNode(
ISD::MUL, dl,
N.getValueType(), IdxN, VScale,
4572 if (ElementMul != 1) {
4573 if (ElementMul.isPowerOf2()) {
4574 unsigned Amt = ElementMul.logBase2();
4577 DAG.getShiftAmountConstant(Amt,
N.getValueType(), dl),
4580 SDValue Scale =
DAG.getConstant(ElementMul.getZExtValue(), dl,
4582 IdxN =
DAG.getNode(
ISD::MUL, dl,
N.getValueType(), IdxN, Scale,
4592 SDNodeFlags AddFlags;
4596 N =
DAG.getMemBasePlusOffset(
N, IdxN, dl, AddFlags);
4600 if (IsVectorGEP && !
N.getValueType().isVector()) {
4602 N =
DAG.getSplat(VT, dl,
N);
4613 N =
DAG.getPtrExtendInReg(
N, dl, PtrMemTy);
4618void SelectionDAGBuilder::visitAlloca(
const AllocaInst &
I) {
4625 Type *Ty =
I.getAllocatedType();
4626 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4627 auto &
DL =
DAG.getDataLayout();
4628 TypeSize TySize =
DL.getTypeAllocSize(Ty);
4629 MaybeAlign Alignment =
I.getAlign();
4635 AllocSize =
DAG.getZExtOrTrunc(AllocSize, dl, IntPtr);
4637 AllocSize =
DAG.getNode(
4639 DAG.getZExtOrTrunc(
DAG.getTypeSize(dl, MVT::i64, TySize), dl, IntPtr));
4644 Align StackAlign =
DAG.getSubtarget().getFrameLowering()->getStackAlign();
4645 if (*Alignment <= StackAlign)
4646 Alignment = std::nullopt;
4648 const uint64_t StackAlignMask = StackAlign.
value() - 1U;
4653 DAG.getConstant(StackAlignMask, dl, IntPtr),
4658 DAG.getSignedConstant(~StackAlignMask, dl, IntPtr));
4662 DAG.getConstant(Alignment ? Alignment->value() : 0, dl, IntPtr)};
4672 return I.getMetadata(LLVMContext::MD_range);
4677 if (std::optional<ConstantRange> CR = CB->getRange())
4681 return std::nullopt;
4686 return CB->getRetNoFPClass();
4690void SelectionDAGBuilder::visitLoad(
const LoadInst &
I) {
4692 return visitAtomicLoad(
I);
4694 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4695 const Value *SV =
I.getOperand(0);
4700 if (Arg->hasSwiftErrorAttr())
4701 return visitLoadFromSwiftError(
I);
4705 if (Alloca->isSwiftError())
4706 return visitLoadFromSwiftError(
I);
4712 Type *Ty =
I.getType();
4716 unsigned NumValues = ValueVTs.
size();
4720 Align Alignment =
I.getAlign();
4721 AAMDNodes AAInfo =
I.getAAMetadata();
4723 bool isVolatile =
I.isVolatile();
4728 bool ConstantMemory =
false;
4735 BatchAA->pointsToConstantMemory(MemoryLocation(
4740 Root =
DAG.getEntryNode();
4741 ConstantMemory =
true;
4745 Root =
DAG.getRoot();
4756 unsigned ChainI = 0;
4757 for (
unsigned i = 0; i != NumValues; ++i, ++ChainI) {
4773 MachinePointerInfo PtrInfo =
4775 ? MachinePointerInfo(SV, Offsets[i].getKnownMinValue())
4776 : MachinePointerInfo();
4778 SDValue A =
DAG.getObjectPtrOffset(dl, Ptr, Offsets[i]);
4779 SDValue L =
DAG.getLoad(MemVTs[i], dl, Root,
A, PtrInfo, Alignment,
4780 MMOFlags, AAInfo, Ranges);
4781 Chains[ChainI] =
L.getValue(1);
4783 if (MemVTs[i] != ValueVTs[i])
4784 L =
DAG.getPtrExtOrTrunc(L, dl, ValueVTs[i]);
4786 if (MDNode *NoFPClassMD =
I.getMetadata(LLVMContext::MD_nofpclass)) {
4787 uint64_t FPTestInt =
4789 cast<ConstantAsMetadata>(NoFPClassMD->getOperand(0))->getValue())
4791 if (FPTestInt != fcNone) {
4792 SDValue FPTestConst =
4793 DAG.getTargetConstant(FPTestInt, SDLoc(), MVT::i32);
4794 L = DAG.getNode(ISD::AssertNoFPClass, dl, L.getValueType(), L,
4801 if (!ConstantMemory) {
4807 PendingLoads.push_back(Chain);
4811 DAG.getVTList(ValueVTs), Values));
4814void SelectionDAGBuilder::visitStoreToSwiftError(
const StoreInst &
I) {
4815 assert(
DAG.getTargetLoweringInfo().supportSwiftError() &&
4816 "call visitStoreToSwiftError when backend supports swifterror");
4819 SmallVector<uint64_t, 4>
Offsets;
4820 const Value *SrcV =
I.getOperand(0);
4822 SrcV->
getType(), ValueVTs,
nullptr, &Offsets, 0);
4823 assert(ValueVTs.
size() == 1 && Offsets[0] == 0 &&
4824 "expect a single EVT for swifterror");
4833 SDValue(Src.getNode(), Src.getResNo()));
4834 DAG.setRoot(CopyNode);
4837void SelectionDAGBuilder::visitLoadFromSwiftError(
const LoadInst &
I) {
4838 assert(
DAG.getTargetLoweringInfo().supportSwiftError() &&
4839 "call visitLoadFromSwiftError when backend supports swifterror");
4842 !
I.hasMetadata(LLVMContext::MD_nontemporal) &&
4843 !
I.hasMetadata(LLVMContext::MD_invariant_load) &&
4844 "Support volatile, non temporal, invariant for load_from_swift_error");
4846 const Value *SV =
I.getOperand(0);
4847 Type *Ty =
I.getType();
4850 !
BatchAA->pointsToConstantMemory(MemoryLocation(
4852 I.getAAMetadata()))) &&
4853 "load_from_swift_error should not be constant memory");
4856 SmallVector<uint64_t, 4>
Offsets;
4858 ValueVTs,
nullptr, &Offsets, 0);
4859 assert(ValueVTs.
size() == 1 && Offsets[0] == 0 &&
4860 "expect a single EVT for swifterror");
4870void SelectionDAGBuilder::visitStore(
const StoreInst &
I) {
4872 return visitAtomicStore(
I);
4874 const Value *SrcV =
I.getOperand(0);
4875 const Value *PtrV =
I.getOperand(1);
4877 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
4882 if (Arg->hasSwiftErrorAttr())
4883 return visitStoreToSwiftError(
I);
4887 if (Alloca->isSwiftError())
4888 return visitStoreToSwiftError(
I);
4895 SrcV->
getType(), ValueVTs, &MemVTs, &Offsets);
4896 unsigned NumValues = ValueVTs.
size();
4909 Align Alignment =
I.getAlign();
4910 AAMDNodes AAInfo =
I.getAAMetadata();
4914 unsigned ChainI = 0;
4915 for (
unsigned i = 0; i != NumValues; ++i, ++ChainI) {
4925 MachinePointerInfo PtrInfo =
4927 ? MachinePointerInfo(PtrV, Offsets[i].getKnownMinValue())
4928 : MachinePointerInfo();
4932 if (MemVTs[i] != ValueVTs[i])
4933 Val =
DAG.getPtrExtOrTrunc(Val, dl, MemVTs[i]);
4935 DAG.getStore(Root, dl, Val,
Add, PtrInfo, Alignment, MMOFlags, AAInfo);
4936 Chains[ChainI] = St;
4942 DAG.setRoot(StoreNode);
4945void SelectionDAGBuilder::visitMaskedStore(
const CallInst &
I,
4946 bool IsCompressing) {
4949 Value *Src0Operand =
I.getArgOperand(0);
4950 Value *PtrOperand =
I.getArgOperand(1);
4951 Value *MaskOperand =
I.getArgOperand(2);
4952 Align Alignment =
I.getParamAlign(1).valueOrOne();
4962 if (
I.hasMetadata(LLVMContext::MD_nontemporal))
4965 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
4966 MachinePointerInfo(PtrOperand), MMOFlags,
4970 const auto &TLI =
DAG.getTargetLoweringInfo();
4973 !IsCompressing &&
TTI->hasConditionalLoadStoreForType(
4974 I.getArgOperand(0)->getType(),
true)
4980 DAG.setRoot(StoreNode);
5010 C =
C->getSplatValue();
5024 if (!
GEP ||
GEP->getParent() != CurBB)
5027 if (
GEP->getNumOperands() != 2)
5030 const Value *BasePtr =
GEP->getPointerOperand();
5031 const Value *IndexVal =
GEP->getOperand(
GEP->getNumOperands() - 1);
5037 TypeSize ScaleVal =
DL.getTypeAllocSize(
GEP->getResultElementType());
5042 if (ScaleVal != 1 &&
5054void SelectionDAGBuilder::visitMaskedScatter(
const CallInst &
I) {
5058 const Value *Ptr =
I.getArgOperand(1);
5062 Align Alignment =
I.getParamAlign(1).valueOrOne();
5063 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5072 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
5082 EVT IdxVT =
Index.getValueType();
5090 SDValue Scatter =
DAG.getMaskedScatter(
DAG.getVTList(MVT::Other), VT, sdl,
5092 DAG.setRoot(Scatter);
5096void SelectionDAGBuilder::visitMaskedLoad(
const CallInst &
I,
bool IsExpanding) {
5099 Value *PtrOperand =
I.getArgOperand(0);
5100 Value *MaskOperand =
I.getArgOperand(1);
5101 Value *Src0Operand =
I.getArgOperand(2);
5102 Align Alignment =
I.getParamAlign(0).valueOrOne();
5110 AAMDNodes AAInfo =
I.getAAMetadata();
5117 SDValue InChain = AddToChain ?
DAG.getRoot() :
DAG.getEntryNode();
5120 if (
I.hasMetadata(LLVMContext::MD_nontemporal))
5122 if (
I.hasMetadata(LLVMContext::MD_invariant_load))
5125 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
5126 MachinePointerInfo(PtrOperand), MMOFlags,
5129 const auto &TLI =
DAG.getTargetLoweringInfo();
5136 TTI->hasConditionalLoadStoreForType(Src0Operand->
getType(),
5141 DAG.getMaskedLoad(VT, sdl, InChain, Ptr,
Offset, Mask, Src0, VT, MMO,
5148void SelectionDAGBuilder::visitMaskedGather(
const CallInst &
I) {
5152 const Value *Ptr =
I.getArgOperand(0);
5156 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5158 Align Alignment =
I.getParamAlign(0).valueOrOne();
5169 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
5181 EVT IdxVT =
Index.getValueType();
5190 DAG.getMaskedGather(
DAG.getVTList(VT, MVT::Other), VT, sdl,
Ops, MMO,
5206 SDVTList VTs =
DAG.getVTList(MemVT, MVT::i1, MVT::Other);
5208 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5211 MachineFunction &MF =
DAG.getMachineFunction();
5213 MachinePointerInfo(
I.getPointerOperand()), Flags, MemVT.
getStoreSize(),
5214 DAG.getEVTAlign(MemVT), AAMDNodes(),
nullptr, SSID, SuccessOrdering,
5218 dl, MemVT, VTs, InChain,
5226 DAG.setRoot(OutChain);
5229void SelectionDAGBuilder::visitAtomicRMW(
const AtomicRMWInst &
I) {
5232 switch (
I.getOperation()) {
5280 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5283 MachineFunction &MF =
DAG.getMachineFunction();
5285 MachinePointerInfo(
I.getPointerOperand()), Flags, MemVT.
getStoreSize(),
5286 DAG.getEVTAlign(MemVT), AAMDNodes(),
nullptr, SSID, Ordering);
5289 DAG.getAtomic(NT, dl, MemVT, InChain,
5296 DAG.setRoot(OutChain);
5299void SelectionDAGBuilder::visitFence(
const FenceInst &
I) {
5301 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5304 Ops[1] =
DAG.getTargetConstant((
unsigned)
I.getOrdering(), dl,
5306 Ops[2] =
DAG.getTargetConstant(
I.getSyncScopeID(), dl,
5313void SelectionDAGBuilder::visitAtomicLoad(
const LoadInst &
I) {
5320 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5331 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
5332 MachinePointerInfo(
I.getPointerOperand()), Flags, MemVT.
getStoreSize(),
5333 I.getAlign(), AAMDNodes(), Ranges, SSID, Order);
5343 L =
DAG.getPtrExtOrTrunc(L, dl, VT);
5346 DAG.setRoot(OutChain);
5349void SelectionDAGBuilder::visitAtomicStore(
const StoreInst &
I) {
5357 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5367 MachineFunction &MF =
DAG.getMachineFunction();
5369 MachinePointerInfo(
I.getPointerOperand()), Flags, MemVT.
getStoreSize(),
5370 I.getAlign(), AAMDNodes(),
nullptr, SSID, Ordering);
5374 Val =
DAG.getPtrExtOrTrunc(Val, dl, MemVT);
5381 DAG.setRoot(OutChain);
5389std::pair<bool, bool>
5390SelectionDAGBuilder::getTargetIntrinsicCallProperties(
const CallBase &
I) {
5392 bool HasChain = !
F->doesNotAccessMemory();
5394 HasChain &&
F->onlyReadsMemory() &&
F->willReturn() &&
F->doesNotThrow();
5396 return {HasChain, OnlyLoad};
5400 const CallBase &
I,
bool HasChain,
bool OnlyLoad,
5402 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5409 Ops.push_back(
DAG.getRoot());
5422 for (
unsigned i = 0, e =
I.arg_size(); i != e; ++i) {
5423 const Value *Arg =
I.getArgOperand(i);
5424 if (!
I.paramHasAttr(i, Attribute::ImmArg)) {
5432 assert(CI->getBitWidth() <= 64 &&
5433 "large intrinsic immediates not handled");
5434 Ops.push_back(
DAG.getTargetConstant(*CI, SDLoc(), VT));
5441 if (std::optional<OperandBundleUse> Bundle =
5443 auto *Sym = Bundle->Inputs[0].get();
5446 Ops.push_back(SDSym);
5449 if (std::optional<OperandBundleUse> Bundle =
5451 Value *Token = Bundle->Inputs[0].get();
5453 assert(
Ops.back().getValueType() != MVT::Glue &&
5454 "Did not expect another glue node here.");
5457 Ops.push_back(ConvControlToken);
5465 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5473 return DAG.getVTList(ValueVTs);
5477SDValue SelectionDAGBuilder::getTargetNonMemIntrinsicNode(
5500 if (
I.getType()->isVoidTy())
5515void SelectionDAGBuilder::visitTargetIntrinsic(
const CallInst &
I,
5517 auto [HasChain, OnlyLoad] = getTargetIntrinsicCallProperties(
I);
5521 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
5524 TargetLowering::IntrinsicInfo *
Info = !Infos.
empty() ? &Infos[0] :
nullptr;
5527 getTargetIntrinsicOperands(
I, HasChain, OnlyLoad, Info);
5528 SDVTList VTs = getTargetIntrinsicVTList(
I, HasChain);
5533 Flags.copyFMF(*FPMO);
5534 SelectionDAG::FlagInserter FlagsInserter(
DAG, Flags);
5541 if (!Infos.
empty()) {
5544 MachineFunction &MF =
DAG.getMachineFunction();
5546 for (
const auto &Info : Infos) {
5549 MachinePointerInfo MPI;
5551 MPI = MachinePointerInfo(
Info.ptrVal,
Info.offset);
5552 else if (
Info.fallbackAddressSpace)
5553 MPI = MachinePointerInfo(*
Info.fallbackAddressSpace);
5554 EVT MemVT =
Info.memVT;
5556 if (
Size.hasValue() && !
Size.getValue())
5558 Align Alignment =
Info.align.value_or(
DAG.getEVTAlign(MemVT));
5560 MPI,
Info.flags,
Size, Alignment,
I.getAAMetadata(),
5568 Result = getTargetNonMemIntrinsicNode(*
I.getType(), HasChain,
Ops, VTs);
5571 Result = handleTargetIntrinsicRet(
I, HasChain, OnlyLoad, Result);
5628 SDValue TwoToFractionalPartOfX;
5705 if (
Op.getValueType() == MVT::f32 &&
5729 if (
Op.getValueType() == MVT::f32 &&
5828 if (
Op.getValueType() == MVT::f32 &&
5912 return DAG.
getNode(
ISD::FADD, dl, MVT::f32, LogOfExponent, Log2ofMantissa);
5925 if (
Op.getValueType() == MVT::f32 &&
6002 return DAG.
getNode(
ISD::FADD, dl, MVT::f32, LogOfExponent, Log10ofMantissa);
6013 if (
Op.getValueType() == MVT::f32 &&
6026 bool IsExp10 =
false;
6027 if (
LHS.getValueType() == MVT::f32 &&
RHS.getValueType() == MVT::f32 &&
6031 IsExp10 = LHSC->isExactlyValue(Ten);
6058 unsigned Val = RHSC->getSExtValue();
6087 CurSquare, CurSquare);
6092 if (RHSC->getSExtValue() < 0)
6106 EVT VT =
LHS.getValueType();
6129 if ((ScaleInt > 0 || (Saturating &&
Signed)) &&
6133 Opcode, VT, ScaleInt);
6168 switch (
N.getOpcode()) {
6172 Op.getValueType().getSizeInBits());
6197bool SelectionDAGBuilder::EmitFuncArgumentDbgValue(
6204 MachineFunction &MF =
DAG.getMachineFunction();
6205 const TargetInstrInfo *
TII =
DAG.getSubtarget().getInstrInfo();
6209 auto MakeVRegDbgValue = [&](
Register Reg, DIExpression *FragExpr,
6214 auto &Inst =
TII->get(TargetOpcode::DBG_INSTR_REF);
6221 auto *NewDIExpr = FragExpr;
6228 return BuildMI(MF,
DL, Inst,
false, MOs, Variable, NewDIExpr);
6231 auto &Inst =
TII->get(TargetOpcode::DBG_VALUE);
6232 return BuildMI(MF,
DL, Inst, Indirect,
Reg, Variable, FragExpr);
6236 if (Kind == FuncArgumentDbgValueKind::Value) {
6241 if (!IsInEntryBlock)
6257 bool VariableIsFunctionInputArg =
Variable->isParameter() &&
6258 !
DL->getInlinedAt();
6260 if (!IsInPrologue && !VariableIsFunctionInputArg)
6294 if (VariableIsFunctionInputArg) {
6296 if (ArgNo >=
FuncInfo.DescribedArgs.size())
6297 FuncInfo.DescribedArgs.resize(ArgNo + 1,
false);
6298 else if (!IsInPrologue &&
FuncInfo.DescribedArgs.test(ArgNo))
6299 return !NodeMap[
V].getNode();
6304 bool IsIndirect =
false;
6305 std::optional<MachineOperand>
Op;
6307 int FI =
FuncInfo.getArgumentFrameIndex(Arg);
6308 if (FI != std::numeric_limits<int>::max())
6312 if (!
Op &&
N.getNode()) {
6315 if (ArgRegsAndSizes.
size() == 1)
6316 Reg = ArgRegsAndSizes.
front().first;
6319 MachineRegisterInfo &RegInfo = MF.
getRegInfo();
6326 IsIndirect =
Kind != FuncArgumentDbgValueKind::Value;
6330 if (!
Op &&
N.getNode()) {
6334 if (FrameIndexSDNode *FINode =
6341 auto splitMultiRegDbgValue =
6354 uint64_t ExprFragmentSizeInBits = ExprFragmentInfo->SizeInBits;
6357 if (
Offset >= ExprFragmentSizeInBits)
6361 if (
Offset + RegFragmentSizeInBits > ExprFragmentSizeInBits) {
6362 RegFragmentSizeInBits = ExprFragmentSizeInBits -
Offset;
6367 Expr,
Offset, RegFragmentSizeInBits);
6371 if (!FragmentExpr) {
6372 SDDbgValue *SDV =
DAG.getConstantDbgValue(
6374 DAG.AddDbgValue(SDV,
false);
6377 MachineInstr *NewMI = MakeVRegDbgValue(
6378 Reg, *FragmentExpr, Kind != FuncArgumentDbgValueKind::Value);
6379 FuncInfo.ArgDbgValues.push_back(NewMI);
6388 if (VMI !=
FuncInfo.ValueMap.end()) {
6389 const auto &TLI =
DAG.getTargetLoweringInfo();
6390 RegsForValue RFV(
V->getContext(), TLI,
DAG.getDataLayout(), VMI->second,
6391 V->getType(), std::nullopt);
6392 if (RFV.occupiesMultipleRegs())
6393 return splitMultiRegDbgValue(RFV.getRegsAndSizes());
6396 IsIndirect =
Kind != FuncArgumentDbgValueKind::Value;
6397 }
else if (ArgRegsAndSizes.
size() > 1) {
6400 return splitMultiRegDbgValue(ArgRegsAndSizes);
6408 "Expected inlined-at fields to agree");
6409 MachineInstr *NewMI =
nullptr;
6412 NewMI = MakeVRegDbgValue(
Op->getReg(), Expr, IsIndirect);
6414 NewMI =
BuildMI(MF,
DL,
TII->get(TargetOpcode::DBG_VALUE),
true, *
Op,
6418 FuncInfo.ArgDbgValues.push_back(NewMI);
6427 unsigned DbgSDNodeOrder) {
6439 return DAG.getFrameIndexDbgValue(Variable, Expr, FISDN->getIndex(),
6440 false, dl, DbgSDNodeOrder);
6442 return DAG.getDbgValue(Variable, Expr,
N.getNode(),
N.getResNo(),
6443 false, dl, DbgSDNodeOrder);
6448 case Intrinsic::smul_fix:
6450 case Intrinsic::umul_fix:
6452 case Intrinsic::smul_fix_sat:
6454 case Intrinsic::umul_fix_sat:
6456 case Intrinsic::sdiv_fix:
6458 case Intrinsic::udiv_fix:
6460 case Intrinsic::sdiv_fix_sat:
6462 case Intrinsic::udiv_fix_sat:
6475 "expected call_preallocated_setup Value");
6476 for (
const auto *U : PreallocatedSetup->
users()) {
6478 const Function *Fn = UseCall->getCalledFunction();
6479 if (!Fn || Fn->
getIntrinsicID() != Intrinsic::call_preallocated_arg) {
6489bool SelectionDAGBuilder::visitEntryValueDbgValue(
6499 auto ArgIt =
FuncInfo.ValueMap.find(Arg);
6500 if (ArgIt ==
FuncInfo.ValueMap.end()) {
6502 dbgs() <<
"Dropping dbg.value: expression is entry_value but "
6503 "couldn't find an associated register for the Argument\n");
6506 Register ArgVReg = ArgIt->getSecond();
6508 for (
auto [PhysReg, VirtReg] :
FuncInfo.RegInfo->liveins())
6509 if (ArgVReg == VirtReg || ArgVReg == PhysReg) {
6510 SDDbgValue *SDV =
DAG.getVRegDbgValue(
6511 Variable, Expr, PhysReg,
false , DbgLoc, SDNodeOrder);
6512 DAG.AddDbgValue(SDV,
false );
6515 LLVM_DEBUG(
dbgs() <<
"Dropping dbg.value: expression is entry_value but "
6516 "couldn't find a physical register\n");
6521void SelectionDAGBuilder::visitConvergenceControl(
const CallInst &
I,
6524 switch (Intrinsic) {
6525 case Intrinsic::experimental_convergence_anchor:
6528 case Intrinsic::experimental_convergence_entry:
6531 case Intrinsic::experimental_convergence_loop: {
6533 auto *Token = Bundle->Inputs[0].get();
6541void SelectionDAGBuilder::visitVectorHistogram(
const CallInst &
I,
6542 unsigned IntrinsicID) {
6545 assert(IntrinsicID == Intrinsic::experimental_vector_histogram_add &&
6546 "Tried to lower unsupported histogram type");
6552 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
6553 DataLayout TargetDL =
DAG.getDataLayout();
6555 Align Alignment =
DAG.getEVTAlign(VT);
6568 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
6569 MachinePointerInfo(AS),
6580 EVT IdxVT =
Index.getValueType();
6591 SDValue ID =
DAG.getTargetConstant(IntrinsicID, sdl, MVT::i32);
6594 SDValue Histogram =
DAG.getMaskedHistogram(
DAG.getVTList(MVT::Other), VT, sdl,
6598 DAG.setRoot(Histogram);
6601void SelectionDAGBuilder::visitVectorExtractLastActive(
const CallInst &
I,
6603 assert(Intrinsic == Intrinsic::experimental_vector_extract_last_active &&
6604 "Tried lowering invalid vector extract last");
6606 const DataLayout &Layout =
DAG.getDataLayout();
6610 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
6620 EVT BoolVT =
Mask.getValueType().getScalarType();
6622 Result =
DAG.getSelect(sdl, ResVT, AnyActive, Result, PassThru);
6629void SelectionDAGBuilder::visitIntrinsicCall(
const CallInst &
I,
6631 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
6638 Flags.copyFMF(*FPOp);
6640 switch (Intrinsic) {
6643 visitTargetIntrinsic(
I, Intrinsic);
6645 case Intrinsic::vscale: {
6650 case Intrinsic::vastart: visitVAStart(
I);
return;
6651 case Intrinsic::vaend: visitVAEnd(
I);
return;
6652 case Intrinsic::vacopy: visitVACopy(
I);
return;
6653 case Intrinsic::returnaddress:
6658 case Intrinsic::addressofreturnaddress:
6663 case Intrinsic::sponentry:
6668 case Intrinsic::frameaddress:
6673 case Intrinsic::read_volatile_register:
6674 case Intrinsic::read_register: {
6675 Value *
Reg =
I.getArgOperand(0);
6681 DAG.getVTList(VT, MVT::Other), Chain,
RegName);
6686 case Intrinsic::write_register: {
6687 Value *
Reg =
I.getArgOperand(0);
6688 Value *RegValue =
I.getArgOperand(1);
6696 case Intrinsic::memcpy:
6697 case Intrinsic::memcpy_inline: {
6703 "memcpy_inline needs constant size");
6705 Align DstAlign = MCI.getDestAlign().valueOrOne();
6706 Align SrcAlign = MCI.getSourceAlign().valueOrOne();
6707 Align Alignment = std::min(DstAlign, SrcAlign);
6708 bool isVol = MCI.isVolatile();
6712 SDValue MC =
DAG.getMemcpy(Root, sdl, Dst, Src,
Size, Alignment, isVol,
6713 MCI.isForceInlined(), &
I, std::nullopt,
6714 MachinePointerInfo(
I.getArgOperand(0)),
6715 MachinePointerInfo(
I.getArgOperand(1)),
6717 updateDAGForMaybeTailCall(MC);
6720 case Intrinsic::memset:
6721 case Intrinsic::memset_inline: {
6727 "memset_inline needs constant size");
6729 Align DstAlign = MSII.getDestAlign().valueOrOne();
6730 bool isVol = MSII.isVolatile();
6733 Root, sdl, Dst, Value,
Size, DstAlign, isVol, MSII.isForceInlined(),
6734 &
I, MachinePointerInfo(
I.getArgOperand(0)),
I.getAAMetadata());
6735 updateDAGForMaybeTailCall(MC);
6738 case Intrinsic::memmove: {
6744 Align DstAlign = MMI.getDestAlign().valueOrOne();
6745 Align SrcAlign = MMI.getSourceAlign().valueOrOne();
6746 Align Alignment = std::min(DstAlign, SrcAlign);
6747 bool isVol = MMI.isVolatile();
6751 SDValue MM =
DAG.getMemmove(Root, sdl, Op1, Op2, Op3, Alignment, isVol, &
I,
6753 MachinePointerInfo(
I.getArgOperand(0)),
6754 MachinePointerInfo(
I.getArgOperand(1)),
6756 updateDAGForMaybeTailCall(MM);
6759 case Intrinsic::memcpy_element_unordered_atomic: {
6765 Type *LengthTy =
MI.getLength()->getType();
6766 unsigned ElemSz =
MI.getElementSizeInBytes();
6770 isTC, MachinePointerInfo(
MI.getRawDest()),
6771 MachinePointerInfo(
MI.getRawSource()));
6772 updateDAGForMaybeTailCall(MC);
6775 case Intrinsic::memmove_element_unordered_atomic: {
6781 Type *LengthTy =
MI.getLength()->getType();
6782 unsigned ElemSz =
MI.getElementSizeInBytes();
6786 isTC, MachinePointerInfo(
MI.getRawDest()),
6787 MachinePointerInfo(
MI.getRawSource()));
6788 updateDAGForMaybeTailCall(MC);
6791 case Intrinsic::memset_element_unordered_atomic: {
6797 Type *LengthTy =
MI.getLength()->getType();
6798 unsigned ElemSz =
MI.getElementSizeInBytes();
6802 isTC, MachinePointerInfo(
MI.getRawDest()));
6803 updateDAGForMaybeTailCall(MC);
6806 case Intrinsic::call_preallocated_setup: {
6808 SDValue SrcValue =
DAG.getSrcValue(PreallocatedCall);
6815 case Intrinsic::call_preallocated_arg: {
6817 SDValue SrcValue =
DAG.getSrcValue(PreallocatedCall);
6831 case Intrinsic::eh_typeid_for: {
6834 unsigned TypeID =
DAG.getMachineFunction().getTypeIDFor(GV);
6835 Res =
DAG.getConstant(
TypeID, sdl, MVT::i32);
6840 case Intrinsic::eh_return_i32:
6841 case Intrinsic::eh_return_i64:
6842 DAG.getMachineFunction().setCallsEHReturn(
true);
6849 case Intrinsic::eh_unwind_init:
6850 DAG.getMachineFunction().setCallsUnwindInit(
true);
6852 case Intrinsic::eh_dwarf_cfa:
6857 case Intrinsic::eh_sjlj_callsite: {
6859 assert(
FuncInfo.getCurrentCallSite() == 0 &&
"Overlapping call sites!");
6864 case Intrinsic::eh_sjlj_functioncontext: {
6866 MachineFrameInfo &MFI =
DAG.getMachineFunction().getFrameInfo();
6869 int FI =
FuncInfo.StaticAllocaMap[FnCtx];
6873 case Intrinsic::eh_sjlj_setjmp: {
6878 DAG.getVTList(MVT::i32, MVT::Other),
Ops);
6880 DAG.setRoot(
Op.getValue(1));
6883 case Intrinsic::eh_sjlj_longjmp:
6887 case Intrinsic::eh_sjlj_setup_dispatch:
6891 case Intrinsic::masked_gather:
6892 visitMaskedGather(
I);
6894 case Intrinsic::masked_load:
6897 case Intrinsic::masked_scatter:
6898 visitMaskedScatter(
I);
6900 case Intrinsic::masked_store:
6901 visitMaskedStore(
I);
6903 case Intrinsic::masked_expandload:
6904 visitMaskedLoad(
I,
true );
6906 case Intrinsic::masked_compressstore:
6907 visitMaskedStore(
I,
true );
6909 case Intrinsic::powi:
6913 case Intrinsic::log:
6916 case Intrinsic::log2:
6920 case Intrinsic::log10:
6924 case Intrinsic::exp:
6927 case Intrinsic::exp2:
6931 case Intrinsic::pow:
6935 case Intrinsic::sqrt:
6936 case Intrinsic::fabs:
6937 case Intrinsic::sin:
6938 case Intrinsic::cos:
6939 case Intrinsic::tan:
6940 case Intrinsic::asin:
6941 case Intrinsic::acos:
6942 case Intrinsic::atan:
6943 case Intrinsic::sinh:
6944 case Intrinsic::cosh:
6945 case Intrinsic::tanh:
6946 case Intrinsic::exp10:
6947 case Intrinsic::floor:
6948 case Intrinsic::ceil:
6949 case Intrinsic::trunc:
6950 case Intrinsic::rint:
6951 case Intrinsic::nearbyint:
6952 case Intrinsic::round:
6953 case Intrinsic::roundeven:
6954 case Intrinsic::canonicalize: {
6957 switch (Intrinsic) {
6959 case Intrinsic::sqrt: Opcode =
ISD::FSQRT;
break;
6960 case Intrinsic::fabs: Opcode =
ISD::FABS;
break;
6961 case Intrinsic::sin: Opcode =
ISD::FSIN;
break;
6962 case Intrinsic::cos: Opcode =
ISD::FCOS;
break;
6963 case Intrinsic::tan: Opcode =
ISD::FTAN;
break;
6964 case Intrinsic::asin: Opcode =
ISD::FASIN;
break;
6965 case Intrinsic::acos: Opcode =
ISD::FACOS;
break;
6966 case Intrinsic::atan: Opcode =
ISD::FATAN;
break;
6967 case Intrinsic::sinh: Opcode =
ISD::FSINH;
break;
6968 case Intrinsic::cosh: Opcode =
ISD::FCOSH;
break;
6969 case Intrinsic::tanh: Opcode =
ISD::FTANH;
break;
6970 case Intrinsic::exp10: Opcode =
ISD::FEXP10;
break;
6971 case Intrinsic::floor: Opcode =
ISD::FFLOOR;
break;
6972 case Intrinsic::ceil: Opcode =
ISD::FCEIL;
break;
6973 case Intrinsic::trunc: Opcode =
ISD::FTRUNC;
break;
6974 case Intrinsic::rint: Opcode =
ISD::FRINT;
break;
6976 case Intrinsic::round: Opcode =
ISD::FROUND;
break;
6983 getValue(
I.getArgOperand(0)).getValueType(),
6987 case Intrinsic::atan2:
6989 getValue(
I.getArgOperand(0)).getValueType(),
6993 case Intrinsic::lround:
6994 case Intrinsic::llround:
6995 case Intrinsic::lrint:
6996 case Intrinsic::llrint: {
6999 switch (Intrinsic) {
7001 case Intrinsic::lround: Opcode =
ISD::LROUND;
break;
7003 case Intrinsic::lrint: Opcode =
ISD::LRINT;
break;
7004 case Intrinsic::llrint: Opcode =
ISD::LLRINT;
break;
7013 case Intrinsic::minnum:
7015 getValue(
I.getArgOperand(0)).getValueType(),
7019 case Intrinsic::maxnum:
7021 getValue(
I.getArgOperand(0)).getValueType(),
7025 case Intrinsic::minimum:
7027 getValue(
I.getArgOperand(0)).getValueType(),
7031 case Intrinsic::maximum:
7033 getValue(
I.getArgOperand(0)).getValueType(),
7037 case Intrinsic::minimumnum:
7039 getValue(
I.getArgOperand(0)).getValueType(),
7043 case Intrinsic::maximumnum:
7045 getValue(
I.getArgOperand(0)).getValueType(),
7049 case Intrinsic::copysign:
7051 getValue(
I.getArgOperand(0)).getValueType(),
7055 case Intrinsic::ldexp:
7057 getValue(
I.getArgOperand(0)).getValueType(),
7061 case Intrinsic::modf:
7062 case Intrinsic::sincos:
7063 case Intrinsic::sincospi:
7064 case Intrinsic::frexp: {
7066 switch (Intrinsic) {
7069 case Intrinsic::sincos:
7072 case Intrinsic::sincospi:
7075 case Intrinsic::modf:
7078 case Intrinsic::frexp:
7084 SDVTList VTs =
DAG.getVTList(ValueVTs);
7086 &
I,
DAG.getNode(Opcode, sdl, VTs,
getValue(
I.getArgOperand(0)), Flags));
7089 case Intrinsic::arithmetic_fence: {
7091 getValue(
I.getArgOperand(0)).getValueType(),
7095 case Intrinsic::fma:
7101#define INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC) \
7102 case Intrinsic::INTRINSIC:
7103#include "llvm/IR/ConstrainedOps.def"
7106#define BEGIN_REGISTER_VP_INTRINSIC(VPID, ...) case Intrinsic::VPID:
7107#include "llvm/IR/VPIntrinsics.def"
7110 case Intrinsic::fptrunc_round: {
7114 std::optional<RoundingMode> RoundMode =
7122 SelectionDAG::FlagInserter FlagsInserter(
DAG, Flags);
7127 DAG.getTargetConstant((
int)*RoundMode, sdl, MVT::i32));
7132 case Intrinsic::fmuladd: {
7137 getValue(
I.getArgOperand(0)).getValueType(),
7144 getValue(
I.getArgOperand(0)).getValueType(),
7160 case Intrinsic::fptosi_sat: {
7167 case Intrinsic::fptoui_sat: {
7174 case Intrinsic::convert_from_arbitrary_fp: {
7179 const fltSemantics *SrcSem =
7182 DAG.getContext()->emitError(
7183 "convert_from_arbitrary_fp: not implemented format '" + FormatStr +
7194 DAG.getTargetConstant(
static_cast<int>(SemEnum), sdl, MVT::i32);
7199 case Intrinsic::set_rounding:
7205 case Intrinsic::is_fpclass: {
7206 const DataLayout DLayout =
DAG.getDataLayout();
7208 EVT ArgVT = TLI.
getValueType(DLayout,
I.getArgOperand(0)->getType());
7211 MachineFunction &MF =
DAG.getMachineFunction();
7215 Flags.setNoFPExcept(
7216 !
F.getAttributes().hasFnAttr(llvm::Attribute::StrictFP));
7232 case Intrinsic::get_fpenv: {
7233 const DataLayout DLayout =
DAG.getDataLayout();
7235 Align TempAlign =
DAG.getEVTAlign(EnvVT);
7250 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
7253 Chain =
DAG.getGetFPEnv(Chain, sdl, Temp, EnvVT, MMO);
7254 Res =
DAG.getLoad(EnvVT, sdl, Chain, Temp, MPI);
7260 case Intrinsic::set_fpenv: {
7261 const DataLayout DLayout =
DAG.getDataLayout();
7264 Align TempAlign =
DAG.getEVTAlign(EnvVT);
7277 Chain =
DAG.getStore(Chain, sdl, Env, Temp, MPI, TempAlign,
7279 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
7282 Chain =
DAG.getSetFPEnv(Chain, sdl, Temp, EnvVT, MMO);
7287 case Intrinsic::reset_fpenv:
7290 case Intrinsic::get_fpmode:
7299 case Intrinsic::set_fpmode:
7304 case Intrinsic::reset_fpmode: {
7309 case Intrinsic::pcmarker: {
7314 case Intrinsic::readcyclecounter: {
7317 DAG.getVTList(MVT::i64, MVT::Other),
Op);
7322 case Intrinsic::readsteadycounter: {
7325 DAG.getVTList(MVT::i64, MVT::Other),
Op);
7330 case Intrinsic::bitreverse:
7332 getValue(
I.getArgOperand(0)).getValueType(),
7335 case Intrinsic::bswap:
7337 getValue(
I.getArgOperand(0)).getValueType(),
7340 case Intrinsic::cttz: {
7348 case Intrinsic::ctlz: {
7356 case Intrinsic::ctpop: {
7362 case Intrinsic::fshl:
7363 case Intrinsic::fshr: {
7364 bool IsFSHL =
Intrinsic == Intrinsic::fshl;
7368 EVT VT =
X.getValueType();
7379 case Intrinsic::clmul: {
7385 case Intrinsic::sadd_sat: {
7391 case Intrinsic::uadd_sat: {
7397 case Intrinsic::ssub_sat: {
7403 case Intrinsic::usub_sat: {
7409 case Intrinsic::sshl_sat:
7410 case Intrinsic::ushl_sat: {
7414 EVT ShiftTy =
DAG.getTargetLoweringInfo().getShiftAmountTy(
7419 if (!
I.getType()->isVectorTy() && Op2.
getValueType() != ShiftTy) {
7422 "Unexpected shift type");
7431 case Intrinsic::smul_fix:
7432 case Intrinsic::umul_fix:
7433 case Intrinsic::smul_fix_sat:
7434 case Intrinsic::umul_fix_sat: {
7442 case Intrinsic::sdiv_fix:
7443 case Intrinsic::udiv_fix:
7444 case Intrinsic::sdiv_fix_sat:
7445 case Intrinsic::udiv_fix_sat: {
7450 Op1, Op2, Op3,
DAG, TLI));
7453 case Intrinsic::smax: {
7459 case Intrinsic::smin: {
7465 case Intrinsic::umax: {
7471 case Intrinsic::umin: {
7477 case Intrinsic::abs: {
7483 case Intrinsic::scmp: {
7490 case Intrinsic::ucmp: {
7497 case Intrinsic::stackaddress:
7498 case Intrinsic::stacksave: {
7503 Res =
DAG.getNode(SDOpcode, sdl,
DAG.getVTList(VT, MVT::Other),
Op);
7508 case Intrinsic::stackrestore:
7512 case Intrinsic::get_dynamic_area_offset: {
7521 case Intrinsic::stackguard: {
7522 MachineFunction &MF =
DAG.getMachineFunction();
7528 Res =
DAG.getPtrExtOrTrunc(Res, sdl, PtrTy);
7532 LLVMContext &Ctx = *
DAG.getContext();
7533 Ctx.
diagnose(DiagnosticInfoGeneric(
"unable to lower stackguard"));
7540 MachinePointerInfo(
Global, 0), Align,
7549 case Intrinsic::stackprotector: {
7551 MachineFunction &MF =
DAG.getMachineFunction();
7571 Chain, sdl, Src, FIN,
7578 case Intrinsic::objectsize:
7581 case Intrinsic::is_constant:
7584 case Intrinsic::annotation:
7585 case Intrinsic::ptr_annotation:
7586 case Intrinsic::launder_invariant_group:
7587 case Intrinsic::strip_invariant_group:
7592 case Intrinsic::type_test:
7593 case Intrinsic::public_type_test:
7595 "LowerTypeTests pass before code generation");
7598 case Intrinsic::assume:
7599 case Intrinsic::experimental_noalias_scope_decl:
7600 case Intrinsic::var_annotation:
7601 case Intrinsic::sideeffect:
7606 case Intrinsic::codeview_annotation: {
7608 MachineFunction &MF =
DAG.getMachineFunction();
7617 case Intrinsic::init_trampoline: {
7625 Ops[4] =
DAG.getSrcValue(
I.getArgOperand(0));
7633 case Intrinsic::adjust_trampoline:
7638 case Intrinsic::gcroot: {
7639 assert(
DAG.getMachineFunction().getFunction().hasGC() &&
7640 "only valid in functions with gc specified, enforced by Verifier");
7642 const Value *Alloca =
I.getArgOperand(0)->stripPointerCasts();
7649 case Intrinsic::gcread:
7650 case Intrinsic::gcwrite:
7652 case Intrinsic::get_rounding:
7658 case Intrinsic::expect:
7659 case Intrinsic::expect_with_probability:
7665 case Intrinsic::ubsantrap:
7666 case Intrinsic::debugtrap:
7667 case Intrinsic::trap: {
7668 StringRef TrapFuncName =
7669 I.getAttributes().getFnAttr(
"trap-func-name").getValueAsString();
7670 if (TrapFuncName.
empty()) {
7671 switch (Intrinsic) {
7672 case Intrinsic::trap:
7675 case Intrinsic::debugtrap:
7678 case Intrinsic::ubsantrap:
7681 DAG.getTargetConstant(
7687 DAG.addNoMergeSiteInfo(
DAG.getRoot().getNode(),
7688 I.hasFnAttr(Attribute::NoMerge));
7692 if (Intrinsic == Intrinsic::ubsantrap) {
7693 Value *Arg =
I.getArgOperand(0);
7697 TargetLowering::CallLoweringInfo CLI(
DAG);
7698 CLI.setDebugLoc(sdl).setChain(
getRoot()).setLibCallee(
7700 DAG.getExternalSymbol(TrapFuncName.
data(),
7703 CLI.NoMerge =
I.hasFnAttr(Attribute::NoMerge);
7709 case Intrinsic::allow_runtime_check:
7710 case Intrinsic::allow_ubsan_check:
7714 case Intrinsic::uadd_with_overflow:
7715 case Intrinsic::sadd_with_overflow:
7716 case Intrinsic::usub_with_overflow:
7717 case Intrinsic::ssub_with_overflow:
7718 case Intrinsic::umul_with_overflow:
7719 case Intrinsic::smul_with_overflow: {
7721 switch (Intrinsic) {
7723 case Intrinsic::uadd_with_overflow:
Op =
ISD::UADDO;
break;
7724 case Intrinsic::sadd_with_overflow:
Op =
ISD::SADDO;
break;
7725 case Intrinsic::usub_with_overflow:
Op =
ISD::USUBO;
break;
7726 case Intrinsic::ssub_with_overflow:
Op =
ISD::SSUBO;
break;
7727 case Intrinsic::umul_with_overflow:
Op =
ISD::UMULO;
break;
7728 case Intrinsic::smul_with_overflow:
Op =
ISD::SMULO;
break;
7736 SDVTList VTs =
DAG.getVTList(ResultVT, OverflowVT);
7740 case Intrinsic::prefetch: {
7755 std::nullopt, Flags);
7761 DAG.setRoot(Result);
7764 case Intrinsic::lifetime_start:
7765 case Intrinsic::lifetime_end: {
7766 bool IsStart = (
Intrinsic == Intrinsic::lifetime_start);
7772 if (!LifetimeObject)
7777 auto SI =
FuncInfo.StaticAllocaMap.find(LifetimeObject);
7778 if (SI ==
FuncInfo.StaticAllocaMap.end())
7782 Res =
DAG.getLifetimeNode(IsStart, sdl,
getRoot(), FrameIndex);
7786 case Intrinsic::pseudoprobe: {
7794 case Intrinsic::invariant_start:
7799 case Intrinsic::invariant_end:
7802 case Intrinsic::clear_cache: {
7807 {InputChain, StartVal, EndVal});
7812 case Intrinsic::donothing:
7813 case Intrinsic::seh_try_begin:
7814 case Intrinsic::seh_scope_begin:
7815 case Intrinsic::seh_try_end:
7816 case Intrinsic::seh_scope_end:
7819 case Intrinsic::experimental_stackmap:
7822 case Intrinsic::experimental_patchpoint_void:
7823 case Intrinsic::experimental_patchpoint:
7826 case Intrinsic::experimental_gc_statepoint:
7829 case Intrinsic::experimental_gc_result:
7832 case Intrinsic::experimental_gc_relocate:
7835 case Intrinsic::instrprof_cover:
7837 case Intrinsic::instrprof_increment:
7839 case Intrinsic::instrprof_timestamp:
7841 case Intrinsic::instrprof_value_profile:
7843 case Intrinsic::instrprof_mcdc_parameters:
7845 case Intrinsic::instrprof_mcdc_tvbitmap_update:
7847 case Intrinsic::localescape: {
7848 MachineFunction &MF =
DAG.getMachineFunction();
7849 const TargetInstrInfo *
TII =
DAG.getSubtarget().getInstrInfo();
7853 for (
unsigned Idx = 0,
E =
I.arg_size(); Idx <
E; ++Idx) {
7859 "can only escape static allocas");
7864 TII->get(TargetOpcode::LOCAL_ESCAPE))
7872 case Intrinsic::localrecover: {
7874 MachineFunction &MF =
DAG.getMachineFunction();
7880 unsigned(Idx->getLimitedValue(std::numeric_limits<int>::max()));
7884 Value *
FP =
I.getArgOperand(1);
7890 SDValue OffsetSym =
DAG.getMCSymbol(FrameAllocSym, PtrVT);
7895 SDValue Add =
DAG.getMemBasePlusOffset(FPVal, OffsetVal, sdl);
7901 case Intrinsic::fake_use: {
7902 Value *
V =
I.getArgOperand(0);
7907 auto FakeUseValue = [&]() ->
SDValue {
7921 if (!FakeUseValue || FakeUseValue.isUndef())
7924 Ops[1] = FakeUseValue;
7933 case Intrinsic::reloc_none: {
7938 DAG.getTargetExternalSymbol(
7944 case Intrinsic::cond_loop: {
7954 case Intrinsic::eh_exceptionpointer:
7955 case Intrinsic::eh_exceptioncode: {
7961 SDValue N =
DAG.getCopyFromReg(
DAG.getEntryNode(), sdl, VReg, PtrVT);
7962 if (Intrinsic == Intrinsic::eh_exceptioncode)
7963 N =
DAG.getZExtOrTrunc(
N, sdl, MVT::i32);
7967 case Intrinsic::xray_customevent: {
7970 const auto &Triple =
DAG.getTarget().getTargetTriple();
7979 SDVTList NodeTys =
DAG.getVTList(MVT::Other, MVT::Glue);
7981 Ops.push_back(LogEntryVal);
7982 Ops.push_back(StrSizeVal);
7983 Ops.push_back(Chain);
7989 MachineSDNode *MN =
DAG.getMachineNode(TargetOpcode::PATCHABLE_EVENT_CALL,
7992 DAG.setRoot(patchableNode);
7996 case Intrinsic::xray_typedevent: {
7999 const auto &Triple =
DAG.getTarget().getTargetTriple();
8011 SDVTList NodeTys =
DAG.getVTList(MVT::Other, MVT::Glue);
8013 Ops.push_back(LogTypeId);
8014 Ops.push_back(LogEntryVal);
8015 Ops.push_back(StrSizeVal);
8016 Ops.push_back(Chain);
8022 MachineSDNode *MN =
DAG.getMachineNode(
8023 TargetOpcode::PATCHABLE_TYPED_EVENT_CALL, sdl, NodeTys,
Ops);
8025 DAG.setRoot(patchableNode);
8029 case Intrinsic::experimental_deoptimize:
8032 case Intrinsic::stepvector:
8035 case Intrinsic::vector_reduce_fadd:
8036 case Intrinsic::vector_reduce_fmul:
8037 case Intrinsic::vector_reduce_add:
8038 case Intrinsic::vector_reduce_mul:
8039 case Intrinsic::vector_reduce_and:
8040 case Intrinsic::vector_reduce_or:
8041 case Intrinsic::vector_reduce_xor:
8042 case Intrinsic::vector_reduce_smax:
8043 case Intrinsic::vector_reduce_smin:
8044 case Intrinsic::vector_reduce_umax:
8045 case Intrinsic::vector_reduce_umin:
8046 case Intrinsic::vector_reduce_fmax:
8047 case Intrinsic::vector_reduce_fmin:
8048 case Intrinsic::vector_reduce_fmaximum:
8049 case Intrinsic::vector_reduce_fminimum:
8050 visitVectorReduce(
I, Intrinsic);
8053 case Intrinsic::icall_branch_funnel: {
8059 I.getArgOperand(1),
Offset,
DAG.getDataLayout()));
8062 "llvm.icall.branch.funnel operand must be a GlobalValue");
8063 Ops.push_back(
DAG.getTargetGlobalAddress(
Base, sdl, MVT::i64, 0));
8065 struct BranchFunnelTarget {
8071 for (
unsigned Op = 1,
N =
I.arg_size();
Op !=
N;
Op += 2) {
8074 if (ElemBase !=
Base)
8076 "to the same GlobalValue");
8082 "llvm.icall.branch.funnel operand must be a GlobalValue");
8088 [](
const BranchFunnelTarget &
T1,
const BranchFunnelTarget &T2) {
8089 return T1.Offset < T2.Offset;
8092 for (
auto &
T : Targets) {
8093 Ops.push_back(
DAG.getTargetConstant(
T.Offset, sdl, MVT::i32));
8094 Ops.push_back(
T.Target);
8097 Ops.push_back(
DAG.getRoot());
8098 SDValue N(
DAG.getMachineNode(TargetOpcode::ICALL_BRANCH_FUNNEL, sdl,
8107 case Intrinsic::wasm_landingpad_index:
8113 case Intrinsic::aarch64_settag:
8114 case Intrinsic::aarch64_settag_zero: {
8115 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
8116 bool ZeroMemory =
Intrinsic == Intrinsic::aarch64_settag_zero;
8119 getValue(
I.getArgOperand(1)), MachinePointerInfo(
I.getArgOperand(0)),
8125 case Intrinsic::amdgcn_cs_chain: {
8130 Type *RetTy =
I.getType();
8140 for (
unsigned Idx : {2, 3, 1}) {
8141 TargetLowering::ArgListEntry Arg(
getValue(
I.getOperand(Idx)),
8143 Arg.setAttributes(&
I, Idx);
8144 Args.push_back(Arg);
8147 assert(Args[0].IsInReg &&
"SGPR args should be marked inreg");
8148 assert(!Args[1].IsInReg &&
"VGPR args should not be marked inreg");
8149 Args[2].IsInReg =
true;
8152 for (
unsigned Idx = 4; Idx <
I.arg_size(); ++Idx) {
8153 TargetLowering::ArgListEntry Arg(
getValue(
I.getOperand(Idx)),
8155 Arg.setAttributes(&
I, Idx);
8156 Args.push_back(Arg);
8159 TargetLowering::CallLoweringInfo CLI(
DAG);
8162 .setCallee(CC, RetTy, Callee, std::move(Args))
8165 .setConvergent(
I.isConvergent());
8167 std::pair<SDValue, SDValue>
Result =
8171 "Should've lowered as tail call");
8176 case Intrinsic::amdgcn_call_whole_wave: {
8178 bool isTailCall =
I.isTailCall();
8181 for (
unsigned Idx = 1; Idx <
I.arg_size(); ++Idx) {
8182 TargetLowering::ArgListEntry Arg(
getValue(
I.getArgOperand(Idx)),
8183 I.getArgOperand(Idx)->getType());
8184 Arg.setAttributes(&
I, Idx);
8191 Args.push_back(Arg);
8196 auto *Token = Bundle->Inputs[0].get();
8197 ConvControlToken =
getValue(Token);
8200 TargetLowering::CallLoweringInfo CLI(
DAG);
8204 getValue(
I.getArgOperand(0)), std::move(Args))
8208 .setConvergent(
I.isConvergent())
8209 .setConvergenceControlToken(ConvControlToken);
8212 std::pair<SDValue, SDValue>
Result =
8215 if (
Result.first.getNode())
8219 case Intrinsic::ptrmask: {
8235 auto HighOnes =
DAG.getNode(
8236 ISD::SHL, sdl, PtrVT,
DAG.getAllOnesConstant(sdl, PtrVT),
8237 DAG.getShiftAmountConstant(
Mask.getValueType().getFixedSizeInBits(),
8240 DAG.getZExtOrTrunc(Mask, sdl, PtrVT), HighOnes);
8241 }
else if (
Mask.getValueType() != PtrVT)
8242 Mask =
DAG.getPtrExtOrTrunc(Mask, sdl, PtrVT);
8248 case Intrinsic::threadlocal_address: {
8252 case Intrinsic::get_active_lane_mask: {
8256 EVT ElementVT =
Index.getValueType();
8267 SDValue VectorIndex =
DAG.getSplat(VecTy, sdl, Index);
8268 SDValue VectorTripCount =
DAG.getSplat(VecTy, sdl, TripCount);
8269 SDValue VectorStep =
DAG.getStepVector(sdl, VecTy);
8272 SDValue SetCC =
DAG.getSetCC(sdl, CCVT, VectorInduction,
8277 case Intrinsic::experimental_get_vector_length: {
8279 "Expected positive VF");
8284 EVT CountVT =
Count.getValueType();
8287 visitTargetIntrinsic(
I, Intrinsic);
8296 if (CountVT.
bitsLT(VT)) {
8301 SDValue MaxEVL =
DAG.getElementCount(sdl, CountVT,
8311 case Intrinsic::vector_partial_reduce_add: {
8319 case Intrinsic::vector_partial_reduce_fadd: {
8327 case Intrinsic::experimental_cttz_elts: {
8329 EVT OpVT =
Op.getValueType();
8336 SDValue AllZero =
DAG.getConstant(0, sdl, OpVT);
8345 case Intrinsic::vector_insert: {
8353 if (
Index.getValueType() != VectorIdxTy)
8354 Index =
DAG.getVectorIdxConstant(
Index->getAsZExtVal(), sdl);
8361 case Intrinsic::vector_extract: {
8369 if (
Index.getValueType() != VectorIdxTy)
8370 Index =
DAG.getVectorIdxConstant(
Index->getAsZExtVal(), sdl);
8376 case Intrinsic::experimental_vector_match: {
8382 EVT ResVT =
Mask.getValueType();
8388 visitTargetIntrinsic(
I, Intrinsic);
8392 SDValue Ret =
DAG.getConstant(0, sdl, ResVT);
8394 for (
unsigned i = 0; i < SearchSize; ++i) {
8397 DAG.getVectorIdxConstant(i, sdl));
8400 Ret =
DAG.getNode(
ISD::OR, sdl, ResVT, Ret, Cmp);
8406 case Intrinsic::vector_reverse:
8407 visitVectorReverse(
I);
8409 case Intrinsic::vector_splice_left:
8410 case Intrinsic::vector_splice_right:
8411 visitVectorSplice(
I);
8413 case Intrinsic::callbr_landingpad:
8414 visitCallBrLandingPad(
I);
8416 case Intrinsic::vector_interleave2:
8417 visitVectorInterleave(
I, 2);
8419 case Intrinsic::vector_interleave3:
8420 visitVectorInterleave(
I, 3);
8422 case Intrinsic::vector_interleave4:
8423 visitVectorInterleave(
I, 4);
8425 case Intrinsic::vector_interleave5:
8426 visitVectorInterleave(
I, 5);
8428 case Intrinsic::vector_interleave6:
8429 visitVectorInterleave(
I, 6);
8431 case Intrinsic::vector_interleave7:
8432 visitVectorInterleave(
I, 7);
8434 case Intrinsic::vector_interleave8:
8435 visitVectorInterleave(
I, 8);
8437 case Intrinsic::vector_deinterleave2:
8438 visitVectorDeinterleave(
I, 2);
8440 case Intrinsic::vector_deinterleave3:
8441 visitVectorDeinterleave(
I, 3);
8443 case Intrinsic::vector_deinterleave4:
8444 visitVectorDeinterleave(
I, 4);
8446 case Intrinsic::vector_deinterleave5:
8447 visitVectorDeinterleave(
I, 5);
8449 case Intrinsic::vector_deinterleave6:
8450 visitVectorDeinterleave(
I, 6);
8452 case Intrinsic::vector_deinterleave7:
8453 visitVectorDeinterleave(
I, 7);
8455 case Intrinsic::vector_deinterleave8:
8456 visitVectorDeinterleave(
I, 8);
8458 case Intrinsic::experimental_vector_compress:
8460 getValue(
I.getArgOperand(0)).getValueType(),
8465 case Intrinsic::experimental_convergence_anchor:
8466 case Intrinsic::experimental_convergence_entry:
8467 case Intrinsic::experimental_convergence_loop:
8468 visitConvergenceControl(
I, Intrinsic);
8470 case Intrinsic::experimental_vector_histogram_add: {
8471 visitVectorHistogram(
I, Intrinsic);
8474 case Intrinsic::experimental_vector_extract_last_active: {
8475 visitVectorExtractLastActive(
I, Intrinsic);
8478 case Intrinsic::loop_dependence_war_mask:
8483 DAG.getConstant(0, sdl, MVT::i64)));
8485 case Intrinsic::loop_dependence_raw_mask:
8490 DAG.getConstant(0, sdl, MVT::i64)));
8492 case Intrinsic::masked_udiv:
8498 case Intrinsic::masked_sdiv:
8504 case Intrinsic::masked_urem:
8510 case Intrinsic::masked_srem:
8519void SelectionDAGBuilder::pushFPOpOutChain(
SDValue Result,
8535 PendingConstrainedFP.push_back(OutChain);
8538 PendingConstrainedFPStrict.push_back(OutChain);
8543void SelectionDAGBuilder::visitConstrainedFPIntrinsic(
8557 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8559 SDVTList VTs =
DAG.getVTList(VT, MVT::Other);
8563 Flags.setNoFPExcept(
true);
8566 Flags.copyFMF(*FPOp);
8571#define DAG_INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC, DAGN) \
8572 case Intrinsic::INTRINSIC: \
8573 Opcode = ISD::STRICT_##DAGN; \
8575#include "llvm/IR/ConstrainedOps.def"
8576 case Intrinsic::experimental_constrained_fmuladd: {
8583 pushFPOpOutChain(
Mul, EB);
8606 if (
DAG.isKnownNeverNaN(Opers[1]) &&
DAG.isKnownNeverNaN(Opers[2]))
8614 pushFPOpOutChain(Result, EB);
8621 std::optional<unsigned> ResOPC;
8623 case Intrinsic::vp_ctlz: {
8625 ResOPC = IsZeroUndef ? ISD::VP_CTLZ_ZERO_UNDEF : ISD::VP_CTLZ;
8628 case Intrinsic::vp_cttz: {
8630 ResOPC = IsZeroUndef ? ISD::VP_CTTZ_ZERO_UNDEF : ISD::VP_CTTZ;
8633 case Intrinsic::vp_cttz_elts: {
8635 ResOPC = IsZeroPoison ? ISD::VP_CTTZ_ELTS_ZERO_UNDEF : ISD::VP_CTTZ_ELTS;
8638#define HELPER_MAP_VPID_TO_VPSD(VPID, VPSD) \
8639 case Intrinsic::VPID: \
8640 ResOPC = ISD::VPSD; \
8642#include "llvm/IR/VPIntrinsics.def"
8647 "Inconsistency: no SDNode available for this VPIntrinsic!");
8649 if (*ResOPC == ISD::VP_REDUCE_SEQ_FADD ||
8650 *ResOPC == ISD::VP_REDUCE_SEQ_FMUL) {
8652 return *ResOPC == ISD::VP_REDUCE_SEQ_FADD ? ISD::VP_REDUCE_FADD
8653 : ISD::VP_REDUCE_FMUL;
8659void SelectionDAGBuilder::visitVPLoad(
8671 Alignment =
DAG.getEVTAlign(VT);
8674 SDValue InChain = AddToChain ?
DAG.getRoot() :
DAG.getEntryNode();
8675 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8678 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
8679 MachinePointerInfo(PtrOperand), MMOFlags,
8681 LD =
DAG.getLoadVP(VT,
DL, InChain, OpValues[0], OpValues[1], OpValues[2],
8688void SelectionDAGBuilder::visitVPLoadFF(
8691 assert(OpValues.
size() == 3 &&
"Unexpected number of operands");
8701 Alignment =
DAG.getEVTAlign(VT);
8704 SDValue InChain = AddToChain ?
DAG.getRoot() :
DAG.getEntryNode();
8705 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
8708 LD =
DAG.getLoadFFVP(VT,
DL, InChain, OpValues[0], OpValues[1], OpValues[2],
8713 setValue(&VPIntrin,
DAG.getMergeValues({LD.getValue(0), Trunc},
DL));
8716void SelectionDAGBuilder::visitVPGather(
8720 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8732 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
8734 *Alignment, AAInfo, Ranges);
8744 EVT IdxVT =
Index.getValueType();
8750 LD =
DAG.getGatherVP(
8751 DAG.getVTList(VT, MVT::Other), VT,
DL,
8752 {DAG.getRoot(), Base, Index, Scale, OpValues[1], OpValues[2]}, MMO,
8758void SelectionDAGBuilder::visitVPStore(
8762 EVT VT = OpValues[0].getValueType();
8767 Alignment =
DAG.getEVTAlign(VT);
8770 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8773 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
8774 MachinePointerInfo(PtrOperand), MMOFlags,
8783void SelectionDAGBuilder::visitVPScatter(
8786 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8788 EVT VT = OpValues[0].getValueType();
8798 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
8800 *Alignment, AAInfo);
8810 EVT IdxVT =
Index.getValueType();
8816 ST =
DAG.getScatterVP(
DAG.getVTList(MVT::Other), VT,
DL,
8817 {getMemoryRoot(), OpValues[0], Base, Index, Scale,
8818 OpValues[2], OpValues[3]},
8824void SelectionDAGBuilder::visitVPStridedLoad(
8836 SDValue InChain = AddToChain ?
DAG.getRoot() :
DAG.getEntryNode();
8838 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8841 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
8843 *Alignment, AAInfo, Ranges);
8845 SDValue LD =
DAG.getStridedLoadVP(VT,
DL, InChain, OpValues[0], OpValues[1],
8846 OpValues[2], OpValues[3], MMO,
8854void SelectionDAGBuilder::visitVPStridedStore(
8858 EVT VT = OpValues[0].getValueType();
8864 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8867 MachineMemOperand *MMO =
DAG.getMachineFunction().getMachineMemOperand(
8869 *Alignment, AAInfo);
8873 DAG.getUNDEF(OpValues[1].getValueType()), OpValues[2], OpValues[3],
8881void SelectionDAGBuilder::visitVPCmp(
const VPCmpIntrinsic &VPIntrin) {
8882 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8895 "Unexpected target EVL type");
8900 SimplifyQuery SQ(
DAG.getDataLayout(), &VPIntrin);
8907 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
8910 Condition, MaskOp, EVL));
8913void SelectionDAGBuilder::visitVectorPredicationIntrinsic(
8921 return visitVPCmp(*CmpI);
8924 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
8926 SDVTList VTs =
DAG.getVTList(ValueVTs);
8932 "Unexpected target EVL type");
8936 for (
unsigned I = 0;
I < VPIntrin.
arg_size(); ++
I) {
8938 if (
I == EVLParamPos)
8945 SDNodeFlags SDFlags;
8953 visitVPLoad(VPIntrin, ValueVTs[0], OpValues);
8955 case ISD::VP_LOAD_FF:
8956 visitVPLoadFF(VPIntrin, ValueVTs[0], ValueVTs[1], OpValues);
8958 case ISD::VP_GATHER:
8959 visitVPGather(VPIntrin, ValueVTs[0], OpValues);
8961 case ISD::EXPERIMENTAL_VP_STRIDED_LOAD:
8962 visitVPStridedLoad(VPIntrin, ValueVTs[0], OpValues);
8965 visitVPStore(VPIntrin, OpValues);
8967 case ISD::VP_SCATTER:
8968 visitVPScatter(VPIntrin, OpValues);
8970 case ISD::EXPERIMENTAL_VP_STRIDED_STORE:
8971 visitVPStridedStore(VPIntrin, OpValues);
8973 case ISD::VP_FMULADD: {
8974 assert(OpValues.
size() == 5 &&
"Unexpected number of operands");
8975 SDNodeFlags SDFlags;
8980 setValue(&VPIntrin,
DAG.getNode(ISD::VP_FMA,
DL, VTs, OpValues, SDFlags));
8983 ISD::VP_FMUL,
DL, VTs,
8984 {OpValues[0], OpValues[1], OpValues[3], OpValues[4]}, SDFlags);
8986 DAG.getNode(ISD::VP_FADD,
DL, VTs,
8987 {
Mul, OpValues[2], OpValues[3], OpValues[4]}, SDFlags);
8992 case ISD::VP_IS_FPCLASS: {
8993 const DataLayout DLayout =
DAG.getDataLayout();
8995 auto Constant = OpValues[1]->getAsZExtVal();
8998 {OpValues[0],
Check, OpValues[2], OpValues[3]});
9002 case ISD::VP_INTTOPTR: {
9013 case ISD::VP_PTRTOINT: {
9015 EVT DestVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
9028 case ISD::VP_CTLZ_ZERO_UNDEF:
9030 case ISD::VP_CTTZ_ZERO_UNDEF:
9031 case ISD::VP_CTTZ_ELTS_ZERO_UNDEF:
9032 case ISD::VP_CTTZ_ELTS: {
9034 DAG.getNode(Opcode,
DL, VTs, {OpValues[0], OpValues[2], OpValues[3]});
9052 unsigned CallSiteIndex =
FuncInfo.getCurrentCallSite();
9053 if (CallSiteIndex) {
9067 assert(BeginLabel &&
"BeginLabel should've been set");
9081 assert(
II &&
"II should've been set");
9092std::pair<SDValue, SDValue>
9106 std::pair<SDValue, SDValue> Result = TLI.
LowerCallTo(CLI);
9109 "Non-null chain expected with non-tail call!");
9110 assert((Result.second.getNode() || !Result.first.getNode()) &&
9111 "Null value expected with tail call!");
9113 if (!Result.second.getNode()) {
9120 PendingExports.clear();
9122 DAG.setRoot(Result.second);
9140 if (!isMustTailCall &&
9141 Caller->getFnAttribute(
"disable-tail-calls").getValueAsBool())
9147 if (
DAG.getTargetLoweringInfo().supportSwiftError() &&
9148 Caller->getAttributes().hasAttrSomewhere(Attribute::SwiftError))
9157 bool isTailCall,
bool isMustTailCall,
9160 auto &
DL =
DAG.getDataLayout();
9167 const Value *SwiftErrorVal =
nullptr;
9174 const Value *V = *
I;
9177 if (V->getType()->isEmptyTy())
9182 Entry.setAttributes(&CB,
I - CB.
arg_begin());
9194 Args.push_back(Entry);
9205 Value *V = Bundle->Inputs[0];
9207 Entry.IsCFGuardTarget =
true;
9208 Args.push_back(Entry);
9221 "Target doesn't support calls with kcfi operand bundles.");
9229 auto *Token = Bundle->Inputs[0].get();
9230 ConvControlToken =
getValue(Token);
9241 .
setCallee(RetTy, FTy, Callee, std::move(Args), CB)
9254 "This target doesn't support calls with ptrauth operand bundles.");
9258 std::pair<SDValue, SDValue> Result =
lowerInvokable(CLI, EHPadBB);
9260 if (Result.first.getNode()) {
9275 DAG.setRoot(CopyNode);
9291 LoadTy, Builder.DAG.getDataLayout()))
9292 return Builder.getValue(LoadCst);
9298 bool ConstantMemory =
false;
9301 if (Builder.BatchAA && Builder.BatchAA->pointsToConstantMemory(PtrVal)) {
9302 Root = Builder.DAG.getEntryNode();
9303 ConstantMemory =
true;
9306 Root = Builder.DAG.getRoot();
9311 Builder.DAG.getLoad(LoadVT, Builder.getCurSDLoc(), Root, Ptr,
9314 if (!ConstantMemory)
9315 Builder.PendingLoads.push_back(LoadVal.
getValue(1));
9321void SelectionDAGBuilder::processIntegerCallValue(
const Instruction &
I,
9324 EVT VT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
9335bool SelectionDAGBuilder::visitMemCmpBCmpCall(
const CallInst &
I) {
9336 const Value *
LHS =
I.getArgOperand(0), *
RHS =
I.getArgOperand(1);
9337 const Value *
Size =
I.getArgOperand(2);
9340 EVT CallVT =
DAG.getTargetLoweringInfo().getValueType(
DAG.getDataLayout(),
9346 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9350 if (Res.first.getNode()) {
9351 processIntegerCallValue(
I, Res.first,
true);
9365 auto hasFastLoadsAndCompare = [&](
unsigned NumBits) {
9366 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
9388 switch (NumBitsToCompare) {
9400 LoadVT = hasFastLoadsAndCompare(NumBitsToCompare);
9413 LoadL =
DAG.getBitcast(CmpVT, LoadL);
9414 LoadR =
DAG.getBitcast(CmpVT, LoadR);
9418 processIntegerCallValue(
I, Cmp,
false);
9427bool SelectionDAGBuilder::visitMemChrCall(
const CallInst &
I) {
9428 const Value *Src =
I.getArgOperand(0);
9429 const Value *
Char =
I.getArgOperand(1);
9430 const Value *
Length =
I.getArgOperand(2);
9432 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9433 std::pair<SDValue, SDValue> Res =
9436 MachinePointerInfo(Src));
9437 if (Res.first.getNode()) {
9451bool SelectionDAGBuilder::visitMemCCpyCall(
const CallInst &
I) {
9452 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9459 processIntegerCallValue(
I, Res.first,
true);
9471bool SelectionDAGBuilder::visitMemPCpyCall(
const CallInst &
I) {
9476 Align DstAlign =
DAG.InferPtrAlign(Dst).valueOrOne();
9477 Align SrcAlign =
DAG.InferPtrAlign(Src).valueOrOne();
9479 Align Alignment = std::min(DstAlign, SrcAlign);
9488 Root, sdl, Dst, Src,
Size, Alignment,
false,
false,
nullptr,
9489 std::nullopt, MachinePointerInfo(
I.getArgOperand(0)),
9490 MachinePointerInfo(
I.getArgOperand(1)),
I.getAAMetadata());
9492 "** memcpy should not be lowered as TailCall in mempcpy context **");
9496 Size =
DAG.getSExtOrTrunc(
Size, sdl, Dst.getValueType());
9509bool SelectionDAGBuilder::visitStrCpyCall(
const CallInst &
I,
bool isStpcpy) {
9510 const Value *Arg0 =
I.getArgOperand(0), *Arg1 =
I.getArgOperand(1);
9512 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9515 MachinePointerInfo(Arg0), MachinePointerInfo(Arg1), isStpcpy, &
I);
9516 if (Res.first.getNode()) {
9518 DAG.setRoot(Res.second);
9530bool SelectionDAGBuilder::visitStrCmpCall(
const CallInst &
I) {
9531 const Value *Arg0 =
I.getArgOperand(0), *Arg1 =
I.getArgOperand(1);
9533 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9536 MachinePointerInfo(Arg0), MachinePointerInfo(Arg1), &
I);
9537 if (Res.first.getNode()) {
9538 processIntegerCallValue(
I, Res.first,
true);
9551bool SelectionDAGBuilder::visitStrLenCall(
const CallInst &
I) {
9552 const Value *Arg0 =
I.getArgOperand(0);
9554 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9557 if (Res.first.getNode()) {
9558 processIntegerCallValue(
I, Res.first,
false);
9571bool SelectionDAGBuilder::visitStrNLenCall(
const CallInst &
I) {
9572 const Value *Arg0 =
I.getArgOperand(0), *Arg1 =
I.getArgOperand(1);
9574 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9575 std::pair<SDValue, SDValue> Res =
9578 MachinePointerInfo(Arg0));
9579 if (Res.first.getNode()) {
9580 processIntegerCallValue(
I, Res.first,
false);
9593bool SelectionDAGBuilder::visitStrstrCall(
const CallInst &
I) {
9594 const SelectionDAGTargetInfo &TSI =
DAG.getSelectionDAGInfo();
9595 const Value *Arg0 =
I.getArgOperand(0), *Arg1 =
I.getArgOperand(1);
9599 processIntegerCallValue(
I, Res.first,
false);
9611bool SelectionDAGBuilder::visitUnaryFloatCall(
const CallInst &
I,
9616 if (!
I.onlyReadsMemory() ||
I.isStrictFP())
9633bool SelectionDAGBuilder::visitBinaryFloatCall(
const CallInst &
I,
9638 if (!
I.onlyReadsMemory() ||
I.isStrictFP())
9651void SelectionDAGBuilder::visitCall(
const CallInst &
I) {
9653 if (
I.isInlineAsm()) {
9660 if (Function *
F =
I.getCalledFunction()) {
9661 if (
F->isDeclaration()) {
9663 if (
unsigned IID =
F->getIntrinsicID()) {
9664 visitIntrinsicCall(
I, IID);
9675 if (!
I.isNoBuiltin() && !
F->hasLocalLinkage() &&
F->hasName() &&
9676 LibInfo->getLibFunc(*
F, Func) &&
LibInfo->hasOptimizedCodeGen(Func)) {
9680 if (visitMemCmpBCmpCall(
I))
9683 case LibFunc_copysign:
9684 case LibFunc_copysignf:
9685 case LibFunc_copysignl:
9688 if (
I.onlyReadsMemory()) {
9733 case LibFunc_atan2f:
9734 case LibFunc_atan2l:
9759 case LibFunc_sqrt_finite:
9760 case LibFunc_sqrtf_finite:
9761 case LibFunc_sqrtl_finite:
9778 case LibFunc_exp10f:
9779 case LibFunc_exp10l:
9784 case LibFunc_ldexpf:
9785 case LibFunc_ldexpl:
9789 case LibFunc_strstr:
9790 if (visitStrstrCall(
I))
9793 case LibFunc_memcmp:
9794 if (visitMemCmpBCmpCall(
I))
9797 case LibFunc_memccpy:
9798 if (visitMemCCpyCall(
I))
9801 case LibFunc_mempcpy:
9802 if (visitMemPCpyCall(
I))
9805 case LibFunc_memchr:
9806 if (visitMemChrCall(
I))
9809 case LibFunc_strcpy:
9810 if (visitStrCpyCall(
I,
false))
9813 case LibFunc_stpcpy:
9814 if (visitStrCpyCall(
I,
true))
9817 case LibFunc_strcmp:
9818 if (visitStrCmpCall(
I))
9821 case LibFunc_strlen:
9822 if (visitStrLenCall(
I))
9825 case LibFunc_strnlen:
9826 if (visitStrNLenCall(
I))
9850 if (
I.hasDeoptState())
9867 const Value *Discriminator = PAB->Inputs[1];
9869 assert(
Key->getType()->isIntegerTy(32) &&
"Invalid ptrauth key");
9870 assert(Discriminator->getType()->isIntegerTy(64) &&
9871 "Invalid ptrauth discriminator");
9876 if (CalleeCPA->isKnownCompatibleWith(
Key, Discriminator,
9877 DAG.getDataLayout()))
9917 for (
const auto &Code : Codes)
9932 SDISelAsmOperandInfo &MatchingOpInfo,
9934 if (OpInfo.ConstraintVT == MatchingOpInfo.ConstraintVT)
9940 std::pair<unsigned, const TargetRegisterClass *> MatchRC =
9942 OpInfo.ConstraintVT);
9943 std::pair<unsigned, const TargetRegisterClass *> InputRC =
9945 MatchingOpInfo.ConstraintVT);
9946 const bool OutOpIsIntOrFP =
9947 OpInfo.ConstraintVT.isInteger() || OpInfo.ConstraintVT.isFloatingPoint();
9948 const bool InOpIsIntOrFP = MatchingOpInfo.ConstraintVT.isInteger() ||
9949 MatchingOpInfo.ConstraintVT.isFloatingPoint();
9950 if ((OutOpIsIntOrFP != InOpIsIntOrFP) || (MatchRC.second != InputRC.second)) {
9953 " with a matching output constraint of"
9954 " incompatible type!");
9956 MatchingOpInfo.ConstraintVT = OpInfo.ConstraintVT;
9963 SDISelAsmOperandInfo &OpInfo,
9976 const Value *OpVal = OpInfo.CallOperandVal;
9994 DL.getPrefTypeAlign(Ty),
false,
9997 Chain = DAG.
getTruncStore(Chain, Location, OpInfo.CallOperand, StackSlot,
10000 OpInfo.CallOperand = StackSlot;
10013static std::optional<unsigned>
10015 SDISelAsmOperandInfo &OpInfo,
10016 SDISelAsmOperandInfo &RefOpInfo) {
10027 return std::nullopt;
10031 unsigned AssignedReg;
10034 &
TRI, RefOpInfo.ConstraintCode, RefOpInfo.ConstraintVT);
10037 return std::nullopt;
10042 const MVT RegVT = *
TRI.legalclasstypes_begin(*RC);
10044 if (OpInfo.ConstraintVT != MVT::Other && RegVT != MVT::Untyped) {
10053 !
TRI.isTypeLegalForClass(*RC, OpInfo.ConstraintVT)) {
10058 if (RegVT.
getSizeInBits() == OpInfo.ConstraintVT.getSizeInBits()) {
10063 OpInfo.CallOperand =
10065 OpInfo.ConstraintVT = RegVT;
10069 }
else if (RegVT.
isInteger() && OpInfo.ConstraintVT.isFloatingPoint()) {
10072 OpInfo.CallOperand =
10074 OpInfo.ConstraintVT = VT;
10081 if (OpInfo.isMatchingInputConstraint())
10082 return std::nullopt;
10084 EVT ValueVT = OpInfo.ConstraintVT;
10085 if (OpInfo.ConstraintVT == MVT::Other)
10089 unsigned NumRegs = 1;
10090 if (OpInfo.ConstraintVT != MVT::Other)
10105 I = std::find(
I, RC->
end(), AssignedReg);
10106 if (
I == RC->
end()) {
10109 return {AssignedReg};
10113 for (; NumRegs; --NumRegs, ++
I) {
10114 assert(
I != RC->
end() &&
"Ran out of registers to allocate!");
10119 OpInfo.AssignedRegs =
RegsForValue(Regs, RegVT, ValueVT);
10120 return std::nullopt;
10125 const std::vector<SDValue> &AsmNodeOperands) {
10128 for (; OperandNo; --OperandNo) {
10130 unsigned OpFlag = AsmNodeOperands[CurOp]->getAsZExtVal();
10133 (
F.isRegDefKind() ||
F.isRegDefEarlyClobberKind() ||
F.isMemKind()) &&
10134 "Skipped past definitions?");
10135 CurOp +=
F.getNumOperandRegisters() + 1;
10143 unsigned Flags = 0;
10146 explicit ExtraFlags(
const CallBase &
Call) {
10148 if (
IA->hasSideEffects())
10150 if (
IA->isAlignStack())
10152 if (
IA->canThrow())
10159 void update(
const TargetLowering::AsmOperandInfo &OpInfo) {
10175 unsigned get()
const {
return Flags; }
10199struct ConstraintDecisionInfo {
10201 std::vector<SDValue> AsmNodeOperands;
10203 bool HasSideEffect =
false;
10206 SmallVector<char> Buffer;
10207 raw_svector_ostream ErrorMsg;
10209 ConstraintDecisionInfo() : ErrorMsg(Buffer) {}
10219 ExtraFlags &ExtraInfo) {
10220 for (
auto &
T : TargetConstraints) {
10221 Info.ConstraintOperands.push_back(SDISelAsmOperandInfo(
T));
10222 SDISelAsmOperandInfo &OpInfo = Info.ConstraintOperands.back();
10224 if (OpInfo.CallOperandVal)
10225 OpInfo.CallOperand = Builder.getValue(OpInfo.CallOperandVal);
10227 if (!Info.HasSideEffect)
10228 Info.HasSideEffect = OpInfo.hasMemory(TLI);
10240 Info.ErrorMsg <<
"constraint '" <<
T.ConstraintCode
10241 <<
"' expects an integer constant expression";
10245 ExtraInfo.update(
T);
10259 IA->collectAsmStrs(AsmStrs);
10262 for (SDISelAsmOperandInfo &OpInfo : Info.ConstraintOperands) {
10270 if (OpInfo.hasMatchingInput()) {
10271 SDISelAsmOperandInfo &
Input =
10272 Info.ConstraintOperands[OpInfo.MatchingInput];
10303 if (OpInfo.isIndirect &&
isFunction(OpInfo.CallOperand) &&
10306 OpInfo.isIndirect =
false;
10313 !OpInfo.isIndirect) {
10314 assert((OpInfo.isMultipleAlternative ||
10316 "Can only indirectify direct input operands!");
10323 OpInfo.CallOperandVal =
nullptr;
10326 OpInfo.isIndirect =
true;
10338 SDLoc DL = Builder.getCurSDLoc();
10339 for (SDISelAsmOperandInfo &OpInfo : Info.ConstraintOperands) {
10341 SDISelAsmOperandInfo &RefOpInfo =
10342 OpInfo.isMatchingInputConstraint()
10343 ? Info.ConstraintOperands[OpInfo.getMatchedOperand()]
10349 const char *
RegName =
TRI.getName(*RegError);
10350 Info.ErrorMsg <<
"register '" <<
RegName <<
"' allocated for constraint '"
10351 << OpInfo.ConstraintCode
10352 <<
"' does not match required type";
10356 auto DetectWriteToReservedRegister = [&]() {
10361 if (
Reg.isPhysical() &&
TRI.isInlineAsmReadOnlyReg(MF,
Reg)) {
10362 Info.ErrorMsg <<
"write to reserved register '"
10363 <<
TRI.getRegAsmName(
Reg) <<
"'";
10372 !OpInfo.isMatchingInputConstraint())) &&
10373 "Only address as input operand is allowed.");
10375 switch (OpInfo.Type) {
10381 "Failed to convert memory constraint code to constraint id.");
10386 Info.AsmNodeOperands.push_back(
10388 Info.AsmNodeOperands.push_back(OpInfo.CallOperand);
10393 if (OpInfo.AssignedRegs.Regs.empty()) {
10394 Info.ErrorMsg <<
"could not allocate output register for "
10395 <<
"constraint '" << OpInfo.ConstraintCode <<
"'";
10399 if (DetectWriteToReservedRegister())
10404 OpInfo.AssignedRegs.AddInlineAsmOperands(
10407 false, 0,
DL, DAG, Info.AsmNodeOperands);
10413 SDValue InOperandVal = OpInfo.CallOperand;
10415 if (OpInfo.isMatchingInputConstraint()) {
10419 Info.AsmNodeOperands);
10421 if (Flag.isRegDefKind() || Flag.isRegDefEarlyClobberKind()) {
10422 if (OpInfo.isIndirect) {
10424 Info.ErrorMsg <<
"inline asm not supported yet: cannot handle "
10425 <<
"tied indirect register inputs";
10435 MVT RegVT = R->getSimpleValueType(0);
10439 :
TRI.getMinimalPhysRegClass(TiedReg);
10440 for (
unsigned I = 0,
E = Flag.getNumOperandRegisters();
I !=
E; ++
I)
10447 &Info.Glue, &
Call);
10449 OpInfo.getMatchedOperand(),
DL, DAG,
10450 Info.AsmNodeOperands);
10454 assert(Flag.isMemKind() &&
"Unknown matching constraint!");
10455 assert(Flag.getNumOperandRegisters() == 1 &&
10456 "Unexpected number of operands");
10460 Flag.clearMemConstraint();
10461 Flag.setMatchingOp(OpInfo.getMatchedOperand());
10464 Info.AsmNodeOperands.push_back(Info.AsmNodeOperands[CurOp + 1]);
10475 std::vector<SDValue>
Ops;
10481 Info.ErrorMsg <<
"value out of range for constraint '"
10482 << OpInfo.ConstraintCode <<
"'";
10486 Info.ErrorMsg <<
"invalid operand for inline asm constraint '"
10487 << OpInfo.ConstraintCode <<
"'";
10500 assert((OpInfo.isIndirect ||
10502 "Operand must be indirect to be a mem!");
10505 "Memory operands expect pointer values");
10510 "Failed to convert memory constraint code to constraint id.");
10515 Info.AsmNodeOperands.push_back(
10517 Info.AsmNodeOperands.push_back(InOperandVal);
10525 "Failed to convert memory constraint code to constraint id.");
10529 SDValue AsmOp = InOperandVal;
10541 Info.AsmNodeOperands.push_back(
10543 Info.AsmNodeOperands.push_back(AsmOp);
10549 Info.ErrorMsg <<
"unknown asm constraint '" << OpInfo.ConstraintCode
10555 if (OpInfo.isIndirect) {
10556 Info.ErrorMsg <<
"cannot handle indirect register inputs yet for "
10557 <<
"constraint '" << OpInfo.ConstraintCode <<
"'";
10562 if (OpInfo.AssignedRegs.Regs.empty()) {
10563 Info.ErrorMsg <<
"could not allocate input reg for constraint '"
10564 << OpInfo.ConstraintCode <<
"'";
10568 if (DetectWriteToReservedRegister())
10571 OpInfo.AssignedRegs.getCopyToRegs(InOperandVal, DAG,
DL, Info.Chain,
10572 &Info.Glue, &
Call);
10573 OpInfo.AssignedRegs.AddInlineAsmOperands(
10581 if (!OpInfo.AssignedRegs.Regs.empty())
10582 OpInfo.AssignedRegs.AddInlineAsmOperands(
10599 ExtraFlags ExtraInfo(
Call);
10602 Info.HasSideEffect = IA->hasSideEffects();
10608 Info.Chain = Info.HasSideEffect ? Builder.getRoot() : DAG.
getRoot();
10612 if (IsCallBr || EmitEHLabels)
10616 Info.Chain = Builder.getControlRoot();
10619 Info.Chain = Builder.lowerStartEH(Info.Chain, EHPadBB, Info.BeginLabel);
10625 Info.AsmNodeOperands.push_back(
SDValue());
10632 const MDNode *SrcLoc =
Call.getMetadata(
"srcloc");
10633 Info.AsmNodeOperands.push_back(DAG.
getMDNode(SrcLoc));
10637 Info.AsmNodeOperands.push_back(
10646void SelectionDAGBuilder::visitInlineAsm(
const CallBase &
Call,
10648 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
10650 DAG.getDataLayout(),
DAG.getSubtarget().getRegisterInfo(),
Call);
10653 "InvokeInst must have an EHPadBB");
10655 ConstraintDecisionInfo
Info;
10658 return emitInlineAsmError(
Call,
Info.ErrorMsg.str());
10666 Info.AsmNodeOperands.push_back(Glue);
10672 Info.AsmNodeOperands);
10684 ResultTypes = StructResult->elements();
10685 else if (!CallResultType->
isVoidTy())
10686 ResultTypes =
ArrayRef(CallResultType);
10688 auto CurResultType = ResultTypes.
begin();
10689 auto handleRegAssign = [&](
SDValue V) {
10690 assert(CurResultType != ResultTypes.
end() &&
"Unexpected value");
10691 assert((*CurResultType)->isSized() &&
"Unexpected unsized type");
10692 EVT ResultVT = TLI.
getValueType(
DAG.getDataLayout(), *CurResultType);
10704 if (ResultVT !=
V.getValueType() &&
10707 else if (ResultVT !=
V.getValueType() && ResultVT.
isInteger() &&
10708 V.getValueType().isInteger()) {
10714 assert(ResultVT ==
V.getValueType() &&
"Asm result value mismatch!");
10720 for (SDISelAsmOperandInfo &OpInfo :
Info.ConstraintOperands) {
10724 if (OpInfo.AssignedRegs.
Regs.empty())
10727 switch (OpInfo.ConstraintType) {
10731 Chain, &Glue, &
Call);
10743 assert(
false &&
"Unexpected unknown constraint");
10747 if (OpInfo.isIndirect) {
10748 const Value *Ptr = OpInfo.CallOperandVal;
10749 assert(Ptr &&
"Expected value CallOperandVal for indirect asm operand");
10751 MachinePointerInfo(Ptr));
10758 handleRegAssign(V);
10760 handleRegAssign(Val);
10766 if (!ResultValues.
empty()) {
10767 assert(CurResultType == ResultTypes.
end() &&
10768 "Mismatch in number of ResultTypes");
10770 "Mismatch in number of output operands in asm result");
10773 DAG.getVTList(ResultVTs), ResultValues);
10778 if (!OutChains.
empty())
10782 Chain = lowerEndEH(Chain,
II, EHPadBB,
Info.BeginLabel);
10785 if (ResultValues.
empty() ||
Info.HasSideEffect || !OutChains.
empty() ||
10787 DAG.setRoot(Chain);
10790void SelectionDAGBuilder::emitInlineAsmError(
const CallBase &
Call,
10791 const Twine &Message) {
10792 LLVMContext &Ctx = *
DAG.getContext();
10796 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
10800 if (ValueVTs.
empty())
10804 for (
const EVT &VT : ValueVTs)
10805 Ops.push_back(
DAG.getUNDEF(VT));
10810void SelectionDAGBuilder::visitVAStart(
const CallInst &
I) {
10814 DAG.getSrcValue(
I.getArgOperand(0))));
10817void SelectionDAGBuilder::visitVAArg(
const VAArgInst &
I) {
10818 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
10819 const DataLayout &
DL =
DAG.getDataLayout();
10823 DL.getABITypeAlign(
I.getType()).value());
10824 DAG.setRoot(
V.getValue(1));
10826 if (
I.getType()->isPointerTy())
10827 V =
DAG.getPtrExtOrTrunc(
10832void SelectionDAGBuilder::visitVAEnd(
const CallInst &
I) {
10836 DAG.getSrcValue(
I.getArgOperand(0))));
10839void SelectionDAGBuilder::visitVACopy(
const CallInst &
I) {
10844 DAG.getSrcValue(
I.getArgOperand(0)),
10845 DAG.getSrcValue(
I.getArgOperand(1))));
10851 std::optional<ConstantRange> CR =
getRange(
I);
10853 if (!CR || CR->isFullSet() || CR->isEmptySet() || CR->isUpperWrapped())
10856 APInt Lo = CR->getUnsignedMin();
10857 if (!
Lo.isMinValue())
10860 APInt Hi = CR->getUnsignedMax();
10861 unsigned Bits = std::max(
Hi.getActiveBits(),
10869 DAG.getValueType(SmallVT));
10870 unsigned NumVals =
Op.getNode()->getNumValues();
10876 Ops.push_back(ZExt);
10877 for (
unsigned I = 1;
I != NumVals; ++
I)
10878 Ops.push_back(
Op.getValue(
I));
10880 return DAG.getMergeValues(
Ops,
SL);
10890 SDValue TestConst =
DAG.getTargetConstant(Classes,
SDLoc(), MVT::i32);
10898 for (
unsigned I = 0, E =
Ops.size();
I != E; ++
I) {
10901 MergeOp, TestConst);
10904 return DAG.getMergeValues(
Ops,
SL);
10915 unsigned ArgIdx,
unsigned NumArgs,
SDValue Callee,
Type *ReturnTy,
10918 Args.reserve(NumArgs);
10922 for (
unsigned ArgI = ArgIdx, ArgE = ArgIdx + NumArgs;
10923 ArgI != ArgE; ++ArgI) {
10924 const Value *V =
Call->getOperand(ArgI);
10926 assert(!V->getType()->isEmptyTy() &&
"Empty type passed to intrinsic.");
10929 Entry.setAttributes(
Call, ArgI);
10930 Args.push_back(Entry);
10935 .
setCallee(
Call->getCallingConv(), ReturnTy, Callee, std::move(Args),
10964 for (
unsigned I = StartIdx;
I <
Call.arg_size();
I++) {
10973 Ops.push_back(Builder.getValue(
Call.getArgOperand(
I)));
10979void SelectionDAGBuilder::visitStackmap(
const CallInst &CI) {
11005 Ops.push_back(Chain);
11006 Ops.push_back(InGlue);
11013 assert(
ID.getValueType() == MVT::i64);
11015 DAG.getTargetConstant(
ID->getAsZExtVal(),
DL,
ID.getValueType());
11016 Ops.push_back(IDConst);
11022 Ops.push_back(ShadConst);
11028 SDVTList NodeTys =
DAG.getVTList(MVT::Other, MVT::Glue);
11032 Chain =
DAG.getCALLSEQ_END(Chain, 0, 0, InGlue,
DL);
11037 DAG.setRoot(Chain);
11040 FuncInfo.MF->getFrameInfo().setHasStackMap();
11044void SelectionDAGBuilder::visitPatchpoint(
const CallBase &CB,
11061 Callee =
DAG.getIntPtrConstant(ConstCallee->getZExtValue(), dl,
11064 Callee =
DAG.getTargetGlobalAddress(SymbolicCallee->getGlobal(),
11065 SDLoc(SymbolicCallee),
11066 SymbolicCallee->getValueType(0));
11076 "Not enough arguments provided to the patchpoint intrinsic");
11079 unsigned NumCallArgs = IsAnyRegCC ? 0 : NumArgs;
11083 TargetLowering::CallLoweringInfo CLI(
DAG);
11088 SDNode *CallEnd =
Result.second.getNode();
11097 "Expected a callseq node.");
11099 bool HasGlue =
Call->getGluedNode();
11124 Ops.push_back(Callee);
11130 NumCallRegArgs = IsAnyRegCC ? NumArgs : NumCallRegArgs;
11131 Ops.push_back(
DAG.getTargetConstant(NumCallRegArgs, dl, MVT::i32));
11134 Ops.push_back(
DAG.getTargetConstant((
unsigned)CC, dl, MVT::i32));
11139 for (
unsigned i = NumMetaOpers, e = NumMetaOpers + NumArgs; i !=
e; ++i)
11150 if (IsAnyRegCC && HasDef) {
11152 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
11155 assert(ValueVTs.
size() == 1 &&
"Expected only one return value type.");
11160 NodeTys =
DAG.getVTList(ValueVTs);
11162 NodeTys =
DAG.getVTList(MVT::Other, MVT::Glue);
11179 if (IsAnyRegCC && HasDef) {
11182 DAG.ReplaceAllUsesOfValuesWith(From, To, 2);
11188 FuncInfo.MF->getFrameInfo().setHasPatchPoint();
11191void SelectionDAGBuilder::visitVectorReduce(
const CallInst &
I,
11193 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
11196 if (
I.arg_size() > 1)
11201 SDNodeFlags SDFlags;
11205 switch (Intrinsic) {
11206 case Intrinsic::vector_reduce_fadd:
11214 case Intrinsic::vector_reduce_fmul:
11222 case Intrinsic::vector_reduce_add:
11225 case Intrinsic::vector_reduce_mul:
11228 case Intrinsic::vector_reduce_and:
11231 case Intrinsic::vector_reduce_or:
11234 case Intrinsic::vector_reduce_xor:
11237 case Intrinsic::vector_reduce_smax:
11240 case Intrinsic::vector_reduce_smin:
11243 case Intrinsic::vector_reduce_umax:
11246 case Intrinsic::vector_reduce_umin:
11249 case Intrinsic::vector_reduce_fmax:
11252 case Intrinsic::vector_reduce_fmin:
11255 case Intrinsic::vector_reduce_fmaximum:
11258 case Intrinsic::vector_reduce_fminimum:
11272 Attrs.push_back(Attribute::SExt);
11274 Attrs.push_back(Attribute::ZExt);
11276 Attrs.push_back(Attribute::InReg);
11278 return AttributeList::get(CLI.
RetTy->
getContext(), AttributeList::ReturnIndex,
11286std::pair<SDValue, SDValue>
11300 "Only supported for non-aggregate returns");
11303 for (
Type *Ty : RetOrigTys)
11312 RetOrigTys.
swap(OldRetOrigTys);
11313 RetVTs.
swap(OldRetVTs);
11314 Offsets.swap(OldOffsets);
11316 for (
size_t i = 0, e = OldRetVTs.
size(); i != e; ++i) {
11317 EVT RetVT = OldRetVTs[i];
11321 unsigned RegisterVTByteSZ = RegisterVT.
getSizeInBits() / 8;
11322 RetOrigTys.
append(NumRegs, OldRetOrigTys[i]);
11323 RetVTs.
append(NumRegs, RegisterVT);
11324 for (
unsigned j = 0; j != NumRegs; ++j)
11337 int DemoteStackIdx = -100;
11350 ArgListEntry Entry(DemoteStackSlot, StackSlotPtrType);
11351 Entry.IsSRet =
true;
11352 Entry.Alignment = Alignment;
11364 for (
unsigned I = 0, E = RetVTs.
size();
I != E; ++
I) {
11366 if (NeedsRegBlock) {
11367 Flags.setInConsecutiveRegs();
11368 if (
I == RetVTs.
size() - 1)
11369 Flags.setInConsecutiveRegsLast();
11371 EVT VT = RetVTs[
I];
11375 for (
unsigned i = 0; i != NumRegs; ++i) {
11389 CLI.
Ins.push_back(Ret);
11398 if (Arg.IsSwiftError) {
11404 CLI.
Ins.push_back(Ret);
11412 for (
unsigned i = 0, e = Args.size(); i != e; ++i) {
11416 Type *FinalType = Args[i].Ty;
11417 if (Args[i].IsByVal)
11418 FinalType = Args[i].IndirectType;
11421 for (
unsigned Value = 0, NumValues = OrigArgTys.
size();
Value != NumValues;
11424 Type *ArgTy = OrigArgTy;
11425 if (Args[i].Ty != Args[i].OrigTy) {
11426 assert(
Value == 0 &&
"Only supported for non-aggregate arguments");
11427 ArgTy = Args[i].Ty;
11432 Args[i].Node.getResNo() +
Value);
11439 Flags.setOrigAlign(OriginalAlignment);
11444 Flags.setPointer();
11447 if (Args[i].IsZExt)
11449 if (Args[i].IsSExt)
11451 if (Args[i].IsNoExt)
11453 if (Args[i].IsInReg) {
11460 Flags.setHvaStart();
11466 if (Args[i].IsSRet)
11468 if (Args[i].IsSwiftSelf)
11469 Flags.setSwiftSelf();
11470 if (Args[i].IsSwiftAsync)
11471 Flags.setSwiftAsync();
11472 if (Args[i].IsSwiftError)
11473 Flags.setSwiftError();
11474 if (Args[i].IsCFGuardTarget)
11475 Flags.setCFGuardTarget();
11476 if (Args[i].IsByVal)
11478 if (Args[i].IsByRef)
11480 if (Args[i].IsPreallocated) {
11481 Flags.setPreallocated();
11489 if (Args[i].IsInAlloca) {
11490 Flags.setInAlloca();
11499 if (Args[i].IsByVal || Args[i].IsInAlloca || Args[i].IsPreallocated) {
11500 unsigned FrameSize =
DL.getTypeAllocSize(Args[i].IndirectType);
11501 Flags.setByValSize(FrameSize);
11504 if (
auto MA = Args[i].Alignment)
11508 }
else if (
auto MA = Args[i].Alignment) {
11511 MemAlign = OriginalAlignment;
11513 Flags.setMemAlign(MemAlign);
11514 if (Args[i].IsNest)
11517 Flags.setInConsecutiveRegs();
11520 unsigned NumParts =
11525 if (Args[i].IsSExt)
11527 else if (Args[i].IsZExt)
11532 if (Args[i].IsReturned && !
Op.getValueType().isVector() &&
11537 Args[i].Ty->getPointerAddressSpace())) &&
11538 RetVTs.
size() == NumValues &&
"unexpected use of 'returned'");
11551 CLI.
RetZExt == Args[i].IsZExt))
11552 Flags.setReturned();
11558 for (
unsigned j = 0; j != NumParts; ++j) {
11564 j * Parts[j].
getValueType().getStoreSize().getKnownMinValue());
11565 if (NumParts > 1 && j == 0)
11569 if (j == NumParts - 1)
11573 CLI.
Outs.push_back(MyFlags);
11574 CLI.
OutVals.push_back(Parts[j]);
11577 if (NeedsRegBlock &&
Value == NumValues - 1)
11578 CLI.
Outs[CLI.
Outs.size() - 1].Flags.setInConsecutiveRegsLast();
11590 "LowerCall didn't return a valid chain!");
11592 "LowerCall emitted a return value for a tail call!");
11594 "LowerCall didn't emit the correct number of values!");
11606 for (
unsigned i = 0, e = CLI.
Ins.size(); i != e; ++i) {
11607 assert(InVals[i].
getNode() &&
"LowerCall emitted a null value!");
11608 assert(
EVT(CLI.
Ins[i].VT) == InVals[i].getValueType() &&
11609 "LowerCall emitted a value with the wrong type!");
11619 unsigned NumValues = RetVTs.
size();
11620 ReturnValues.
resize(NumValues);
11627 for (
unsigned i = 0; i < NumValues; ++i) {
11634 DemoteStackIdx, Offsets[i]),
11636 ReturnValues[i] = L;
11637 Chains[i] = L.getValue(1);
11644 std::optional<ISD::NodeType> AssertOp;
11649 unsigned CurReg = 0;
11650 for (
EVT VT : RetVTs) {
11656 CLI.
DAG, CLI.
DL, &InVals[CurReg], NumRegs, RegisterVT, VT,
nullptr,
11664 if (ReturnValues.
empty())
11670 return std::make_pair(Res, CLI.
Chain);
11687 if (
N->getNumValues() == 1) {
11695 "Lowering returned the wrong number of results!");
11698 for (
unsigned I = 0, E =
N->getNumValues();
I != E; ++
I)
11712 "Copy from a reg to the same reg!");
11713 assert(!Reg.isPhysical() &&
"Is a physreg");
11719 RegsForValue RFV(V->getContext(), TLI,
DAG.getDataLayout(), Reg, V->getType(),
11724 auto PreferredExtendIt =
FuncInfo.PreferredExtendType.find(V);
11725 if (PreferredExtendIt !=
FuncInfo.PreferredExtendType.end())
11726 ExtendType = PreferredExtendIt->second;
11729 PendingExports.push_back(Chain);
11741 return A->use_empty();
11743 const BasicBlock &Entry =
A->getParent()->front();
11744 for (
const User *U :
A->users())
11753 std::pair<const AllocaInst *, const StoreInst *>>;
11765 enum StaticAllocaInfo {
Unknown, Clobbered, Elidable };
11767 unsigned NumArgs = FuncInfo->
Fn->
arg_size();
11768 StaticAllocas.
reserve(NumArgs * 2);
11770 auto GetInfoIfStaticAlloca = [&](
const Value *V) -> StaticAllocaInfo * {
11773 V = V->stripPointerCasts();
11775 if (!AI || !AI->isStaticAlloca() || !FuncInfo->
StaticAllocaMap.count(AI))
11778 return &Iter.first->second;
11795 if (
I.isDebugOrPseudoInst())
11799 for (
const Use &U :
I.operands()) {
11800 if (StaticAllocaInfo *Info = GetInfoIfStaticAlloca(U))
11801 *Info = StaticAllocaInfo::Clobbered;
11807 if (StaticAllocaInfo *Info = GetInfoIfStaticAlloca(
SI->getValueOperand()))
11808 *Info = StaticAllocaInfo::Clobbered;
11811 const Value *Dst =
SI->getPointerOperand()->stripPointerCasts();
11812 StaticAllocaInfo *Info = GetInfoIfStaticAlloca(Dst);
11818 if (*Info != StaticAllocaInfo::Unknown)
11826 const Value *Val =
SI->getValueOperand()->stripPointerCasts();
11829 if (!Arg || Arg->hasPassPointeeByValueCopyAttr() ||
11831 DL.getTypeStoreSize(Arg->
getType()) != *AllocaSize ||
11832 !
DL.typeSizeEqualsStoreSize(Arg->
getType()) ||
11833 ArgCopyElisionCandidates.count(Arg)) {
11834 *Info = StaticAllocaInfo::Clobbered;
11838 LLVM_DEBUG(
dbgs() <<
"Found argument copy elision candidate: " << *AI
11842 *Info = StaticAllocaInfo::Elidable;
11843 ArgCopyElisionCandidates.insert({Arg, {AI,
SI}});
11848 if (ArgCopyElisionCandidates.size() == NumArgs)
11872 auto ArgCopyIter = ArgCopyElisionCandidates.find(&Arg);
11873 assert(ArgCopyIter != ArgCopyElisionCandidates.end());
11874 const AllocaInst *AI = ArgCopyIter->second.first;
11875 int FixedIndex = FINode->getIndex();
11877 int OldIndex = AllocaIndex;
11881 dbgs() <<
" argument copy elision failed due to bad fixed stack "
11887 LLVM_DEBUG(
dbgs() <<
" argument copy elision failed: alignment of alloca "
11888 "greater than stack argument alignment ("
11889 <<
DebugStr(RequiredAlignment) <<
" vs "
11897 dbgs() <<
"Eliding argument copy from " << Arg <<
" to " << *AI <<
'\n'
11898 <<
" Replacing frame index " << OldIndex <<
" with " << FixedIndex
11904 AllocaIndex = FixedIndex;
11905 ArgCopyElisionFrameIndexMap.
insert({OldIndex, FixedIndex});
11906 for (
SDValue ArgVal : ArgVals)
11910 const StoreInst *
SI = ArgCopyIter->second.second;
11923void SelectionDAGISel::LowerArguments(
const Function &
F) {
11924 SelectionDAG &DAG =
SDB->DAG;
11925 SDLoc dl =
SDB->getCurSDLoc();
11930 if (
F.hasFnAttribute(Attribute::Naked))
11935 MVT ValueVT =
TLI->getPointerTy(
DL,
DL.getAllocaAddrSpace());
11937 ISD::ArgFlagsTy
Flags;
11939 MVT RegisterVT =
TLI->getRegisterType(*DAG.
getContext(), ValueVT);
11940 ISD::InputArg RetArg(Flags, RegisterVT, ValueVT,
F.getReturnType(),
true,
11950 ArgCopyElisionCandidates);
11953 for (
const Argument &Arg :
F.args()) {
11954 unsigned ArgNo = Arg.getArgNo();
11957 bool isArgValueUsed = !Arg.
use_empty();
11959 if (Arg.hasAttribute(Attribute::ByVal))
11960 FinalType = Arg.getParamByValType();
11961 bool NeedsRegBlock =
TLI->functionArgumentNeedsConsecutiveRegisters(
11962 FinalType,
F.getCallingConv(),
F.isVarArg(),
DL);
11963 for (
unsigned Value = 0, NumValues =
Types.size();
Value != NumValues;
11966 EVT VT =
TLI->getValueType(
DL, ArgTy);
11967 ISD::ArgFlagsTy
Flags;
11970 Flags.setPointer();
11973 if (Arg.hasAttribute(Attribute::ZExt))
11975 if (Arg.hasAttribute(Attribute::SExt))
11977 if (Arg.hasAttribute(Attribute::InReg)) {
11984 Flags.setHvaStart();
11990 if (Arg.hasAttribute(Attribute::StructRet))
11992 if (Arg.hasAttribute(Attribute::SwiftSelf))
11993 Flags.setSwiftSelf();
11994 if (Arg.hasAttribute(Attribute::SwiftAsync))
11995 Flags.setSwiftAsync();
11996 if (Arg.hasAttribute(Attribute::SwiftError))
11997 Flags.setSwiftError();
11998 if (Arg.hasAttribute(Attribute::ByVal))
12000 if (Arg.hasAttribute(Attribute::ByRef))
12002 if (Arg.hasAttribute(Attribute::InAlloca)) {
12003 Flags.setInAlloca();
12011 if (Arg.hasAttribute(Attribute::Preallocated)) {
12012 Flags.setPreallocated();
12024 const Align OriginalAlignment(
12025 TLI->getABIAlignmentForCallingConv(ArgTy,
DL));
12026 Flags.setOrigAlign(OriginalAlignment);
12029 Type *ArgMemTy =
nullptr;
12030 if (
Flags.isByVal() ||
Flags.isInAlloca() ||
Flags.isPreallocated() ||
12033 ArgMemTy = Arg.getPointeeInMemoryValueType();
12035 uint64_t MemSize =
DL.getTypeAllocSize(ArgMemTy);
12040 if (
auto ParamAlign = Arg.getParamStackAlign())
12041 MemAlign = *ParamAlign;
12042 else if ((ParamAlign = Arg.getParamAlign()))
12043 MemAlign = *ParamAlign;
12045 MemAlign =
TLI->getByValTypeAlignment(ArgMemTy,
DL);
12046 if (
Flags.isByRef())
12047 Flags.setByRefSize(MemSize);
12049 Flags.setByValSize(MemSize);
12050 }
else if (
auto ParamAlign = Arg.getParamStackAlign()) {
12051 MemAlign = *ParamAlign;
12053 MemAlign = OriginalAlignment;
12055 Flags.setMemAlign(MemAlign);
12057 if (Arg.hasAttribute(Attribute::Nest))
12060 Flags.setInConsecutiveRegs();
12061 if (ArgCopyElisionCandidates.count(&Arg))
12062 Flags.setCopyElisionCandidate();
12063 if (Arg.hasAttribute(Attribute::Returned))
12064 Flags.setReturned();
12066 MVT RegisterVT =
TLI->getRegisterTypeForCallingConv(
12067 *
CurDAG->getContext(),
F.getCallingConv(), VT);
12068 unsigned NumRegs =
TLI->getNumRegistersForCallingConv(
12069 *
CurDAG->getContext(),
F.getCallingConv(), VT);
12070 for (
unsigned i = 0; i != NumRegs; ++i) {
12074 ISD::InputArg MyFlags(
12075 Flags, RegisterVT, VT, ArgTy, isArgValueUsed, ArgNo,
12077 if (NumRegs > 1 && i == 0)
12078 MyFlags.Flags.setSplit();
12081 MyFlags.Flags.setOrigAlign(
Align(1));
12082 if (i == NumRegs - 1)
12083 MyFlags.Flags.setSplitEnd();
12087 if (NeedsRegBlock &&
Value == NumValues - 1)
12088 Ins[Ins.
size() - 1].Flags.setInConsecutiveRegsLast();
12094 SDValue NewRoot =
TLI->LowerFormalArguments(
12095 DAG.
getRoot(),
F.getCallingConv(),
F.isVarArg(), Ins, dl, DAG, InVals);
12099 "LowerFormalArguments didn't return a valid chain!");
12101 "LowerFormalArguments didn't emit the correct number of values!");
12103 for (
unsigned i = 0, e = Ins.
size(); i != e; ++i) {
12105 "LowerFormalArguments emitted a null value!");
12107 "LowerFormalArguments emitted a value with the wrong type!");
12119 MVT VT =
TLI->getPointerTy(
DL,
DL.getAllocaAddrSpace());
12120 MVT RegVT =
TLI->getRegisterType(*
CurDAG->getContext(), VT);
12121 std::optional<ISD::NodeType> AssertOp;
12124 F.getCallingConv(), AssertOp);
12126 MachineFunction&
MF =
SDB->DAG.getMachineFunction();
12127 MachineRegisterInfo&
RegInfo =
MF.getRegInfo();
12129 RegInfo.createVirtualRegister(
TLI->getRegClassFor(RegVT));
12130 FuncInfo->DemoteRegister = SRetReg;
12132 SDB->DAG.getCopyToReg(NewRoot,
SDB->getCurSDLoc(), SRetReg, ArgValue);
12140 DenseMap<int, int> ArgCopyElisionFrameIndexMap;
12141 for (
const Argument &Arg :
F.args()) {
12145 unsigned NumValues = ValueVTs.
size();
12146 if (NumValues == 0)
12153 if (Ins[i].
Flags.isCopyElisionCandidate()) {
12154 unsigned NumParts = 0;
12155 for (EVT VT : ValueVTs)
12156 NumParts +=
TLI->getNumRegistersForCallingConv(*
CurDAG->getContext(),
12157 F.getCallingConv(), VT);
12161 ArrayRef(&InVals[i], NumParts), ArgHasUses);
12166 bool isSwiftErrorArg =
12167 TLI->supportSwiftError() &&
12168 Arg.hasAttribute(Attribute::SwiftError);
12169 if (!ArgHasUses && !isSwiftErrorArg) {
12170 SDB->setUnusedArgValue(&Arg, InVals[i]);
12173 if (FrameIndexSDNode *FI =
12175 FuncInfo->setArgumentFrameIndex(&Arg, FI->getIndex());
12178 for (
unsigned Val = 0; Val != NumValues; ++Val) {
12179 EVT VT = ValueVTs[Val];
12180 MVT PartVT =
TLI->getRegisterTypeForCallingConv(*
CurDAG->getContext(),
12181 F.getCallingConv(), VT);
12182 unsigned NumParts =
TLI->getNumRegistersForCallingConv(
12183 *
CurDAG->getContext(),
F.getCallingConv(), VT);
12188 if (ArgHasUses || isSwiftErrorArg) {
12189 std::optional<ISD::NodeType> AssertOp;
12190 if (Arg.hasAttribute(Attribute::SExt))
12192 else if (Arg.hasAttribute(Attribute::ZExt))
12197 NewRoot,
F.getCallingConv(), AssertOp);
12200 if (NoFPClass !=
fcNone) {
12202 static_cast<uint64_t
>(NoFPClass), dl, MVT::i32);
12204 OutVal, SDNoFPClass);
12213 if (ArgValues.
empty())
12217 if (FrameIndexSDNode *FI =
12219 FuncInfo->setArgumentFrameIndex(&Arg, FI->getIndex());
12222 SDB->getCurSDLoc());
12224 SDB->setValue(&Arg, Res);
12234 if (LoadSDNode *LNode =
12236 if (FrameIndexSDNode *FI =
12238 FuncInfo->setArgumentFrameIndex(&Arg, FI->getIndex());
12266 FuncInfo->InitializeRegForValue(&Arg);
12267 SDB->CopyToExportRegsIfNeeded(&Arg);
12271 if (!Chains.
empty()) {
12278 assert(i == InVals.
size() &&
"Argument register count mismatch!");
12282 if (!ArgCopyElisionFrameIndexMap.
empty()) {
12283 for (MachineFunction::VariableDbgInfo &VI :
12284 MF->getInStackSlotVariableDbgInfo()) {
12285 auto I = ArgCopyElisionFrameIndexMap.
find(
VI.getStackSlot());
12286 if (
I != ArgCopyElisionFrameIndexMap.
end())
12287 VI.updateStackSlot(
I->second);
12302SelectionDAGBuilder::HandlePHINodesInSuccessorBlocks(
const BasicBlock *LLVMBB) {
12303 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
12305 SmallPtrSet<MachineBasicBlock *, 4> SuccsHandled;
12311 MachineBasicBlock *SuccMBB =
FuncInfo.getMBB(SuccBB);
12315 if (!SuccsHandled.
insert(SuccMBB).second)
12323 for (
const PHINode &PN : SuccBB->phis()) {
12325 if (PN.use_empty())
12329 if (PN.getType()->isEmptyTy())
12333 const Value *PHIOp = PN.getIncomingValueForBlock(LLVMBB);
12338 RegOut =
FuncInfo.CreateRegs(&PN);
12356 "Didn't codegen value into a register!??");
12366 for (EVT VT : ValueVTs) {
12368 for (
unsigned i = 0; i != NumRegisters; ++i)
12370 Reg += NumRegisters;
12390void SelectionDAGBuilder::updateDAGForMaybeTailCall(
SDValue MaybeTC) {
12392 if (MaybeTC.
getNode() !=
nullptr)
12393 DAG.setRoot(MaybeTC);
12398void SelectionDAGBuilder::lowerWorkItem(SwitchWorkListItem W,
Value *
Cond,
12401 MachineFunction *CurMF =
FuncInfo.MF;
12402 MachineBasicBlock *NextMBB =
nullptr;
12407 unsigned Size =
W.LastCluster -
W.FirstCluster + 1;
12409 BranchProbabilityInfo *BPI =
FuncInfo.BPI;
12411 if (
Size == 2 &&
W.MBB == SwitchMBB) {
12419 CaseCluster &
Small = *
W.FirstCluster;
12420 CaseCluster &
Big = *
W.LastCluster;
12424 const APInt &SmallValue =
Small.Low->getValue();
12425 const APInt &BigValue =
Big.Low->getValue();
12428 APInt CommonBit = BigValue ^ SmallValue;
12435 DAG.getConstant(CommonBit,
DL, VT));
12437 DL, MVT::i1,
Or,
DAG.getConstant(BigValue | SmallValue,
DL, VT),
12443 addSuccessorWithProb(SwitchMBB,
Small.MBB,
Small.Prob +
Big.Prob);
12445 addSuccessorWithProb(
12446 SwitchMBB, DefaultMBB,
12450 addSuccessorWithProb(SwitchMBB, DefaultMBB);
12458 DAG.getBasicBlock(DefaultMBB));
12460 DAG.setRoot(BrCond);
12472 [](
const CaseCluster &a,
const CaseCluster &b) {
12473 return a.Prob != b.Prob ?
12475 a.Low->getValue().slt(b.Low->getValue());
12482 if (
I->Prob >
W.LastCluster->Prob)
12484 if (
I->Kind ==
CC_Range &&
I->MBB == NextMBB) {
12492 BranchProbability DefaultProb =
W.DefaultProb;
12493 BranchProbability UnhandledProbs = DefaultProb;
12495 UnhandledProbs +=
I->Prob;
12497 MachineBasicBlock *CurMBB =
W.MBB;
12499 bool FallthroughUnreachable =
false;
12500 MachineBasicBlock *Fallthrough;
12501 if (
I ==
W.LastCluster) {
12503 Fallthrough = DefaultMBB;
12508 CurMF->
insert(BBI, Fallthrough);
12512 UnhandledProbs -=
I->Prob;
12517 JumpTableHeader *JTH = &
SL->JTCases[
I->JTCasesIndex].first;
12518 SwitchCG::JumpTable *JT = &
SL->JTCases[
I->JTCasesIndex].second;
12521 MachineBasicBlock *JumpMBB = JT->
MBB;
12522 CurMF->
insert(BBI, JumpMBB);
12524 auto JumpProb =
I->Prob;
12525 auto FallthroughProb = UnhandledProbs;
12533 if (*SI == DefaultMBB) {
12534 JumpProb += DefaultProb / 2;
12535 FallthroughProb -= DefaultProb / 2;
12553 if (FallthroughUnreachable) {
12560 addSuccessorWithProb(CurMBB, Fallthrough, FallthroughProb);
12561 addSuccessorWithProb(CurMBB, JumpMBB, JumpProb);
12570 if (CurMBB == SwitchMBB) {
12578 BitTestBlock *BTB = &
SL->BitTestCases[
I->BTCasesIndex];
12581 for (BitTestCase &BTC : BTB->
Cases)
12593 BTB->
Prob += DefaultProb / 2;
12597 if (FallthroughUnreachable)
12601 if (CurMBB == SwitchMBB) {
12608 const Value *
RHS, *
LHS, *MHS;
12610 if (
I->Low ==
I->High) {
12625 if (FallthroughUnreachable)
12629 CaseBlock CB(CC,
LHS,
RHS, MHS,
I->MBB, Fallthrough, CurMBB,
12632 if (CurMBB == SwitchMBB)
12635 SL->SwitchCases.push_back(CB);
12640 CurMBB = Fallthrough;
12644void SelectionDAGBuilder::splitWorkItem(
SwitchWorkList &WorkList,
12645 const SwitchWorkListItem &W,
12648 assert(
W.FirstCluster->Low->getValue().slt(
W.LastCluster->Low->getValue()) &&
12649 "Clusters not sorted?");
12650 assert(
W.LastCluster -
W.FirstCluster + 1 >= 2 &&
"Too small to split!");
12652 auto [LastLeft, FirstRight, LeftProb, RightProb] =
12653 SL->computeSplitWorkItemInfo(W);
12658 assert(PivotCluster >
W.FirstCluster);
12659 assert(PivotCluster <=
W.LastCluster);
12664 const ConstantInt *Pivot = PivotCluster->Low;
12673 MachineBasicBlock *LeftMBB;
12674 if (FirstLeft == LastLeft && FirstLeft->Kind ==
CC_Range &&
12675 FirstLeft->Low ==
W.GE &&
12676 (FirstLeft->High->getValue() + 1LL) == Pivot->
getValue()) {
12677 LeftMBB = FirstLeft->MBB;
12679 LeftMBB =
FuncInfo.MF->CreateMachineBasicBlock(
W.MBB->getBasicBlock());
12680 FuncInfo.MF->insert(BBI, LeftMBB);
12682 {LeftMBB, FirstLeft, LastLeft,
W.GE, Pivot,
W.DefaultProb / 2});
12690 MachineBasicBlock *RightMBB;
12691 if (FirstRight == LastRight && FirstRight->Kind ==
CC_Range &&
12692 W.LT && (FirstRight->High->getValue() + 1ULL) ==
W.LT->getValue()) {
12693 RightMBB = FirstRight->MBB;
12695 RightMBB =
FuncInfo.MF->CreateMachineBasicBlock(
W.MBB->getBasicBlock());
12696 FuncInfo.MF->insert(BBI, RightMBB);
12698 {RightMBB, FirstRight, LastRight, Pivot,
W.LT,
W.DefaultProb / 2});
12704 CaseBlock CB(
ISD::SETLT,
Cond, Pivot,
nullptr, LeftMBB, RightMBB,
W.MBB,
12707 if (
W.MBB == SwitchMBB)
12710 SL->SwitchCases.push_back(CB);
12735 MachineBasicBlock *SwitchMBB =
FuncInfo.MBB;
12743 unsigned PeeledCaseIndex = 0;
12744 bool SwitchPeeled =
false;
12745 for (
unsigned Index = 0;
Index < Clusters.size(); ++
Index) {
12746 CaseCluster &CC = Clusters[
Index];
12747 if (CC.
Prob < TopCaseProb)
12749 TopCaseProb = CC.
Prob;
12750 PeeledCaseIndex =
Index;
12751 SwitchPeeled =
true;
12756 LLVM_DEBUG(
dbgs() <<
"Peeled one top case in switch stmt, prob: "
12757 << TopCaseProb <<
"\n");
12762 MachineBasicBlock *PeeledSwitchMBB =
12764 FuncInfo.MF->insert(BBI, PeeledSwitchMBB);
12767 auto PeeledCaseIt = Clusters.begin() + PeeledCaseIndex;
12768 SwitchWorkListItem
W = {SwitchMBB, PeeledCaseIt, PeeledCaseIt,
12769 nullptr,
nullptr, TopCaseProb.
getCompl()};
12770 lowerWorkItem(W,
SI.getCondition(), SwitchMBB, PeeledSwitchMBB);
12772 Clusters.erase(PeeledCaseIt);
12773 for (CaseCluster &CC : Clusters) {
12775 dbgs() <<
"Scale the probablity for one cluster, before scaling: "
12776 << CC.
Prob <<
"\n");
12780 PeeledCaseProb = TopCaseProb;
12781 return PeeledSwitchMBB;
12784void SelectionDAGBuilder::visitSwitch(
const SwitchInst &
SI) {
12786 BranchProbabilityInfo *BPI =
FuncInfo.BPI;
12788 Clusters.reserve(
SI.getNumCases());
12789 for (
auto I :
SI.cases()) {
12790 MachineBasicBlock *Succ =
FuncInfo.getMBB(
I.getCaseSuccessor());
12791 const ConstantInt *CaseVal =
I.getCaseValue();
12792 BranchProbability Prob =
12794 : BranchProbability(1,
SI.getNumCases() + 1);
12798 MachineBasicBlock *DefaultMBB =
FuncInfo.getMBB(
SI.getDefaultDest());
12807 MachineBasicBlock *PeeledSwitchMBB =
12808 peelDominantCaseCluster(SI, Clusters, PeeledCaseProb);
12811 MachineBasicBlock *SwitchMBB =
FuncInfo.MBB;
12812 if (Clusters.empty()) {
12813 assert(PeeledSwitchMBB == SwitchMBB);
12815 if (DefaultMBB != NextBlock(SwitchMBB)) {
12822 SL->findJumpTables(Clusters, &SI,
getCurSDLoc(), DefaultMBB,
DAG.getPSI(),
12824 SL->findBitTestClusters(Clusters, &SI);
12827 dbgs() <<
"Case clusters: ";
12828 for (
const CaseCluster &
C : Clusters) {
12834 C.Low->getValue().print(
dbgs(),
true);
12835 if (
C.Low !=
C.High) {
12837 C.High->getValue().print(
dbgs(),
true);
12844 assert(!Clusters.empty());
12848 auto DefaultProb = getEdgeProbability(PeeledSwitchMBB, DefaultMBB);
12852 DefaultMBB ==
FuncInfo.getMBB(
SI.getDefaultDest()))
12855 {PeeledSwitchMBB,
First,
Last,
nullptr,
nullptr, DefaultProb});
12857 while (!WorkList.
empty()) {
12859 unsigned NumClusters =
W.LastCluster -
W.FirstCluster + 1;
12864 splitWorkItem(WorkList, W,
SI.getCondition(), SwitchMBB);
12868 lowerWorkItem(W,
SI.getCondition(), SwitchMBB, DefaultMBB);
12872void SelectionDAGBuilder::visitStepVector(
const CallInst &
I) {
12873 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
12879void SelectionDAGBuilder::visitVectorReverse(
const CallInst &
I) {
12880 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
12885 assert(VT ==
V.getValueType() &&
"Malformed vector.reverse!");
12894 SmallVector<int, 8>
Mask;
12896 for (
unsigned i = 0; i != NumElts; ++i)
12897 Mask.push_back(NumElts - 1 - i);
12902void SelectionDAGBuilder::visitVectorDeinterleave(
const CallInst &
I,
12911 EVT OutVT = ValueVTs[0];
12915 for (
unsigned i = 0; i != Factor; ++i) {
12916 assert(ValueVTs[i] == OutVT &&
"Expected VTs to be the same");
12918 DAG.getVectorIdxConstant(OutNumElts * i,
DL));
12924 SDValue Even =
DAG.getVectorShuffle(OutVT,
DL, SubVecs[0], SubVecs[1],
12926 SDValue Odd =
DAG.getVectorShuffle(OutVT,
DL, SubVecs[0], SubVecs[1],
12934 DAG.getVTList(ValueVTs), SubVecs);
12938void SelectionDAGBuilder::visitVectorInterleave(
const CallInst &
I,
12941 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
12946 for (
unsigned i = 0; i < Factor; ++i) {
12949 "Expected VTs to be the same");
12967 for (
unsigned i = 0; i < Factor; ++i)
12974void SelectionDAGBuilder::visitFreeze(
const FreezeInst &
I) {
12978 unsigned NumValues = ValueVTs.
size();
12979 if (NumValues == 0)
return;
12984 for (
unsigned i = 0; i != NumValues; ++i)
12989 DAG.getVTList(ValueVTs), Values));
12992void SelectionDAGBuilder::visitVectorSplice(
const CallInst &
I) {
12993 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
12999 const bool IsLeft =
I.getIntrinsicID() == Intrinsic::vector_splice_left;
13014 uint64_t Idx = IsLeft ?
Imm : NumElts -
Imm;
13017 SmallVector<int, 8>
Mask;
13018 for (
unsigned i = 0; i < NumElts; ++i)
13019 Mask.push_back(Idx + i);
13047 assert(
MI->getOpcode() == TargetOpcode::COPY &&
13048 "start of copy chain MUST be COPY");
13049 Reg =
MI->getOperand(1).getReg();
13052 assert(
Reg.isVirtual() &&
"expected COPY of virtual register");
13056 if (
MI->getOpcode() == TargetOpcode::COPY) {
13057 assert(
Reg.isVirtual() &&
"expected COPY of virtual register");
13058 Reg =
MI->getOperand(1).getReg();
13059 assert(
Reg.isPhysical() &&
"expected COPY of physical register");
13062 assert(
MI->getOpcode() == TargetOpcode::INLINEASM_BR &&
13063 "end of copy chain MUST be INLINEASM_BR");
13073void SelectionDAGBuilder::visitCallBrLandingPad(
const CallInst &
I) {
13079 const TargetLowering &TLI =
DAG.getTargetLoweringInfo();
13080 const TargetRegisterInfo *
TRI =
DAG.getSubtarget().getRegisterInfo();
13081 MachineRegisterInfo &MRI =
DAG.getMachineFunction().getRegInfo();
13089 for (
auto &
T : TargetConstraints) {
13090 SDISelAsmOperandInfo OpInfo(
T);
13098 switch (OpInfo.ConstraintType) {
13109 FuncInfo.MBB->addLiveIn(OriginalDef);
13117 ResultVTs.
push_back(OpInfo.ConstraintVT);
13126 ResultVTs.
push_back(OpInfo.ConstraintVT);
13134 DAG.getVTList(ResultVTs), ResultValues);
static unsigned getIntrinsicID(const SDNode *N)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static msgpack::DocNode getNode(msgpack::DocNode DN, msgpack::Type Type, MCValue Val)
This file declares a class to represent arbitrary precision floating point values and provide a varie...
This file implements a class to represent arbitrary precision integral constant values and operations...
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
Function Alias Analysis Results
Atomic ordering constants.
This file contains the simple types necessary to represent the attributes associated with functions a...
static const Function * getParent(const Value *V)
This file implements the BitVector class.
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
This file contains the declarations for the subclasses of Constant, which represent the different fla...
static AttributeList getReturnAttrs(FastISel::CallLoweringInfo &CLI)
Returns an AttributeList representing the attributes applied to the return value of the given call.
static Value * getCondition(Instruction *I)
const HexagonInstrInfo * TII
This file provides various utilities for inspecting and working with the control flow graph in LLVM I...
Module.h This file contains the declarations for the Module class.
static void getRegistersForValue(MachineFunction &MF, MachineIRBuilder &MIRBuilder, GISelAsmOperandInfo &OpInfo, GISelAsmOperandInfo &RefOpInfo)
Assign virtual/physical registers for the specified register operand.
static void computeConstraintToUse(const TargetLowering *TLI, TargetLowering::AsmOperandInfo &OpInfo)
This file defines an InstructionCost class that is used when calculating the cost of an instruction,...
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
Machine Check Debug Module
static bool isUndef(const MachineInstr &MI)
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static const Function * getCalledFunction(const Value *V)
This file provides utility analysis objects describing memory locations.
This file provides utility for Memory Model Relaxation Annotations (MMRAs).
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
static unsigned getAddressSpace(const Value *V, unsigned MaxLookup)
MachineInstr unsigned OpIdx
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
uint64_t IntrinsicInst * II
OptimizedStructLayoutField Field
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
static Type * getValueType(Value *V, bool LookThroughCmp=false)
Returns the "element type" of the given value/instruction V.
static bool hasOnlySelectUsers(const Value *Cond)
static SDValue getLoadStackGuard(SelectionDAG &DAG, const SDLoc &DL, SDValue &Chain)
Create a LOAD_STACK_GUARD node, and let it carry the target specific global variable if there exists ...
static bool getUniformBase(const Value *Ptr, SDValue &Base, SDValue &Index, SDValue &Scale, SelectionDAGBuilder *SDB, const BasicBlock *CurBB, uint64_t ElemSize)
static void failForInvalidBundles(const CallBase &I, StringRef Name, ArrayRef< uint32_t > AllowedBundles)
static void addStackMapLiveVars(const CallBase &Call, unsigned StartIdx, const SDLoc &DL, SmallVectorImpl< SDValue > &Ops, SelectionDAGBuilder &Builder)
Add a stack map intrinsic call's live variable operands to a stackmap or patchpoint target node's ope...
static const unsigned MaxParallelChains
static SDValue expandPow(const SDLoc &dl, SDValue LHS, SDValue RHS, SelectionDAG &DAG, const TargetLowering &TLI, SDNodeFlags Flags)
visitPow - Lower a pow intrinsic.
static const CallBase * FindPreallocatedCall(const Value *PreallocatedSetup)
Given a @llvm.call.preallocated.setup, return the corresponding preallocated call.
static cl::opt< unsigned > SwitchPeelThreshold("switch-peel-threshold", cl::Hidden, cl::init(66), cl::desc("Set the case probability threshold for peeling the case from a " "switch statement. A value greater than 100 will void this " "optimization"))
static cl::opt< bool > InsertAssertAlign("insert-assert-align", cl::init(true), cl::desc("Insert the experimental `assertalign` node."), cl::ReallyHidden)
static unsigned getISDForVPIntrinsic(const VPIntrinsic &VPIntrin)
static bool handleDanglingVariadicDebugInfo(SelectionDAG &DAG, DILocalVariable *Variable, DebugLoc DL, unsigned Order, SmallVectorImpl< Value * > &Values, DIExpression *Expression)
static bool prepareDAGLevelOperands(ConstraintDecisionInfo &Info, const CallBase &Call, SelectionDAGBuilder &Builder, const TargetLowering &TLI, SelectionDAG &DAG)
Prepare DAG-level operands.
static unsigned findMatchingInlineAsmOperand(unsigned OperandNo, const std::vector< SDValue > &AsmNodeOperands)
static void patchMatchingInput(const SDISelAsmOperandInfo &OpInfo, SDISelAsmOperandInfo &MatchingOpInfo, SelectionDAG &DAG)
Make sure that the output operand OpInfo and its corresponding input operand MatchingOpInfo have comp...
static void findUnwindDestinations(FunctionLoweringInfo &FuncInfo, const BasicBlock *EHPadBB, BranchProbability Prob, SmallVectorImpl< std::pair< MachineBasicBlock *, BranchProbability > > &UnwindDests)
When an invoke or a cleanupret unwinds to the next EH pad, there are many places it could ultimately ...
static unsigned FixedPointIntrinsicToOpcode(unsigned Intrinsic)
static BranchProbability scaleCaseProbality(BranchProbability CaseProb, BranchProbability PeeledCaseProb)
static SDValue expandExp2(const SDLoc &dl, SDValue Op, SelectionDAG &DAG, const TargetLowering &TLI, SDNodeFlags Flags)
expandExp2 - Lower an exp2 intrinsic.
static SDValue expandDivFix(unsigned Opcode, const SDLoc &DL, SDValue LHS, SDValue RHS, SDValue Scale, SelectionDAG &DAG, const TargetLowering &TLI)
static SDValue getF32Constant(SelectionDAG &DAG, unsigned Flt, const SDLoc &dl)
getF32Constant - Get 32-bit floating point constant.
static SDValue widenVectorToPartType(SelectionDAG &DAG, SDValue Val, const SDLoc &DL, EVT PartVT)
static SDValue expandLog10(const SDLoc &dl, SDValue Op, SelectionDAG &DAG, const TargetLowering &TLI, SDNodeFlags Flags)
expandLog10 - Lower a log10 intrinsic.
DenseMap< const Argument *, std::pair< const AllocaInst *, const StoreInst * > > ArgCopyElisionMapTy
static void getCopyToPartsVector(SelectionDAG &DAG, const SDLoc &dl, SDValue Val, SDValue *Parts, unsigned NumParts, MVT PartVT, const Value *V, std::optional< CallingConv::ID > CallConv)
getCopyToPartsVector - Create a series of nodes that contain the specified value split into legal par...
static void getUnderlyingArgRegs(SmallVectorImpl< std::pair< Register, TypeSize > > &Regs, const SDValue &N)
static void getCopyToParts(SelectionDAG &DAG, const SDLoc &DL, SDValue Val, SDValue *Parts, unsigned NumParts, MVT PartVT, const Value *V, std::optional< CallingConv::ID > CallConv=std::nullopt, ISD::NodeType ExtendKind=ISD::ANY_EXTEND)
getCopyToParts - Create a series of nodes that contain the specified value split into legal parts.
static SDValue getMemCmpLoad(const Value *PtrVal, MVT LoadVT, SelectionDAGBuilder &Builder)
static SDValue expandLog2(const SDLoc &dl, SDValue Op, SelectionDAG &DAG, const TargetLowering &TLI, SDNodeFlags Flags)
expandLog2 - Lower a log2 intrinsic.
static SDValue getAddressForMemoryInput(SDValue Chain, const SDLoc &Location, SDISelAsmOperandInfo &OpInfo, SelectionDAG &DAG)
Get a direct memory input to behave well as an indirect operand.
static bool isOnlyUsedInEntryBlock(const Argument *A, bool FastISel)
isOnlyUsedInEntryBlock - If the specified argument is only used in the entry block,...
static void diagnosePossiblyInvalidConstraint(LLVMContext &Ctx, const Value *V, const Twine &ErrMsg)
static bool collectInstructionDeps(SmallMapVector< const Instruction *, bool, 8 > *Deps, const Value *V, SmallMapVector< const Instruction *, bool, 8 > *Necessary=nullptr, unsigned Depth=0)
static void findArgumentCopyElisionCandidates(const DataLayout &DL, FunctionLoweringInfo *FuncInfo, ArgCopyElisionMapTy &ArgCopyElisionCandidates)
Scan the entry block of the function in FuncInfo for arguments that look like copies into a local all...
static bool isFunction(SDValue Op)
static SDValue GetExponent(SelectionDAG &DAG, SDValue Op, const TargetLowering &TLI, const SDLoc &dl)
GetExponent - Get the exponent:
static Register FollowCopyChain(MachineRegisterInfo &MRI, Register Reg)
static SDValue ExpandPowI(const SDLoc &DL, SDValue LHS, SDValue RHS, SelectionDAG &DAG)
ExpandPowI - Expand a llvm.powi intrinsic.
static SDValue expandLog(const SDLoc &dl, SDValue Op, SelectionDAG &DAG, const TargetLowering &TLI, SDNodeFlags Flags)
expandLog - Lower a log intrinsic.
static SDValue getCopyFromParts(SelectionDAG &DAG, const SDLoc &DL, const SDValue *Parts, unsigned NumParts, MVT PartVT, EVT ValueVT, const Value *V, SDValue InChain, std::optional< CallingConv::ID > CC=std::nullopt, std::optional< ISD::NodeType > AssertOp=std::nullopt)
getCopyFromParts - Create a value that contains the specified legal parts combined into the value the...
static SDValue getLimitedPrecisionExp2(SDValue t0, const SDLoc &dl, SelectionDAG &DAG)
static bool determineConstraints(ConstraintDecisionInfo &Info, TargetLowering::AsmOperandInfoVector &TargetConstraints, const CallBase &Call, SelectionDAGBuilder &Builder, const TargetLowering &TLI, const TargetMachine &TM, SelectionDAG &DAG, const BasicBlock *EHPadBB)
DetermineConstraints - Find the constraints to use for inline asm operands.
static bool constructOperandInfo(ConstraintDecisionInfo &Info, TargetLowering::AsmOperandInfoVector &TargetConstraints, SelectionDAGBuilder &Builder, const TargetLowering &TLI, ExtraFlags &ExtraInfo)
Construct operand info objects.
static SDValue GetSignificand(SelectionDAG &DAG, SDValue Op, const SDLoc &dl)
GetSignificand - Get the significand and build it into a floating-point number with exponent of 1:
static SDValue expandExp(const SDLoc &dl, SDValue Op, SelectionDAG &DAG, const TargetLowering &TLI, SDNodeFlags Flags)
expandExp - Lower an exp intrinsic.
static const MDNode * getRangeMetadata(const Instruction &I)
static cl::opt< unsigned, true > LimitFPPrecision("limit-float-precision", cl::desc("Generate low-precision inline sequences " "for some float libcalls"), cl::location(LimitFloatPrecision), cl::Hidden, cl::init(0))
static void tryToElideArgumentCopy(FunctionLoweringInfo &FuncInfo, SmallVectorImpl< SDValue > &Chains, DenseMap< int, int > &ArgCopyElisionFrameIndexMap, SmallPtrSetImpl< const Instruction * > &ElidedArgCopyInstrs, ArgCopyElisionMapTy &ArgCopyElisionCandidates, const Argument &Arg, ArrayRef< SDValue > ArgVals, bool &ArgHasUses)
Try to elide argument copies from memory into a local alloca.
static unsigned LimitFloatPrecision
LimitFloatPrecision - Generate low-precision inline sequences for some float libcalls (6,...
static SDValue getCopyFromPartsVector(SelectionDAG &DAG, const SDLoc &DL, const SDValue *Parts, unsigned NumParts, MVT PartVT, EVT ValueVT, const Value *V, SDValue InChain, std::optional< CallingConv::ID > CC)
getCopyFromPartsVector - Create a value that contains the specified legal parts combined into the val...
static bool InBlock(const Value *V, const BasicBlock *BB)
static FPClassTest getNoFPClass(const Instruction &I)
static LLVM_ATTRIBUTE_ALWAYS_INLINE MVT::SimpleValueType getSimpleVT(const uint8_t *MatcherTable, size_t &MatcherIndex)
getSimpleVT - Decode a value in MatcherTable, if it's a VBR encoded value, use GetVBR to decode it.
This file defines the SmallPtrSet class.
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static SymbolRef::Type getType(const Symbol *Sym)
uint16_t RegSizeInBits(const MCRegisterInfo &MRI, MCRegister RegNo)
static const fltSemantics & IEEEsingle()
static LLVM_ABI Semantics SemanticsToEnum(const llvm::fltSemantics &Sem)
static LLVM_ABI const fltSemantics * getArbitraryFPSemantics(StringRef Format)
Returns the fltSemantics for a given arbitrary FP format string, or nullptr if invalid.
Class for arbitrary precision integers.
bool isNonNegative() const
Determine if this APInt Value is non-negative (>= 0)
bool isPowerOf2() const
Check if this APInt's value is a power of two greater than zero.
an instruction to allocate memory on the stack
Align getAlign() const
Return the alignment of the memory that is being allocated by the instruction.
LLVM_ABI std::optional< TypeSize > getAllocationSize(const DataLayout &DL) const
Get allocation size in bytes.
This class represents an incoming formal argument to a Function.
LLVM_ABI bool hasAttribute(Attribute::AttrKind Kind) const
Check if an argument has a given attribute.
unsigned getArgNo() const
Return the index of this formal argument in its containing function.
Represent a constant reference to an array (0 or more elements consecutively in memory),...
size_t size() const
Get the array size.
bool empty() const
Check if the array is empty.
A cache of @llvm.assume calls within a function.
An instruction that atomically checks whether a specified value is in a memory location,...
an instruction that atomically reads a memory location, combines it with another value,...
@ USubCond
Subtract only if no unsigned overflow.
@ FMinimum
*p = minimum(old, v) minimum matches the behavior of llvm.minimum.
@ Min
*p = old <signed v ? old : v
@ USubSat
*p = usub.sat(old, v) usub.sat matches the behavior of llvm.usub.sat.
@ FMaximum
*p = maximum(old, v) maximum matches the behavior of llvm.maximum.
@ UIncWrap
Increment one up to a maximum value.
@ Max
*p = old >signed v ? old : v
@ UMin
*p = old <unsigned v ? old : v
@ FMin
*p = minnum(old, v) minnum matches the behavior of llvm.minnum.
@ UMax
*p = old >unsigned v ? old : v
@ FMaximumNum
*p = maximumnum(old, v) maximumnum matches the behavior of llvm.maximumnum.
@ FMax
*p = maxnum(old, v) maxnum matches the behavior of llvm.maxnum.
@ UDecWrap
Decrement one until a minimum value or zero.
@ FMinimumNum
*p = minimumnum(old, v) minimumnum matches the behavior of llvm.minimumnum.
This class holds the attributes for a particular argument, parameter, function, or return value.
LLVM Basic Block Representation.
const Function * getParent() const
Return the enclosing method, or null if none.
LLVM_ABI InstListType::const_iterator getFirstNonPHIIt() const
Returns an iterator to the first instruction in this block that is not a PHINode instruction.
InstListType::const_iterator const_iterator
LLVM_ABI bool isEntryBlock() const
Return true if this is the entry block of the containing function.
LLVM_ABI InstListType::const_iterator getFirstNonPHIOrDbg(bool SkipPseudoOp=true) const
Returns a pointer to the first instruction in this block that is not a PHINode or a debug intrinsic,...
const Instruction * getTerminator() const LLVM_READONLY
Returns the terminator instruction; assumes that the block is well-formed.
This class is a wrapper over an AAResults, and it is intended to be used only when there are no IR ch...
This class represents a no-op cast from one type to another.
The address of a basic block.
Analysis providing branch probability information.
LLVM_ABI BranchProbability getEdgeProbability(const BasicBlock *Src, unsigned IndexInSuccessors) const
Get an edge's probability, relative to other out-edges of the Src.
LLVM_ABI bool isEdgeHot(const BasicBlock *Src, const BasicBlock *Dst) const
Test if an edge is hot relative to other out-edges of the Src.
static uint32_t getDenominator()
static BranchProbability getOne()
static BranchProbability getUnknown()
uint32_t getNumerator() const
LLVM_ABI uint64_t scale(uint64_t Num) const
Scale a large integer.
BranchProbability getCompl() const
static BranchProbability getZero()
static void normalizeProbabilities(ProbabilityIter Begin, ProbabilityIter End)
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
std::optional< OperandBundleUse > getOperandBundle(StringRef Name) const
Return an operand bundle by name, if present.
CallingConv::ID getCallingConv() const
User::op_iterator arg_begin()
Return the iterator pointing to the beginning of the argument list.
LLVM_ABI bool isMustTailCall() const
Tests if this call site must be tail call optimized.
LLVM_ABI bool isIndirectCall() const
Return true if the callsite is an indirect call.
unsigned countOperandBundlesOfType(StringRef Name) const
Return the number of operand bundles with the tag Name attached to this instruction.
Value * getCalledOperand() const
Value * getArgOperand(unsigned i) const
User::op_iterator arg_end()
Return the iterator pointing to the end of the argument list.
bool isConvergent() const
Determine if the invoke is convergent.
FunctionType * getFunctionType() const
unsigned arg_size() const
AttributeList getAttributes() const
Return the attributes for this call.
LLVM_ABI bool isTailCall() const
Tests if this call site is marked as a tail call.
CallBr instruction, tracking function calls that may not return control but instead transfer it to a ...
This class represents a function call, abstracting a target machine's calling convention.
This class is the base class for the comparison instructions.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Conditional Branch instruction.
Class for constant bytes.
ConstantDataSequential - A vector or array constant whose element type is a simple 1/2/4/8-byte integ...
A constant value that is initialized with an expression using other constant values.
ConstantFP - Floating Point Values [float, double].
This is the shared class of boolean and integer constants.
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
static LLVM_ABI ConstantInt * getFalse(LLVMContext &Context)
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
const APInt & getValue() const
Return the constant as an APInt value reference.
A signed pointer, in the ptrauth sense.
uint64_t getZExtValue() const
Constant Vector Declarations.
This is an important base class in LLVM.
This is the common base class for constrained floating point intrinsics.
LLVM_ABI std::optional< fp::ExceptionBehavior > getExceptionBehavior() const
LLVM_ABI unsigned getNonMetadataArgCount() const
LLVM_ABI bool isEntryValue() const
Check if the expression consists of exactly one entry value operand.
static bool fragmentsOverlap(const FragmentInfo &A, const FragmentInfo &B)
Check if fragments overlap between a pair of FragmentInfos.
static LLVM_ABI DIExpression * appendOpsToArg(const DIExpression *Expr, ArrayRef< uint64_t > Ops, unsigned ArgNo, bool StackValue=false)
Create a copy of Expr by appending the given list of Ops to each instance of the operand DW_OP_LLVM_a...
static LLVM_ABI std::optional< FragmentInfo > getFragmentInfo(expr_op_iterator Start, expr_op_iterator End)
Retrieve the details of this fragment expression.
LLVM_ABI uint64_t getNumLocationOperands() const
Return the number of unique location operands referred to (via DW_OP_LLVM_arg) in this expression; th...
static LLVM_ABI std::optional< DIExpression * > createFragmentExpression(const DIExpression *Expr, unsigned OffsetInBits, unsigned SizeInBits)
Create a DIExpression to describe one part of an aggregate variable that is fragmented across multipl...
static LLVM_ABI const DIExpression * convertToUndefExpression(const DIExpression *Expr)
Removes all elements from Expr that do not apply to an undef debug value, which includes every operat...
static LLVM_ABI DIExpression * prepend(const DIExpression *Expr, uint8_t Flags, int64_t Offset=0)
Prepend DIExpr with a deref and offset operation and optionally turn it into a stack value or/and an ...
static LLVM_ABI DIExpression * prependOpcodes(const DIExpression *Expr, SmallVectorImpl< uint64_t > &Ops, bool StackValue=false, bool EntryValue=false)
Prepend DIExpr with the given opcodes and optionally turn it into a stack value.
Base class for variables.
LLVM_ABI std::optional< uint64_t > getSizeInBits() const
Determines the size of the variable's type.
A parsed version of the target data layout string in and methods for querying it.
Records a position in IR for a source label (DILabel).
Base class for non-instruction debug metadata records that have positions within IR.
DebugLoc getDebugLoc() const
Record of a variable value-assignment, aka a non instruction representation of the dbg....
LocationType getType() const
LLVM_ABI Value * getVariableLocationOp(unsigned OpIdx) const
DIExpression * getExpression() const
DILocalVariable * getVariable() const
LLVM_ABI iterator_range< location_op_iterator > location_ops() const
Get the locations corresponding to the variable referenced by the debug info intrinsic.
LLVM_ABI DILocation * getInlinedAt() const
iterator find(const_arg_type_t< KeyT > Val)
DenseMapIterator< KeyT, ValueT, KeyInfoT, BucketT > iterator
DenseMapIterator< KeyT, ValueT, KeyInfoT, BucketT, true > const_iterator
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
void reserve(size_type NumEntries)
Grow the densemap so that it can contain at least NumEntries items before resizing again.
Diagnostic information for inline asm reporting.
static constexpr ElementCount getFixed(ScalarTy MinVal)
static constexpr ElementCount get(ScalarTy MinVal, bool Scalable)
constexpr bool isScalar() const
Exactly one element.
Lightweight error class with error context and mandatory checking.
Class representing an expression and its matching format.
This instruction compares its operands according to the predicate given to the constructor.
This is a fast-path instruction selection class that generates poor code and doesn't support illegal ...
bool allowReassoc() const
Flag queries.
An instruction for ordering other memory operations.
static LLVM_ABI FixedVectorType * get(Type *ElementType, unsigned NumElts)
This class represents a freeze function that returns random concrete value if an operand is either a ...
FunctionLoweringInfo - This contains information that is global to a function that is used when lower...
BranchProbabilityInfo * BPI
MachineBasicBlock * getMBB(const BasicBlock *BB) const
DenseMap< const AllocaInst *, int > StaticAllocaMap
StaticAllocaMap - Keep track of frame indices for fixed sized allocas in the entry block.
const LiveOutInfo * GetLiveOutRegInfo(Register Reg)
GetLiveOutRegInfo - Gets LiveOutInfo for a register, returning NULL if the register is a PHI destinat...
MachineBasicBlock * MBB
MBB - The current block.
Class to represent function types.
unsigned getNumParams() const
Return the number of fixed parameters this function type requires.
Type * getParamType(unsigned i) const
Parameter type accessors.
Type * getReturnType() const
Data structure describing the variable locations in a function.
const BasicBlock & getEntryBlock() const
FunctionType * getFunctionType() const
Returns the FunctionType for me.
Intrinsic::ID getIntrinsicID() const LLVM_READONLY
getIntrinsicID - This method returns the ID number of the specified function, or Intrinsic::not_intri...
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
bool hasParamAttribute(unsigned ArgNo, Attribute::AttrKind Kind) const
check if an attributes is in the list of attributes.
CallingConv::ID getCallingConv() const
getCallingConv()/setCallingConv(CC) - These method get and set the calling convention of this functio...
Constant * getPersonalityFn() const
Get the personality function associated with this function.
AttributeList getAttributes() const
Return the attribute list for this Function.
bool isIntrinsic() const
isIntrinsic - Returns true if the function's name starts with "llvm.".
bool hasFnAttribute(Attribute::AttrKind Kind) const
Return true if the function has the attribute.
Garbage collection metadata for a single function.
bool hasNoUnsignedSignedWrap() const
bool hasNoUnsignedWrap() const
an instruction for type-safe pointer arithmetic to access elements of arrays and structs
static StringRef dropLLVMManglingEscape(StringRef Name)
If the given string begins with the GlobalValue name mangling escape character '\1',...
bool hasDLLImportStorageClass() const
Module * getParent()
Get the module that this global value is contained inside of...
This instruction compares its operands according to the predicate given to the constructor.
Indirect Branch Instruction.
void setMemConstraint(ConstraintCode C)
setMemConstraint - Augment an existing flag with the constraint code for a memory constraint.
This instruction inserts a struct field of array element value into an aggregate value.
MDNode * getMetadata(unsigned KindID) const
Get the metadata of given kind attached to this Instruction.
LLVM_ABI FastMathFlags getFastMathFlags() const LLVM_READONLY
Convenience function for getting all the fast-math flags, which must be an operator which supports th...
LLVM_ABI AAMDNodes getAAMetadata() const
Returns the AA metadata for this instruction.
@ MIN_INT_BITS
Minimum number of bits that can be specified.
Intrinsic::ID getIntrinsicID() const
Return the intrinsic ID of this intrinsic.
This is an important class for using LLVM in a threaded context.
@ OB_clang_arc_attachedcall
LLVM_ABI void diagnose(const DiagnosticInfo &DI)
Report a message to the currently installed diagnostic handler.
The landingpad instruction holds all of the information necessary to generate correct exception handl...
A helper class to return the specified delimiter string after the first invocation of operator String...
An instruction for reading from memory.
static LocationSize precise(uint64_t Value)
static constexpr LocationSize beforeOrAfterPointer()
Any location before or after the base pointer (but still within the underlying object).
static LocationSize upperBound(uint64_t Value)
LLVM_ABI MCSymbol * createTempSymbol()
Create a temporary symbol with a unique name.
LLVM_ABI MCSymbol * getOrCreateFrameAllocSymbol(const Twine &FuncName, unsigned Idx)
Gets a symbol that will be defined to the final stack offset of a local variable after codegen.
MCSymbol - Instances of this class represent a symbol name in the MC file, and MCSymbols are created ...
@ INVALID_SIMPLE_VALUE_TYPE
uint64_t getScalarSizeInBits() const
unsigned getVectorNumElements() const
bool isVector() const
Return true if this is a vector value type.
bool isInteger() const
Return true if this is an integer or a vector integer type.
TypeSize getSizeInBits() const
Returns the size of the specified MVT in bits.
uint64_t getFixedSizeInBits() const
Return the size of the specified fixed width value type in bits.
ElementCount getVectorElementCount() const
TypeSize getStoreSize() const
Return the number of bytes overwritten by a store of the specified value type.
bool bitsGE(MVT VT) const
Return true if this has no less bits than VT.
bool isScalarInteger() const
Return true if this is an integer, not including vectors.
static MVT getVectorVT(MVT VT, unsigned NumElements)
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
static MVT getIntegerVT(unsigned BitWidth)
void normalizeSuccProbs()
Normalize probabilities of all successors so that the sum of them becomes one.
const BasicBlock * getBasicBlock() const
Return the LLVM basic block that this instance corresponded to originally.
LLVM_ABI void setSuccProbability(succ_iterator I, BranchProbability Prob)
Set successor probability of a given iterator.
succ_iterator succ_begin()
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
SmallVectorImpl< MachineBasicBlock * >::iterator succ_iterator
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
void setIsEHContTarget(bool V=true)
Indicates if this is a target of Windows EH Continuation Guard.
void setIsEHFuncletEntry(bool V=true)
Indicates if this is the entry block of an EH funclet.
MachineInstrBundleIterator< MachineInstr > iterator
void setIsEHScopeEntry(bool V=true)
Indicates if this is the entry block of an EH scope, i.e., the block that that used to have a catchpa...
void setMachineBlockAddressTaken()
Set this block to indicate that its address is used as something other than the target of a terminato...
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
void setIsImmutableObjectIndex(int ObjectIdx, bool IsImmutable)
Marks the immutability of an object.
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
bool hasOpaqueSPAdjustment() const
Returns true if the function contains opaque dynamic stack adjustments.
int getStackProtectorIndex() const
Return the index for the stack protector object.
void setStackProtectorIndex(int I)
void setIsAliasedObjectIndex(int ObjectIdx, bool IsAliased)
Set "maybe pointed to by an LLVM IR value" for an object.
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
void RemoveStackObject(int ObjectIdx)
Remove or mark dead a statically sized stack object.
void setFunctionContextIndex(int I)
const WinEHFuncInfo * getWinEHFuncInfo() const
getWinEHFuncInfo - Return information about how the current function uses Windows exception handling.
bool useDebugInstrRef() const
Returns true if the function's variable locations are tracked with instruction referencing.
void setCallSiteBeginLabel(MCSymbol *BeginLabel, unsigned Site)
Map the begin label for a call site.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MCContext & getContext() const
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
void addCodeViewAnnotation(MCSymbol *Label, MDNode *MD)
Record annotations associated with a particular label.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
bool hasEHFunclets() const
void setHasEHContTarget(bool V)
void addInvoke(MachineBasicBlock *LandingPad, MCSymbol *BeginLabel, MCSymbol *EndLabel)
Provide the begin and end labels of an invoke style call and associate it with a try landing pad bloc...
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
Representation of each machine instruction.
A description of a memory reference used in the backend.
Flags
Flags values. These may be or'd together.
@ MOVolatile
The memory access is volatile.
@ MODereferenceable
The memory access is dereferenceable (i.e., doesn't trap).
@ MOLoad
The memory access reads data.
@ MONonTemporal
The memory access is non-temporal.
@ MOInvariant
The memory access always returns the same value (or traps).
@ MOStore
The memory access writes data.
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
static MachineOperand CreateFI(int Idx)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
const TargetRegisterClass * getRegClass(Register Reg) const
Return the register class of the specified virtual register.
def_iterator def_begin(Register RegNo) const
LLVM_ABI Register createVirtualRegister(const TargetRegisterClass *RegClass, StringRef Name="")
createVirtualRegister - Create and return a new virtual register in the function with the specified r...
LLVM_ABI MCRegister getLiveInPhysReg(Register VReg) const
getLiveInPhysReg - If VReg is a live-in virtual register, return the corresponding live-in physical r...
An SDNode that represents everything that will be needed to construct a MachineInstr.
bool contains(const KeyT &Key) const
std::pair< iterator, bool > try_emplace(const KeyT &Key, Ts &&...Args)
static MemoryLocation getAfter(const Value *Ptr, const AAMDNodes &AATags=AAMDNodes())
Return a location that may access any location after Ptr, while remaining within the underlying objec...
A Module instance is used to store all the information related to an LLVM module.
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
static LLVM_ABI PoisonValue * get(Type *T)
Static factory methods - Return an 'poison' object of the specified type.
Wrapper class representing virtual and physical registers.
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Resume the propagation of an exception.
Return a value (possibly void), from a function.
Holds the information from a dbg_label node through SDISel.
static SDDbgOperand fromNode(SDNode *Node, unsigned ResNo)
static SDDbgOperand fromFrameIdx(unsigned FrameIdx)
static SDDbgOperand fromVReg(Register VReg)
static SDDbgOperand fromConst(const Value *Const)
Holds the information from a dbg_value node through SDISel.
Wrapper class for IR location info (IR ordering and DebugLoc) to be passed into SDNode creation funct...
Represents one node in the SelectionDAG.
unsigned getOpcode() const
Return the SelectionDAG opcode value for this node.
iterator_range< value_op_iterator > op_values() const
unsigned getIROrder() const
Return the node ordering.
uint64_t getAsZExtVal() const
Helper method returns the zero-extended integer value of a ConstantSDNode.
unsigned getNumValues() const
Return the number of values defined/returned by this operator.
const SDValue & getOperand(unsigned Num) const
EVT getValueType(unsigned ResNo) const
Return the type of a specified result.
Unlike LLVM values, Selection DAG nodes may return multiple values as the result of a computation.
SDNode * getNode() const
get the SDNode which holds the desired result
SDValue getValue(unsigned R) const
EVT getValueType() const
Return the ValueType of the referenced return value.
TypeSize getValueSizeInBits() const
Returns the size of the value in bits.
const SDValue & getOperand(unsigned i) const
unsigned getResNo() const
get the index which selects a specific result in the SDNode
MVT getSimpleValueType() const
Return the simple ValueType of the referenced return value.
unsigned getOpcode() const
SelectionDAGBuilder - This is the common target-independent lowering implementation that is parameter...
SDValue getValue(const Value *V)
getValue - Return an SDValue for the given Value.
bool shouldKeepJumpConditionsTogether(const FunctionLoweringInfo &FuncInfo, const CondBrInst &I, Instruction::BinaryOps Opc, const Value *Lhs, const Value *Rhs, TargetLoweringBase::CondMergingParams Params) const
DenseMap< const Constant *, Register > ConstantsOut
void addDanglingDebugInfo(SmallVectorImpl< Value * > &Values, DILocalVariable *Var, DIExpression *Expr, bool IsVariadic, DebugLoc DL, unsigned Order)
Register a dbg_value which relies on a Value which we have not yet seen.
void visitDbgInfo(const Instruction &I)
void clearDanglingDebugInfo()
Clear the dangling debug information map.
SDValue lowerStartEH(SDValue Chain, const BasicBlock *EHPadBB, MCSymbol *&BeginLabel)
void LowerCallTo(const CallBase &CB, SDValue Callee, bool IsTailCall, bool IsMustTailCall, const BasicBlock *EHPadBB=nullptr, const TargetLowering::PtrAuthInfo *PAI=nullptr)
void clear()
Clear out the current SelectionDAG and the associated state and prepare this SelectionDAGBuilder obje...
void visitBitTestHeader(SwitchCG::BitTestBlock &B, MachineBasicBlock *SwitchBB)
visitBitTestHeader - This function emits necessary code to produce value suitable for "bit tests"
void LowerStatepoint(const GCStatepointInst &I, const BasicBlock *EHPadBB=nullptr)
std::unique_ptr< SDAGSwitchLowering > SL
SDValue lowerRangeToAssertZExt(SelectionDAG &DAG, const Instruction &I, SDValue Op)
bool HasTailCall
This is set to true if a call in the current block has been translated as a tail call.
bool ShouldEmitAsBranches(const std::vector< SwitchCG::CaseBlock > &Cases)
If the set of cases should be emitted as a series of branches, return true.
void EmitBranchForMergedCondition(const Value *Cond, MachineBasicBlock *TBB, MachineBasicBlock *FBB, MachineBasicBlock *CurBB, MachineBasicBlock *SwitchBB, BranchProbability TProb, BranchProbability FProb, bool InvertCond)
EmitBranchForMergedCondition - Helper method for FindMergedConditions.
void LowerDeoptimizeCall(const CallInst *CI)
void LowerCallSiteWithDeoptBundle(const CallBase *Call, SDValue Callee, const BasicBlock *EHPadBB)
SwiftErrorValueTracking & SwiftError
Information about the swifterror values used throughout the function.
SDValue getNonRegisterValue(const Value *V)
getNonRegisterValue - Return an SDValue for the given Value, but don't look in FuncInfo....
const TargetTransformInfo * TTI
DenseMap< MachineBasicBlock *, SmallVector< unsigned, 4 > > LPadToCallSiteMap
Map a landing pad to the call site indexes.
SDValue lowerNoFPClassToAssertNoFPClass(SelectionDAG &DAG, const Instruction &I, SDValue Op)
void handleDebugDeclare(Value *Address, DILocalVariable *Variable, DIExpression *Expression, DebugLoc DL)
StatepointLoweringState StatepointLowering
State used while lowering a statepoint sequence (gc_statepoint, gc_relocate, and gc_result).
void visitBitTestCase(SwitchCG::BitTestBlock &BB, MachineBasicBlock *NextMBB, BranchProbability BranchProbToNext, Register Reg, SwitchCG::BitTestCase &B, MachineBasicBlock *SwitchBB)
visitBitTestCase - this function produces one "bit test"
bool canTailCall(const CallBase &CB) const
void populateCallLoweringInfo(TargetLowering::CallLoweringInfo &CLI, const CallBase *Call, unsigned ArgIdx, unsigned NumArgs, SDValue Callee, Type *ReturnTy, AttributeSet RetAttrs, bool IsPatchPoint)
Populate a CallLowerinInfo (into CLI) based on the properties of the call being lowered.
void CopyValueToVirtualRegister(const Value *V, Register Reg, ISD::NodeType ExtendType=ISD::ANY_EXTEND)
void salvageUnresolvedDbgValue(const Value *V, DanglingDebugInfo &DDI)
For the given dangling debuginfo record, perform last-ditch efforts to resolve the debuginfo to somet...
SmallVector< SDValue, 8 > PendingLoads
Loads are not emitted to the program immediately.
GCFunctionInfo * GFI
Garbage collection metadata for the function.
void init(GCFunctionInfo *gfi, BatchAAResults *BatchAA, AssumptionCache *AC, const TargetLibraryInfo *li, const TargetTransformInfo &TTI)
SDValue getRoot()
Similar to getMemoryRoot, but also flushes PendingConstrainedFP(Strict) items.
void ExportFromCurrentBlock(const Value *V)
ExportFromCurrentBlock - If this condition isn't known to be exported from the current basic block,...
DebugLoc getCurDebugLoc() const
void resolveOrClearDbgInfo()
Evict any dangling debug information, attempting to salvage it first.
std::pair< SDValue, SDValue > lowerInvokable(TargetLowering::CallLoweringInfo &CLI, const BasicBlock *EHPadBB=nullptr)
SDValue getMemoryRoot()
Return the current virtual root of the Selection DAG, flushing any PendingLoad items.
void resolveDanglingDebugInfo(const Value *V, SDValue Val)
If we saw an earlier dbg_value referring to V, generate the debug data structures now that we've seen...
SDLoc getCurSDLoc() const
void visit(const Instruction &I)
void dropDanglingDebugInfo(const DILocalVariable *Variable, const DIExpression *Expr)
If we have dangling debug info that describes Variable, or an overlapping part of variable considerin...
SDValue getCopyFromRegs(const Value *V, Type *Ty)
If there was virtual register allocated for the value V emit CopyFromReg of the specified type Ty.
void CopyToExportRegsIfNeeded(const Value *V)
CopyToExportRegsIfNeeded - If the given value has virtual registers created for it,...
void handleKillDebugValue(DILocalVariable *Var, DIExpression *Expr, DebugLoc DbgLoc, unsigned Order)
Create a record for a kill location debug intrinsic.
void visitJumpTable(SwitchCG::JumpTable &JT)
visitJumpTable - Emit JumpTable node in the current MBB
SDValue getFPOperationRoot(fp::ExceptionBehavior EB)
Return the current virtual root of the Selection DAG, flushing PendingConstrainedFP or PendingConstra...
void visitJumpTableHeader(SwitchCG::JumpTable &JT, SwitchCG::JumpTableHeader &JTH, MachineBasicBlock *SwitchBB)
visitJumpTableHeader - This function emits necessary code to produce index in the JumpTable from swit...
void LowerCallSiteWithPtrAuthBundle(const CallBase &CB, const BasicBlock *EHPadBB)
static const unsigned LowestSDNodeOrder
Lowest valid SDNodeOrder.
void LowerDeoptimizingReturn()
FunctionLoweringInfo & FuncInfo
Information about the function as a whole.
void setValue(const Value *V, SDValue NewN)
void FindMergedConditions(const Value *Cond, MachineBasicBlock *TBB, MachineBasicBlock *FBB, MachineBasicBlock *CurBB, MachineBasicBlock *SwitchBB, Instruction::BinaryOps Opc, BranchProbability TProb, BranchProbability FProb, bool InvertCond)
const TargetLibraryInfo * LibInfo
bool isExportableFromCurrentBlock(const Value *V, const BasicBlock *FromBB)
void visitSPDescriptorParent(StackProtectorDescriptor &SPD, MachineBasicBlock *ParentBB)
Codegen a new tail for a stack protector check ParentMBB which has had its tail spliced into a stack ...
bool handleDebugValue(ArrayRef< const Value * > Values, DILocalVariable *Var, DIExpression *Expr, DebugLoc DbgLoc, unsigned Order, bool IsVariadic)
For a given list of Values, attempt to create and record a SDDbgValue in the SelectionDAG.
SDValue getControlRoot()
Similar to getRoot, but instead of flushing all the PendingLoad items, flush all the PendingExports (...
void UpdateSplitBlock(MachineBasicBlock *First, MachineBasicBlock *Last)
When an MBB was split during scheduling, update the references that need to refer to the last resulti...
SDValue getValueImpl(const Value *V)
getValueImpl - Helper function for getValue and getNonRegisterValue.
void visitSwitchCase(SwitchCG::CaseBlock &CB, MachineBasicBlock *SwitchBB)
visitSwitchCase - Emits the necessary code to represent a single node in the binary search tree resul...
void visitSPDescriptorFailure(StackProtectorDescriptor &SPD)
Codegen the failure basic block for a stack protector check.
std::unique_ptr< FunctionLoweringInfo > FuncInfo
SmallPtrSet< const Instruction *, 4 > ElidedArgCopyInstrs
const TargetLowering * TLI
MachineRegisterInfo * RegInfo
std::unique_ptr< SwiftErrorValueTracking > SwiftError
virtual void emitFunctionEntryCode()
std::unique_ptr< SelectionDAGBuilder > SDB
virtual std::pair< SDValue, SDValue > EmitTargetCodeForMemccpy(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Dst, SDValue Src, SDValue C, SDValue Size, const CallInst *CI) const
Emit target-specific code that performs a memccpy, in cases where that is faster than a libcall.
virtual std::pair< SDValue, SDValue > EmitTargetCodeForStrnlen(SelectionDAG &DAG, const SDLoc &DL, SDValue Chain, SDValue Src, SDValue MaxLength, MachinePointerInfo SrcPtrInfo) const
virtual std::pair< SDValue, SDValue > EmitTargetCodeForStrlen(SelectionDAG &DAG, const SDLoc &DL, SDValue Chain, SDValue Src, const CallInst *CI) const
virtual std::pair< SDValue, SDValue > EmitTargetCodeForStrstr(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Op1, SDValue Op2, const CallInst *CI) const
Emit target-specific code that performs a strstr, in cases where that is faster than a libcall.
virtual std::pair< SDValue, SDValue > EmitTargetCodeForMemchr(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Src, SDValue Char, SDValue Length, MachinePointerInfo SrcPtrInfo) const
Emit target-specific code that performs a memchr, in cases where that is faster than a libcall.
virtual std::pair< SDValue, SDValue > EmitTargetCodeForStrcmp(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Op1, SDValue Op2, MachinePointerInfo Op1PtrInfo, MachinePointerInfo Op2PtrInfo, const CallInst *CI) const
Emit target-specific code that performs a strcmp, in cases where that is faster than a libcall.
virtual std::pair< SDValue, SDValue > EmitTargetCodeForMemcmp(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Op1, SDValue Op2, SDValue Op3, const CallInst *CI) const
Emit target-specific code that performs a memcmp/bcmp, in cases where that is faster than a libcall.
virtual SDValue EmitTargetCodeForSetTag(SelectionDAG &DAG, const SDLoc &dl, SDValue Chain, SDValue Addr, SDValue Size, MachinePointerInfo DstPtrInfo, bool ZeroData) const
virtual std::pair< SDValue, SDValue > EmitTargetCodeForStrcpy(SelectionDAG &DAG, const SDLoc &DL, SDValue Chain, SDValue Dest, SDValue Src, MachinePointerInfo DestPtrInfo, MachinePointerInfo SrcPtrInfo, bool isStpcpy, const CallInst *CI) const
Emit target-specific code that performs a strcpy or stpcpy, in cases where that is faster than a libc...
This is used to represent a portion of an LLVM function in a low-level Data Dependence DAG representa...
SDValue getTargetGlobalAddress(const GlobalValue *GV, const SDLoc &DL, EVT VT, int64_t offset=0, unsigned TargetFlags=0)
SDValue getExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT, unsigned Opcode)
Convert Op, which must be of integer type, to the integer type VT, by either any/sign/zero-extending ...
const SDValue & getRoot() const
Return the root tag of the SelectionDAG.
const TargetSubtargetInfo & getSubtarget() const
SDValue getCopyToReg(SDValue Chain, const SDLoc &dl, Register Reg, SDValue N)
LLVM_ABI SDValue getMergeValues(ArrayRef< SDValue > Ops, const SDLoc &dl)
Create a MERGE_VALUES node from the given operands.
LLVM_ABI SDVTList getVTList(EVT VT)
Return an SDVTList that represents the list of values specified.
LLVM_ABI SDValue getShiftAmountConstant(uint64_t Val, EVT VT, const SDLoc &DL)
LLVM_ABI MachineSDNode * getMachineNode(unsigned Opcode, const SDLoc &dl, EVT VT)
These are used for target selectors to create a new node with specified return type(s),...
LLVM_ABI void ExtractVectorElements(SDValue Op, SmallVectorImpl< SDValue > &Args, unsigned Start=0, unsigned Count=0, EVT EltVT=EVT())
Append the extracted elements from Start to Count out of the vector Op in Args.
LLVM_ABI SDValue getConstantPool(const Constant *C, EVT VT, MaybeAlign Align=std::nullopt, int Offs=0, bool isT=false, unsigned TargetFlags=0)
LLVM_ABI SDValue getConstantFP(double Val, const SDLoc &DL, EVT VT, bool isTarget=false)
Create a ConstantFPSDNode wrapping a constant value.
LLVM_ABI SDValue getRegister(Register Reg, EVT VT)
LLVM_ABI SDValue getLoad(EVT VT, const SDLoc &dl, SDValue Chain, SDValue Ptr, MachinePointerInfo PtrInfo, MaybeAlign Alignment=MaybeAlign(), MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr)
Loads are not normal binary operators: their result type is not determined by their operands,...
LLVM_ABI Align getEVTAlign(EVT MemoryVT) const
Compute the default alignment value for the given type.
LLVM_ABI bool shouldOptForSize() const
const TargetLowering & getTargetLoweringInfo() const
static constexpr unsigned MaxRecursionDepth
LLVM_ABI void AddDbgValue(SDDbgValue *DB, bool isParameter)
Add a dbg_value SDNode.
SDValue getUNDEF(EVT VT)
Return an UNDEF node. UNDEF does not have a useful SDLoc.
SDValue getBuildVector(EVT VT, const SDLoc &DL, ArrayRef< SDValue > Ops)
Return an ISD::BUILD_VECTOR node.
LLVM_ABI SDValue getBitcast(EVT VT, SDValue V)
Return a bitcast using the SDLoc of the value operand, and casting to the provided type.
LLVM_ABI SDDbgValue * getDbgValueList(DIVariable *Var, DIExpression *Expr, ArrayRef< SDDbgOperand > Locs, ArrayRef< SDNode * > Dependencies, bool IsIndirect, const DebugLoc &DL, unsigned O, bool IsVariadic)
Creates a SDDbgValue node from a list of locations.
SDValue getCopyFromReg(SDValue Chain, const SDLoc &dl, Register Reg, EVT VT)
LLVM_ABI void setNodeMemRefs(MachineSDNode *N, ArrayRef< MachineMemOperand * > NewMemRefs)
Mutate the specified machine node's memory references to the provided list.
const DataLayout & getDataLayout() const
SDValue getTargetFrameIndex(int FI, EVT VT)
LLVM_ABI SDValue getConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isTarget=false, bool isOpaque=false)
Create a ConstantSDNode wrapping a constant value.
LLVM_ABI SDValue getMemBasePlusOffset(SDValue Base, TypeSize Offset, const SDLoc &DL, const SDNodeFlags Flags=SDNodeFlags())
Returns sum of the base pointer and offset.
LLVM_ABI SDValue getTruncStore(SDValue Chain, const SDLoc &dl, SDValue Val, SDValue Ptr, MachinePointerInfo PtrInfo, EVT SVT, Align Alignment, MachineMemOperand::Flags MMOFlags=MachineMemOperand::MONone, const AAMDNodes &AAInfo=AAMDNodes())
LLVM_ABI SDValue getMDNode(const MDNode *MD)
Return an MDNodeSDNode which holds an MDNode.
LLVM_ABI SDValue getBasicBlock(MachineBasicBlock *MBB)
LLVM_ABI SDValue getEHLabel(const SDLoc &dl, SDValue Root, MCSymbol *Label)
LLVM_ABI SDValue getPtrExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either truncating it or perform...
LLVM_ABI SDValue getAnyExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either any-extending or truncat...
const LibcallLoweringInfo & getLibcalls() const
LLVM_ABI SDValue getIntPtrConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
LLVM_ABI SDValue getValueType(EVT)
LLVM_ABI SDValue getNode(unsigned Opcode, const SDLoc &DL, EVT VT, ArrayRef< SDUse > Ops)
Gets or creates the specified node.
LLVM_ABI SDValue getFPExtendOrRound(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of float type, to the float type VT, by either extending or rounding (by tr...
SDValue getTargetConstant(uint64_t Val, const SDLoc &DL, EVT VT, bool isOpaque=false)
LLVM_ABI SDValue getVectorIdxConstant(uint64_t Val, const SDLoc &DL, bool isTarget=false)
MachineFunction & getMachineFunction() const
LLVM_ABI SDValue getFrameIndex(int FI, EVT VT, bool isTarget=false)
LLVM_ABI SDValue getZExtOrTrunc(SDValue Op, const SDLoc &DL, EVT VT)
Convert Op, which must be of integer type, to the integer type VT, by either zero-extending or trunca...
LLVMContext * getContext() const
const SDValue & setRoot(SDValue N)
Set the current root tag of the SelectionDAG.
LLVM_ABI SDValue getTargetExternalSymbol(const char *Sym, EVT VT, unsigned TargetFlags=0)
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
void swap(SmallVectorImpl &RHS)
void push_back(const T &Elt)
pointer data()
Return a pointer to the vector's buffer, even if empty().
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Encapsulates all of the information needed to generate a stack protector check, and signals to isel w...
MachineBasicBlock * getSuccessMBB()
MachineBasicBlock * getFailureMBB()
MachineBasicBlock * getParentMBB()
bool shouldEmitFunctionBasedCheckStackProtector() const
An instruction for storing to memory.
Represent a constant reference to a string, i.e.
constexpr bool empty() const
Check if the string is empty.
constexpr const char * data() const
Get a pointer to the start of the string (which may not be null terminated).
Information about stack frame layout on the target.
virtual TargetStackID::Value getStackIDForScalableVectors() const
Returns the StackID that scalable vectors should be associated with.
Provides information about what library functions are available for the current target.
virtual Align getByValTypeAlignment(Type *Ty, const DataLayout &DL) const
Returns the desired alignment for ByVal or InAlloca aggregate function arguments in the caller parame...
virtual bool isFMAFasterThanFMulAndFAdd(const MachineFunction &MF, EVT) const
Return true if an FMA operation is faster than a pair of fmul and fadd instructions.
EVT getMemValueType(const DataLayout &DL, Type *Ty, bool AllowUnknown=false) const
Function * getSSPStackGuardCheck(const Module &M, const LibcallLoweringInfo &Libcalls) const
If the target has a standard stack protection check function that performs validation and error handl...
EVT getValueType(const DataLayout &DL, Type *Ty, bool AllowUnknown=false) const
Return the EVT corresponding to this LLVM type.
LegalizeAction
This enum indicates whether operations are valid for a target, and if not, what action should be used...
virtual bool useStackGuardXorFP() const
If this function returns true, stack protection checks should XOR the frame pointer (or whichever poi...
virtual const TargetRegisterClass * getRegClassFor(MVT VT, bool isDivergent=false) const
Return the register class that should be used for the specified value type.
virtual bool isLegalScaleForGatherScatter(uint64_t Scale, uint64_t ElemSize) const
virtual bool isSExtCheaperThanZExt(EVT FromTy, EVT ToTy) const
Return true if sign-extension from FromTy to ToTy is cheaper than zero-extension.
MVT getVectorIdxTy(const DataLayout &DL) const
Returns the type to be used for the index operand of: ISD::INSERT_VECTOR_ELT, ISD::EXTRACT_VECTOR_ELT...
virtual unsigned getNumRegistersForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const
Certain targets require unusual breakdowns of certain types.
virtual bool isZExtFree(Type *FromTy, Type *ToTy) const
Return true if any actual instruction that defines a value of type FromTy implicitly zero-extends the...
virtual MVT getRegisterTypeForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT) const
Certain combinations of ABIs, Targets and features require that types are legal for some operations a...
virtual unsigned getNumRegisters(LLVMContext &Context, EVT VT, std::optional< MVT > RegisterVT=std::nullopt) const
Return the number of registers that this ValueType will eventually require.
MachineMemOperand::Flags getLoadMemOperandFlags(const LoadInst &LI, const DataLayout &DL, AssumptionCache *AC=nullptr, const TargetLibraryInfo *LibInfo=nullptr, CodeGenOptLevel OptLevel=CodeGenOptLevel::Default) const
virtual bool shouldExtendGSIndex(EVT VT, EVT &EltTy) const
Returns true if the index type for a masked gather/scatter requires extending.
virtual unsigned getVectorTypeBreakdownForCallingConv(LLVMContext &Context, CallingConv::ID CC, EVT VT, EVT &IntermediateVT, unsigned &NumIntermediates, MVT &RegisterVT) const
Certain targets such as MIPS require that some types such as vectors are always broken down into scal...
Register getStackPointerRegisterToSaveRestore() const
If a physical register, this specifies the register that llvm.savestack/llvm.restorestack should save...
LegalizeAction getFixedPointOperationAction(unsigned Op, EVT VT, unsigned Scale) const
Some fixed point operations may be natively supported by the target but only for specific scales.
MachineMemOperand::Flags getAtomicMemOperandFlags(const Instruction &AI, const DataLayout &DL) const
virtual bool allowsMisalignedMemoryAccesses(EVT, unsigned AddrSpace=0, Align Alignment=Align(1), MachineMemOperand::Flags Flags=MachineMemOperand::MONone, unsigned *=nullptr) const
Determine if the target supports unaligned memory accesses.
bool isOperationCustom(unsigned Op, EVT VT) const
Return true if the operation uses custom lowering, regardless of whether the type is legal or not.
bool hasBigEndianPartOrdering(EVT VT, const DataLayout &DL) const
When splitting a value of the specified type into parts, does the Lo or Hi part come first?
EVT getShiftAmountTy(EVT LHSTy, const DataLayout &DL) const
Returns the type for the shift amount of a shift opcode.
virtual Align getABIAlignmentForCallingConv(Type *ArgTy, const DataLayout &DL) const
Certain targets have context sensitive alignment requirements, where one type has the alignment requi...
MachineMemOperand::Flags getVPIntrinsicMemOperandFlags(const VPIntrinsic &VPIntrin) const
virtual bool shouldExpandGetActiveLaneMask(EVT VT, EVT OpVT) const
Return true if the @llvm.get.active.lane.mask intrinsic should be expanded using generic code in Sele...
virtual EVT getSetCCResultType(const DataLayout &DL, LLVMContext &Context, EVT VT) const
Return the ValueType of the result of SETCC operations.
virtual EVT getTypeToTransformTo(LLVMContext &Context, EVT VT) const
For types supported by the target, this is an identity function.
bool isTypeLegal(EVT VT) const
Return true if the target has native support for the specified value type.
MVT getProgramPointerTy(const DataLayout &DL) const
Return the type for code pointers, which is determined by the program address space specified through...
virtual MVT getPointerTy(const DataLayout &DL, uint32_t AS=0) const
Return the pointer type for the given address space, defaults to the pointer type from the data layou...
bool isOperationLegal(unsigned Op, EVT VT) const
Return true if the specified operation is legal on this target.
virtual bool shouldExpandVectorMatch(EVT VT, unsigned SearchSize) const
Return true if the @llvm.experimental.vector.match intrinsic should be expanded for vector type ‘VT’ ...
virtual bool isProfitableToCombineMinNumMaxNum(EVT VT) const
virtual MVT getFenceOperandTy(const DataLayout &DL) const
Return the type for operands of fence.
virtual bool shouldExpandGetVectorLength(EVT CountVT, unsigned VF, bool IsScalable) const
bool isOperationLegalOrCustom(unsigned Op, EVT VT, bool LegalOnly=false) const
Return true if the specified operation is legal on this target or can be made legal with custom lower...
virtual MVT hasFastEqualityCompare(unsigned NumBits) const
Return the preferred operand type if the target has a quick way to compare integer values of the give...
MachineMemOperand::Flags getStoreMemOperandFlags(const StoreInst &SI, const DataLayout &DL) const
virtual void getTgtMemIntrinsic(SmallVectorImpl< IntrinsicInfo > &Infos, const CallBase &I, MachineFunction &MF, unsigned Intrinsic) const
Given an intrinsic, checks if on the target the intrinsic will need to map to a MemIntrinsicNode (tou...
virtual bool signExtendConstant(const ConstantInt *C) const
Return true if this constant should be sign extended when promoting to a larger type.
virtual Value * getSDagStackGuard(const Module &M, const LibcallLoweringInfo &Libcalls) const
Return the variable that's previously inserted by insertSSPDeclarations, if any, otherwise return nul...
LegalizeTypeAction getTypeAction(LLVMContext &Context, EVT VT) const
Return how we should legalize values of this type, either it is already legal (return 'Legal') or we ...
virtual Register getExceptionPointerRegister(const Constant *PersonalityFn) const
If a physical register, this returns the register that receives the exception address on entry to an ...
bool supportsUnalignedAtomics() const
Whether the target supports unaligned atomic operations.
std::vector< ArgListEntry > ArgListTy
bool isBeneficialToExpandPowI(int64_t Exponent, bool OptForSize) const
Return true if it is beneficial to expand an @llvm.powi.
MVT getFrameIndexTy(const DataLayout &DL) const
Return the type for frame index, which is determined by the alloca address space specified through th...
virtual Register getExceptionSelectorRegister(const Constant *PersonalityFn) const
If a physical register, this returns the register that receives the exception typeid on entry to a la...
virtual MVT getPointerMemTy(const DataLayout &DL, uint32_t AS=0) const
Return the in-memory pointer type for the given address space, defaults to the pointer type from the ...
MVT getRegisterType(MVT VT) const
Return the type of registers that this ValueType will eventually require.
unsigned getVectorTypeBreakdown(LLVMContext &Context, EVT VT, EVT &IntermediateVT, unsigned &NumIntermediates, MVT &RegisterVT) const
Vector types are broken down into some number of legal first class types.
virtual MVT getVPExplicitVectorLengthTy() const
Returns the type to be used for the EVL/AVL operand of VP nodes: ISD::VP_ADD, ISD::VP_SUB,...
This class defines information used to lower LLVM code to legal SelectionDAG operators that the targe...
virtual bool supportKCFIBundles() const
Return true if the target supports kcfi operand bundles.
virtual bool supportPtrAuthBundles() const
Return true if the target supports ptrauth operand bundles.
virtual bool supportSwiftError() const
Return true if the target supports swifterror attribute.
virtual SDValue visitMaskedLoad(SelectionDAG &DAG, const SDLoc &DL, SDValue Chain, MachineMemOperand *MMO, SDValue &NewLoad, SDValue Ptr, SDValue PassThru, SDValue Mask) const
virtual SDValue emitStackGuardXorFP(SelectionDAG &DAG, SDValue Val, const SDLoc &DL) const
virtual EVT getTypeForExtReturn(LLVMContext &Context, EVT VT, ISD::NodeType) const
Return the type that should be used to zero or sign extend a zeroext/signext integer return value.
virtual InlineAsm::ConstraintCode getInlineAsmMemConstraint(StringRef ConstraintCode) const
std::vector< AsmOperandInfo > AsmOperandInfoVector
SDValue expandIS_FPCLASS(EVT ResultVT, SDValue Op, FPClassTest Test, SDNodeFlags Flags, const SDLoc &DL, SelectionDAG &DAG) const
Expand check for floating point class.
virtual SDValue prepareVolatileOrAtomicLoad(SDValue Chain, const SDLoc &DL, SelectionDAG &DAG) const
This callback is used to prepare for a volatile or atomic load.
virtual ConstraintType getConstraintType(StringRef Constraint) const
Given a constraint, return the type of constraint it is for this target.
virtual bool splitValueIntoRegisterParts(SelectionDAG &DAG, const SDLoc &DL, SDValue Val, SDValue *Parts, unsigned NumParts, MVT PartVT, std::optional< CallingConv::ID > CC) const
Target-specific splitting of values into parts that fit a register storing a legal type.
virtual SDValue joinRegisterPartsIntoValue(SelectionDAG &DAG, const SDLoc &DL, const SDValue *Parts, unsigned NumParts, MVT PartVT, EVT ValueVT, std::optional< CallingConv::ID > CC) const
Target-specific combining of register parts into its original value.
virtual SDValue LowerCall(CallLoweringInfo &, SmallVectorImpl< SDValue > &) const
This hook must be implemented to lower calls into the specified DAG.
std::pair< SDValue, SDValue > LowerCallTo(CallLoweringInfo &CLI) const
This function lowers an abstract call to a function into an actual call.
virtual SDValue LowerAsmOutputForConstraint(SDValue &Chain, SDValue &Glue, const SDLoc &DL, const AsmOperandInfo &OpInfo, SelectionDAG &DAG) const
virtual std::pair< unsigned, const TargetRegisterClass * > getRegForInlineAsmConstraint(const TargetRegisterInfo *TRI, StringRef Constraint, MVT VT) const
Given a physical register constraint (e.g.
virtual AsmOperandInfoVector ParseConstraints(const DataLayout &DL, const TargetRegisterInfo *TRI, const CallBase &Call) const
Split up the constraint string from the inline assembly value into the specific constraints and their...
virtual SDValue LowerOperation(SDValue Op, SelectionDAG &DAG) const
This callback is invoked for operations that are unsupported by the target, which are registered to u...
virtual bool functionArgumentNeedsConsecutiveRegisters(Type *Ty, CallingConv::ID CallConv, bool isVarArg, const DataLayout &DL) const
For some targets, an LLVM struct type must be broken down into multiple simple types,...
virtual void ComputeConstraintToUse(AsmOperandInfo &OpInfo, SDValue Op, SelectionDAG *DAG=nullptr) const
Determines the constraint code and constraint type to use for the specific AsmOperandInfo,...
virtual void CollectTargetIntrinsicOperands(const CallInst &I, SmallVectorImpl< SDValue > &Ops, SelectionDAG &DAG) const
virtual SDValue visitMaskedStore(SelectionDAG &DAG, const SDLoc &DL, SDValue Chain, MachineMemOperand *MMO, SDValue Ptr, SDValue Val, SDValue Mask) const
virtual bool useLoadStackGuardNode(const Module &M) const
If this function returns true, SelectionDAGBuilder emits a LOAD_STACK_GUARD node when it is lowering ...
virtual void LowerAsmOperandForConstraint(SDValue Op, StringRef Constraint, std::vector< SDValue > &Ops, SelectionDAG &DAG) const
Lower the specified operand into the Ops vector.
std::pair< SDValue, SDValue > makeLibCall(SelectionDAG &DAG, RTLIB::LibcallImpl LibcallImpl, EVT RetVT, ArrayRef< SDValue > Ops, MakeLibCallOptions CallOptions, const SDLoc &dl, SDValue Chain=SDValue()) const
Returns a pair of (return value, chain).
virtual void LowerOperationWrapper(SDNode *N, SmallVectorImpl< SDValue > &Results, SelectionDAG &DAG) const
This callback is invoked by the type legalizer to legalize nodes with an illegal operand type but leg...
virtual bool isInlineAsmTargetBranch(const SmallVectorImpl< StringRef > &AsmStrs, unsigned OpNo) const
On x86, return true if the operand with index OpNo is a CALL or JUMP instruction, which can use eithe...
virtual MVT getJumpTableRegTy(const DataLayout &DL) const
virtual bool CanLowerReturn(CallingConv::ID, MachineFunction &, bool, const SmallVectorImpl< ISD::OutputArg > &, LLVMContext &, const Type *RetTy) const
This hook should be implemented to check whether the return values described by the Outs array can fi...
Primary interface to the complete machine description for the target machine.
CodeGenOptLevel getOptLevel() const
Returns the optimization level: None, Less, Default, or Aggressive.
CodeModel::Model getCodeModel() const
Returns the code model.
unsigned NoTrapAfterNoreturn
Do not emit a trap instruction for 'unreachable' IR instructions behind noreturn calls,...
unsigned TrapUnreachable
Emit target-specific trap instruction for 'unreachable' IR instructions.
unsigned getID() const
Return the register class ID number.
const MCPhysReg * iterator
iterator begin() const
begin/end - Return all of the registers in this class.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
virtual const TargetFrameLowering * getFrameLowering() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
static constexpr TypeSize getFixed(ScalarTy ExactSize)
The instances of the Type class are immutable: once they are created, they are never changed.
LLVM_ABI bool isEmptyTy() const
Return true if this type is empty, that is, it has no elements or all of its elements are empty.
bool isVectorTy() const
True if this is an instance of VectorType.
bool isPointerTy() const
True if this is an instance of PointerType.
LLVM_ABI unsigned getPointerAddressSpace() const
Get the address space of this pointer or pointer vector type.
static LLVM_ABI Type * getVoidTy(LLVMContext &C)
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
LLVMContext & getContext() const
Return the LLVMContext in which this type was uniqued.
static LLVM_ABI IntegerType * getInt1Ty(LLVMContext &C)
bool isIntegerTy() const
True if this is an instance of IntegerType.
bool isTokenTy() const
Return true if this is 'token'.
static LLVM_ABI IntegerType * getIntNTy(LLVMContext &C, unsigned N)
bool isFPOrFPVectorTy() const
Return true if this is a FP type or a vector of FP.
bool isVoidTy() const
Return true if this is 'void'.
Unconditional Branch instruction.
This function has undefined behavior.
A Use represents the edge between a Value definition and its users.
Value * getOperand(unsigned i) const
unsigned getNumOperands() const
This class represents the va_arg llvm instruction, which returns an argument of the specified type gi...
LLVM_ABI CmpInst::Predicate getPredicate() const
This is the common base class for vector predication intrinsics.
static LLVM_ABI std::optional< unsigned > getVectorLengthParamPos(Intrinsic::ID IntrinsicID)
LLVM_ABI MaybeAlign getPointerAlignment() const
LLVM Value Representation.
Type * getType() const
All values are typed, get the type of this value.
bool hasOneUse() const
Return true if there is exactly one use of this value.
LLVMContext & getContext() const
All values hold a context through their type.
iterator_range< user_iterator > users()
LLVM_ABI const Value * stripPointerCasts() const
Strip off pointer casts, all-zero GEPs and address space casts.
LLVM_ABI StringRef getName() const
Return a constant reference to the value's name.
Base class of all SIMD vector types.
Type * getElementType() const
constexpr ScalarTy getFixedValue() const
static constexpr bool isKnownLE(const FixedOrScalableQuantity &LHS, const FixedOrScalableQuantity &RHS)
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
const ParentTy * getParent() const
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
constexpr char Args[]
Key for Kernel::Metadata::mArgs.
constexpr char SymbolName[]
Key for Kernel::Metadata::mSymbolName.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ AnyReg
OBSOLETED - Used for stack based JavaScript calls.
@ AMDGPU_CS_Chain
Used on AMDGPUs to give the middle-end more control over argument placement.
@ X86_VectorCall
MSVC calling convention that passes vectors and vector aggregates in SSE registers.
@ C
The default llvm calling convention, compatible with C.
NodeType
ISD::NodeType enum - This enum defines the target-independent operators for a SelectionDAG.
@ MERGE_VALUES
MERGE_VALUES - This node takes multiple discrete operands and returns them all as its individual resu...
@ STACKRESTORE
STACKRESTORE has two operands, an input chain and a pointer to restore to it returns an output chain.
@ STACKSAVE
STACKSAVE - STACKSAVE has one operand, an input chain.
@ CONVERGENCECTRL_ANCHOR
The llvm.experimental.convergence.* intrinsics.
@ STRICT_FSETCC
STRICT_FSETCC/STRICT_FSETCCS - Constrained versions of SETCC, used for floating-point operands only.
@ DELETED_NODE
DELETED_NODE - This is an illegal value that is used to catch errors.
@ SET_FPENV
Sets the current floating-point environment.
@ ATOMIC_LOAD_FMINIMUMNUM
@ LOOP_DEPENDENCE_RAW_MASK
@ VECREDUCE_SEQ_FADD
Generic reduction nodes.
@ COND_LOOP
COND_LOOP is a conditional branch to self, used for implementing efficient conditional traps.
@ EH_SJLJ_LONGJMP
OUTCHAIN = EH_SJLJ_LONGJMP(INCHAIN, buffer) This corresponds to the eh.sjlj.longjmp intrinsic.
@ INSERT_SUBVECTOR
INSERT_SUBVECTOR(VECTOR1, VECTOR2, IDX) - Returns a vector with VECTOR2 inserted into VECTOR1.
@ STACKADDRESS
STACKADDRESS - Represents the llvm.stackaddress intrinsic.
@ BSWAP
Byte Swap and Counting operators.
@ SMULFIX
RESULT = [US]MULFIX(LHS, RHS, SCALE) - Perform fixed point multiplication on 2 integers with the same...
@ VAEND
VAEND, VASTART - VAEND and VASTART have three operands: an input chain, pointer, and a SRCVALUE.
@ ATOMIC_STORE
OUTCHAIN = ATOMIC_STORE(INCHAIN, val, ptr) This corresponds to "store atomic" instruction.
@ RESET_FPENV
Set floating-point environment to default state.
@ ADD
Simple integer binary arithmetic operators.
@ SMULFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
@ SET_FPMODE
Sets the current dynamic floating-point control modes.
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
@ CTTZ_ELTS
Returns the number of number of trailing (least significant) zero elements in a vector.
@ FMA
FMA - Perform a * b + c with no intermediate rounding step.
@ VECTOR_FIND_LAST_ACTIVE
Finds the index of the last active mask element Operands: Mask.
@ FMODF
FMODF - Decomposes the operand into integral and fractional parts, each having the same type and sign...
@ FATAN2
FATAN2 - atan2, inspired by libm.
@ FSINCOSPI
FSINCOSPI - Compute both the sine and cosine times pi more accurately than FSINCOS(pi*x),...
@ INTRINSIC_VOID
OUTCHAIN = INTRINSIC_VOID(INCHAIN, INTRINSICID, arg1, arg2, ...) This node represents a target intrin...
@ EH_SJLJ_SETUP_DISPATCH
OUTCHAIN = EH_SJLJ_SETUP_DISPATCH(INCHAIN) The target initializes the dispatch table here.
@ ATOMIC_CMP_SWAP_WITH_SUCCESS
Val, Success, OUTCHAIN = ATOMIC_CMP_SWAP_WITH_SUCCESS(INCHAIN, ptr, cmp, swap) N.b.
@ SINT_TO_FP
[SU]INT_TO_FP - These operators convert integers (whose interpreted sign depends on the first letter)...
@ CONCAT_VECTORS
CONCAT_VECTORS(VECTOR0, VECTOR1, ...) - Given a number of values of vector type with the same length ...
@ VECREDUCE_FMAX
FMIN/FMAX nodes can have flags, for NaN/NoNaN variants.
@ FADD
Simple binary floating point operators.
@ VECREDUCE_FMAXIMUM
FMINIMUM/FMAXIMUM nodes propatate NaNs and signed zeroes using the llvm.minimum and llvm....
@ ABS
ABS - Determine the unsigned absolute value of a signed integer value of the same bitwidth.
@ ATOMIC_FENCE
OUTCHAIN = ATOMIC_FENCE(INCHAIN, ordering, scope) This corresponds to the fence instruction.
@ RESET_FPMODE
Sets default dynamic floating-point control modes.
@ FMULADD
FMULADD - Performs a * b + c, with, or without, intermediate rounding.
@ FPTRUNC_ROUND
FPTRUNC_ROUND - This corresponds to the fptrunc_round intrinsic.
@ FAKE_USE
FAKE_USE represents a use of the operand but does not do anything.
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
@ BUILD_PAIR
BUILD_PAIR - This is the opposite of EXTRACT_ELEMENT in some ways.
@ CLMUL
Carry-less multiplication operations.
@ INIT_TRAMPOLINE
INIT_TRAMPOLINE - This corresponds to the init_trampoline intrinsic.
@ FLDEXP
FLDEXP - ldexp, inspired by libm (op0 * 2**op1).
@ SDIVFIX
RESULT = [US]DIVFIX(LHS, RHS, SCALE) - Perform fixed point division on 2 integers with the same width...
@ CONVERT_FROM_ARBITRARY_FP
CONVERT_FROM_ARBITRARY_FP - This operator converts from an arbitrary floating-point represented as an...
@ EH_LABEL
EH_LABEL - Represents a label in mid basic block used to track locations needed for debug and excepti...
@ EH_RETURN
OUTCHAIN = EH_RETURN(INCHAIN, OFFSET, HANDLER) - This node represents 'eh_return' gcc dwarf builtin,...
@ ANNOTATION_LABEL
ANNOTATION_LABEL - Represents a mid basic block label used by annotations.
@ SET_ROUNDING
Set rounding mode.
@ CONVERGENCECTRL_GLUE
This does not correspond to any convergence control intrinsic.
@ SIGN_EXTEND
Conversion operators.
@ PREALLOCATED_SETUP
PREALLOCATED_SETUP - This has 2 operands: an input chain and a SRCVALUE with the preallocated call Va...
@ READSTEADYCOUNTER
READSTEADYCOUNTER - This corresponds to the readfixedcounter intrinsic.
@ ADDROFRETURNADDR
ADDROFRETURNADDR - Represents the llvm.addressofreturnaddress intrinsic.
@ BR
Control flow instructions. These all have token chains.
@ VECREDUCE_FADD
These reductions have relaxed evaluation order semantics, and have a single vector operand.
@ CTTZ_ZERO_UNDEF
Bit counting operators with an undefined result for zero inputs.
@ PREFETCH
PREFETCH - This corresponds to a prefetch intrinsic.
@ FSINCOS
FSINCOS - Compute both fsin and fcos as a single operation.
@ SSUBO
Same for subtraction.
@ PREALLOCATED_ARG
PREALLOCATED_ARG - This has 3 operands: an input chain, a SRCVALUE with the preallocated call Value,...
@ BRIND
BRIND - Indirect branch.
@ BR_JT
BR_JT - Jumptable branch.
@ VECTOR_INTERLEAVE
VECTOR_INTERLEAVE(VEC1, VEC2, ...) - Returns N vectors from N input vectors, where N is the factor to...
@ FCANONICALIZE
Returns platform specific canonical encoding of a floating point number.
@ IS_FPCLASS
Performs a check of floating point class property, defined by IEEE-754.
@ SSUBSAT
RESULT = [US]SUBSAT(LHS, RHS) - Perform saturation subtraction on 2 integers with the same bit width ...
@ SELECT
Select(COND, TRUEVAL, FALSEVAL).
@ EXTRACT_ELEMENT
EXTRACT_ELEMENT - This is used to get the lower or upper (determined by a Constant,...
@ SPLAT_VECTOR
SPLAT_VECTOR(VAL) - Returns a vector with the scalar value VAL duplicated in all lanes.
@ VACOPY
VACOPY - VACOPY has 5 operands: an input chain, a destination pointer, a source pointer,...
@ GET_ACTIVE_LANE_MASK
GET_ACTIVE_LANE_MASK - this corrosponds to the llvm.get.active.lane.mask intrinsic.
@ BasicBlock
Various leaf nodes.
@ CopyFromReg
CopyFromReg - This node indicates that the input value is a virtual or physical register that is defi...
@ SADDO
RESULT, BOOL = [SU]ADDO(LHS, RHS) - Overflow-aware nodes for addition.
@ ARITH_FENCE
ARITH_FENCE - This corresponds to a arithmetic fence intrinsic.
@ VECREDUCE_ADD
Integer reductions may have a result type larger than the vector element type.
@ GET_ROUNDING
Returns current rounding mode: -1 Undefined 0 Round to 0 1 Round to nearest, ties to even 2 Round to ...
@ CLEANUPRET
CLEANUPRET - Represents a return from a cleanup block funclet.
@ GET_FPMODE
Reads the current dynamic floating-point control modes.
@ GET_FPENV
Gets the current floating-point environment.
@ SHL
Shift and rotation operations.
@ AssertNoFPClass
AssertNoFPClass - These nodes record if a register contains a float value that is known to be not som...
@ PtrAuthGlobalAddress
A ptrauth constant.
@ EXTRACT_SUBVECTOR
EXTRACT_SUBVECTOR(VECTOR, IDX) - Returns a subvector from VECTOR.
@ EntryToken
EntryToken - This is the marker used to indicate the start of a region.
@ READ_REGISTER
READ_REGISTER, WRITE_REGISTER - This node represents llvm.register on the DAG, which implements the n...
@ EXTRACT_VECTOR_ELT
EXTRACT_VECTOR_ELT(VECTOR, IDX) - Returns a single element from VECTOR identified by the (potentially...
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
@ DEBUGTRAP
DEBUGTRAP - Trap intended to get the attention of a debugger.
@ VSCALE
VSCALE(IMM) - Returns the runtime scaling factor used to calculate the number of elements within a sc...
@ LOCAL_RECOVER
LOCAL_RECOVER - Represents the llvm.localrecover intrinsic.
@ FMINNUM
FMINNUM/FMAXNUM - Perform floating-point minimum maximum on two values, following IEEE-754 definition...
@ UBSANTRAP
UBSANTRAP - Trap with an immediate describing the kind of sanitizer failure.
@ SSHLSAT
RESULT = [US]SHLSAT(LHS, RHS) - Perform saturation left shift.
@ PATCHPOINT
The llvm.experimental.patchpoint.
@ SMULO
Same for multiplication.
@ DYNAMIC_STACKALLOC
DYNAMIC_STACKALLOC - Allocate some number of bytes on the stack aligned to a specified boundary.
@ VECTOR_SPLICE_LEFT
VECTOR_SPLICE_LEFT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1, VEC2) left by OFFSET elements an...
@ SMIN
[US]{MIN/MAX} - Binary minimum or maximum of signed or unsigned integers.
@ MASKED_UDIV
Masked vector arithmetic that returns poison on disabled lanes.
@ VECTOR_REVERSE
VECTOR_REVERSE(VECTOR) - Returns a vector, of the same type as VECTOR, whose elements are shuffled us...
@ SDIVFIXSAT
Same as the corresponding unsaturated fixed point instructions, but the result is clamped between the...
@ FP_EXTEND
X = FP_EXTEND(Y) - Extend a smaller FP type into a larger FP type.
@ VSELECT
Select with a vector condition (op #0) and two vector operands (ops #1 and #2), returning a vector re...
@ PCMARKER
PCMARKER - This corresponds to the pcmarker intrinsic.
@ INLINEASM_BR
INLINEASM_BR - Branching version of inline asm. Used by asm-goto.
@ ATOMIC_LOAD_FMAXIMUMNUM
@ EH_DWARF_CFA
EH_DWARF_CFA - This node represents the pointer to the DWARF Canonical Frame Address (CFA),...
@ FRAMEADDR
FRAMEADDR, RETURNADDR - These nodes represent llvm.frameaddress and llvm.returnaddress on the DAG.
@ STRICT_FP_ROUND
X = STRICT_FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision ...
@ FMINIMUM
FMINIMUM/FMAXIMUM - NaN-propagating minimum/maximum that also treat -0.0 as less than 0....
@ FP_TO_SINT
FP_TO_[US]INT - Convert a floating point value to a signed or unsigned integer.
@ READCYCLECOUNTER
READCYCLECOUNTER - This corresponds to the readcyclecounter intrinsic.
@ RELOC_NONE
Issue a no-op relocation against a given symbol at the current location.
@ AND
Bitwise operators - logical and, logical or, logical xor.
@ TRAP
TRAP - Trapping instruction.
@ INTRINSIC_WO_CHAIN
RESULT = INTRINSIC_WO_CHAIN(INTRINSICID, arg1, arg2, ...) This node represents a target intrinsic fun...
@ SCMP
[US]CMP - 3-way comparison of signed or unsigned integers.
@ VECTOR_SPLICE_RIGHT
VECTOR_SPLICE_RIGHT(VEC1, VEC2, OFFSET) - Shifts CONCAT_VECTORS(VEC1,VEC2) right by OFFSET elements a...
@ STRICT_FADD
Constrained versions of the binary floating point operators.
@ STACKMAP
The llvm.experimental.stackmap intrinsic.
@ FREEZE
FREEZE - FREEZE(VAL) returns an arbitrary value if VAL is UNDEF (or is evaluated to UNDEF),...
@ INSERT_VECTOR_ELT
INSERT_VECTOR_ELT(VECTOR, VAL, IDX) - Returns VECTOR with the element at IDX replaced with VAL.
@ TokenFactor
TokenFactor - This node takes multiple tokens as input and produces a single token result.
@ ATOMIC_SWAP
Val, OUTCHAIN = ATOMIC_SWAP(INCHAIN, ptr, amt) Val, OUTCHAIN = ATOMIC_LOAD_[OpName](INCHAIN,...
@ FFREXP
FFREXP - frexp, extract fractional and exponent component of a floating-point value.
@ FP_ROUND
X = FP_ROUND(Y, TRUNC) - Rounding 'Y' from a larger floating point type down to the precision of the ...
@ VECTOR_COMPRESS
VECTOR_COMPRESS(Vec, Mask, Passthru) consecutively place vector elements based on mask e....
@ SPONENTRY
SPONENTRY - Represents the llvm.sponentry intrinsic.
@ CLEAR_CACHE
llvm.clear_cache intrinsic Operands: Input Chain, Start Addres, End Address Outputs: Output Chain
@ INLINEASM
INLINEASM - Represents an inline asm block.
@ FP_TO_SINT_SAT
FP_TO_[US]INT_SAT - Convert floating point value in operand 0 to a signed or unsigned scalar integer ...
@ EH_SJLJ_SETJMP
RESULT, OUTCHAIN = EH_SJLJ_SETJMP(INCHAIN, buffer) This corresponds to the eh.sjlj....
@ TRUNCATE
TRUNCATE - Completely drop the high bits.
@ BRCOND
BRCOND - Conditional branch.
@ CATCHRET
CATCHRET - Represents a return from a catch block funclet.
@ AssertSext
AssertSext, AssertZext - These nodes record if a register contains a value that has already been zero...
@ FCOPYSIGN
FCOPYSIGN(X, Y) - Return the value of X with the sign of Y.
@ SADDSAT
RESULT = [US]ADDSAT(LHS, RHS) - Perform saturation addition on 2 integers with the same bit width (W)...
@ VECTOR_DEINTERLEAVE
VECTOR_DEINTERLEAVE(VEC1, VEC2, ...) - Returns N vectors from N input vectors, where N is the factor ...
@ GET_DYNAMIC_AREA_OFFSET
GET_DYNAMIC_AREA_OFFSET - get offset from native SP to the address of the most recent dynamic alloca.
@ FMINIMUMNUM
FMINIMUMNUM/FMAXIMUMNUM - minimumnum/maximumnum that is same with FMINNUM_IEEE and FMAXNUM_IEEE besid...
@ ADJUST_TRAMPOLINE
ADJUST_TRAMPOLINE - This corresponds to the adjust_trampoline intrinsic.
@ INTRINSIC_W_CHAIN
RESULT,OUTCHAIN = INTRINSIC_W_CHAIN(INCHAIN, INTRINSICID, arg1, ...) This node represents a target in...
@ BUILD_VECTOR
BUILD_VECTOR(ELT0, ELT1, ELT2, ELT3,...) - Return a fixed-width vector with the specified,...
@ LOOP_DEPENDENCE_WAR_MASK
The llvm.loop.dependence.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
This namespace contains an enum with a value for every intrinsic/builtin function known by LLVM.
Flag
These should be considered private to the implementation of the MCInstrDesc class.
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
bool match(Val *V, const Pattern &P)
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
TwoOps_match< Val_t, Idx_t, Instruction::ExtractElement > m_ExtractElt(const Val_t &Val, const Idx_t &Idx)
Matches ExtractElementInst.
IntrinsicID_match m_VScale()
Matches a call to llvm.vscale().
auto m_Value()
Match an arbitrary value and ignore it.
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
std::pair< JumpTableHeader, JumpTable > JumpTableBlock
void sortAndRangeify(CaseClusterVector &Clusters)
Sort Clusters and merge adjacent cases.
std::vector< CaseCluster > CaseClusterVector
@ CC_Range
A cluster of adjacent case labels with the same destination, or just one case.
@ CC_JumpTable
A cluster of cases suitable for jump table lowering.
@ CC_BitTests
A cluster of cases suitable for bit test lowering.
SmallVector< SwitchWorkListItem, 4 > SwitchWorkList
CaseClusterVector::iterator CaseClusterIt
initializer< Ty > init(const Ty &Val)
LocationClass< Ty > location(Ty &L)
@ DW_OP_LLVM_arg
Only used in LLVM metadata.
ExceptionBehavior
Exception behavior used for floating point operations.
@ ebStrict
This corresponds to "fpexcept.strict".
@ ebMayTrap
This corresponds to "fpexcept.maytrap".
@ ebIgnore
This corresponds to "fpexcept.ignore".
NodeAddr< FuncNode * > Func
friend class Instruction
Iterator for Instructions in a `BasicBlock.
This is an optimization pass for GlobalISel generic memory operations.
@ Low
Lower the current thread's priority such that it does not affect foreground tasks significantly.
unsigned Log2_32_Ceil(uint32_t Value)
Return the ceil log base 2 of the specified value, 32 if the value is zero.
FunctionAddr VTableAddr Value
ISD::CondCode getICmpCondCode(ICmpInst::Predicate Pred)
getICmpCondCode - Return the ISD condition code corresponding to the given LLVM IR integer condition ...
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
auto size(R &&Range, std::enable_if_t< std::is_base_of< std::random_access_iterator_tag, typename std::iterator_traits< decltype(Range.begin())>::iterator_category >::value, void > *=nullptr)
Get the size of a range.
SDValue peekThroughFreeze(SDValue V)
Return the non-frozen source operand of V if it exists.
LLVM_ABI void GetReturnInfo(CallingConv::ID CC, Type *ReturnType, AttributeList attr, SmallVectorImpl< ISD::OutputArg > &Outs, const TargetLowering &TLI, const DataLayout &DL)
Given an LLVM IR type and return type attributes, compute the return value EVTs and flags,...
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
LLVM_ABI bool isOnlyUsedInZeroEqualityComparison(const Instruction *CxtI)
void ComputeValueVTs(const TargetLowering &TLI, const DataLayout &DL, Type *Ty, SmallVectorImpl< EVT > &ValueVTs, SmallVectorImpl< EVT > *MemVTs=nullptr, SmallVectorImpl< TypeSize > *Offsets=nullptr, TypeSize StartingOffset=TypeSize::getZero())
ComputeValueVTs - Given an LLVM IR type, compute a sequence of EVTs that represent all the individual...
LLVM_ABI SDValue peekThroughBitcasts(SDValue V)
Return the non-bitcasted source operand of V if it exists.
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
int countr_one(T Value)
Count the number of ones from the least significant bit to the first zero bit.
LLVM_ABI void diagnoseDontCall(const CallInst &CI)
auto successors(const MachineBasicBlock *BB)
bool isIntOrFPConstant(SDValue V)
Return true if V is either a integer or FP constant.
static ConstantRange getRange(Value *Op, SCCPSolver &Solver, const SmallPtrSetImpl< Value * > &InsertedValues)
Helper for getting ranges from Solver.
void append_range(Container &C, Range &&R)
Wrapper function to append range R to container C.
Value * GetPointerBaseWithConstantOffset(Value *Ptr, int64_t &Offset, const DataLayout &DL, bool AllowNonInbounds=true)
Analyze the specified pointer to see if it can be expressed as a base pointer plus a constant offset.
constexpr bool isUIntN(unsigned N, uint64_t x)
Checks if an unsigned integer fits into the given (dynamic) bit width.
auto cast_or_null(const Y &Val)
constexpr T alignDown(U Value, V Align, W Skew=0)
Returns the largest unsigned integer less than or equal to Value and is Skew mod Align.
gep_type_iterator gep_type_end(const User *GEP)
constexpr auto equal_to(T &&Arg)
Functor variant of std::equal_to that can be used as a UnaryPredicate in functional algorithms like a...
constexpr int popcount(T Value) noexcept
Count the number of set bits in a value.
LLVM_ABI ConstantRange getConstantRangeFromMetadata(const MDNode &RangeMD)
Parse out a conservative ConstantRange from !range metadata.
detail::concat_range< ValueT, RangeTs... > concat(RangeTs &&...Ranges)
Returns a concatenated range across two or more ranges.
bool isScopedEHPersonality(EHPersonality Pers)
Returns true if this personality uses scope-style EH IR instructions: catchswitch,...
int countr_zero(T Val)
Count number of 0's from the least significant bit to the most stopping at the first 1.
void ComputeValueTypes(const DataLayout &DL, Type *Ty, SmallVectorImpl< Type * > &Types, SmallVectorImpl< TypeSize > *Offsets=nullptr, TypeSize StartingOffset=TypeSize::getZero())
Given an LLVM IR type, compute non-aggregate subtypes.
auto dyn_cast_or_null(const Y &Val)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
LLVM_ABI llvm::SmallVector< int, 16 > createStrideMask(unsigned Start, unsigned Stride, unsigned VF)
Create a stride shuffle mask.
@ SPF_ABS
Floating point maxnum.
@ SPF_NABS
Absolute value.
@ SPF_FMAXNUM
Floating point minnum.
@ SPF_UMIN
Signed minimum.
@ SPF_UMAX
Signed maximum.
@ SPF_SMAX
Unsigned minimum.
@ SPF_FMINNUM
Unsigned maximum.
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
detail::zippy< detail::zip_first, T, U, Args... > zip_first(T &&t, U &&u, Args &&...args)
zip iterator that, for the sake of efficiency, assumes the first iteratee to be the shortest.
void sort(IteratorTy Start, IteratorTy End)
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
LLVM_ABI SelectPatternResult matchSelectPattern(Value *V, Value *&LHS, Value *&RHS, Instruction::CastOps *CastOp=nullptr, unsigned Depth=0)
Pattern match integer [SU]MIN, [SU]MAX and ABS idioms, returning the kind and providing the out param...
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
constexpr uint64_t alignTo(uint64_t Size, Align A)
Returns a multiple of A needed to store Size bytes.
generic_gep_type_iterator<> gep_type_iterator
FunctionAddr VTableAddr Count
auto succ_size(const MachineBasicBlock *BB)
bool hasSingleElement(ContainerTy &&C)
Returns true if the given container only contains a single element.
ISD::CondCode getFCmpCondCode(FCmpInst::Predicate Pred)
getFCmpCondCode - Return the ISD condition code corresponding to the given LLVM IR floating-point con...
LLVM_ABI EHPersonality classifyEHPersonality(const Value *Pers)
See if the given exception handling personality function is one that we understand.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
LLVM_ATTRIBUTE_VISIBILITY_DEFAULT AnalysisKey InnerAnalysisManagerProxy< AnalysisManagerT, IRUnitT, ExtraArgTs... >::Key
LLVM_ABI Value * salvageDebugInfoImpl(Instruction &I, uint64_t CurrentLocOps, SmallVectorImpl< uint64_t > &Ops, SmallVectorImpl< Value * > &AdditionalValues)
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
@ Global
Append to llvm.global_dtors.
AtomicOrdering
Atomic ordering for LLVM's memory model.
@ First
Helpers to iterate all locations in the MemoryEffectsBase class.
bool isFuncletEHPersonality(EHPersonality Pers)
Returns true if this is a personality function that invokes handler funclets (which must return to it...
FunctionAddr VTableAddr uintptr_t uintptr_t Data
LLVM_ABI bool isAssignmentTrackingEnabled(const Module &M)
Return true if assignment tracking is enabled for module M.
LLVM_ABI llvm::SmallVector< int, 16 > createInterleaveMask(unsigned VF, unsigned NumVecs)
Create an interleave shuffle mask.
@ UMin
Unsigned integer min implemented in terms of select(cmp()).
@ Or
Bitwise or logical OR of integers.
@ Mul
Product of integers.
@ Sub
Subtraction of integers.
@ SPNB_RETURNS_NAN
NaN behavior not applicable.
@ SPNB_RETURNS_OTHER
Given one NaN input, returns the NaN.
@ SPNB_RETURNS_ANY
Given one NaN input, returns the non-NaN.
bool isInTailCallPosition(const CallBase &Call, const TargetMachine &TM, bool ReturnsFirstArg=false)
Test if the given instruction is in a position to be optimized with a tail-call.
DWARFExpression::Operation Op
ISD::CondCode getFCmpCodeWithoutNaN(ISD::CondCode CC)
getFCmpCodeWithoutNaN - Given an ISD condition code comparing floats, return the equivalent code if w...
ArrayRef(const T &OneElt) -> ArrayRef< T >
bool isAsynchronousEHPersonality(EHPersonality Pers)
Returns true if this personality function catches asynchronous exceptions.
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
LLVM_ABI bool isKnownNeverNaN(const Value *V, const SimplifyQuery &SQ, unsigned Depth=0)
Return true if the floating-point scalar value is not a NaN or if the floating-point vector value has...
LLVM_ABI std::optional< RoundingMode > convertStrToRoundingMode(StringRef)
Returns a valid RoundingMode enumerator when given a string that is valid as input in constrained int...
gep_type_iterator gep_type_begin(const User *GEP)
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
GlobalValue * ExtractTypeInfo(Value *V)
ExtractTypeInfo - Returns the type info, possibly bitcast, encoded in V.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
bool all_equal(std::initializer_list< T > Values)
Returns true if all Values in the initializer lists are equal or the list.
LLVM_ABI Constant * ConstantFoldLoadFromConstPtr(Constant *C, Type *Ty, APInt Offset, const DataLayout &DL)
Return the value that a load from C with offset Offset would produce if it is constant and determinab...
unsigned ComputeLinearIndex(Type *Ty, const unsigned *Indices, const unsigned *IndicesEnd, unsigned CurIndex=0)
Compute the linearized index of a member in a nested aggregate/struct/array.
T bit_floor(T Value)
Returns the largest integral power of two no greater than Value if Value is nonzero.
@ Default
The result value is uniform if and only if all operands are uniform.
LLVM_ABI void reportFatalUsageError(Error Err)
Report a fatal error that does not indicate a bug in LLVM.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
This struct is a compact representation of a valid (non-zero power of two) alignment.
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
TypeSize getStoreSize() const
Return the number of bytes overwritten by a store of the specified value type.
static EVT getVectorVT(LLVMContext &Context, EVT VT, unsigned NumElements, bool IsScalable=false)
Returns the EVT that represents a vector NumElements in length, where each element is of type VT.
uint64_t getScalarStoreSize() const
bool bitsGT(EVT VT) const
Return true if this has more bits than VT.
bool bitsLT(EVT VT) const
Return true if this has less bits than VT.
bool isFloatingPoint() const
Return true if this is a FP or a vector FP type.
ElementCount getVectorElementCount() const
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
unsigned getVectorMinNumElements() const
Given a vector type, return the minimum number of elements it contains.
uint64_t getScalarSizeInBits() const
static LLVM_ABI EVT getEVT(Type *Ty, bool HandleUnknown=false)
Return the value type corresponding to the specified type.
EVT changeVectorElementType(LLVMContext &Context, EVT EltVT) const
Return a VT for a vector type whose attributes match ourselves with the exception of the element type...
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
static EVT getIntegerVT(LLVMContext &Context, unsigned BitWidth)
Returns the EVT that represents an integer with the given number of bits.
bool isRISCVVectorTuple() const
Return true if this is a vector value type.
uint64_t getFixedSizeInBits() const
Return the size of the specified fixed width value type in bits.
bool isFixedLengthVector() const
bool isVector() const
Return true if this is a vector value type.
EVT getScalarType() const
If this is a vector type, return the element type, otherwise return this.
bool bitsGE(EVT VT) const
Return true if this has no less bits than VT.
bool isScalableVector() const
Return true if this is a vector type where the runtime length is machine dependent.
EVT getVectorElementType() const
Given a vector type, return the type of each element.
EVT changeElementType(LLVMContext &Context, EVT EltVT) const
Return a VT for a type whose attributes match ourselves with the exception of the element type that i...
bool isScalarInteger() const
Return true if this is an integer, but not a vector.
unsigned getVectorNumElements() const
Given a vector type, return the number of elements it contains.
bool isInteger() const
Return true if this is an integer or a vector integer type.
void setPointerAddrSpace(unsigned AS)
void setOrigAlign(Align A)
OutputArg - This struct carries flags and a value for a single outgoing (actual) argument or outgoing...
ConstraintPrefix Type
Type - The basic type of the constraint: input/output/clobber/label.
unsigned countMinLeadingZeros() const
Returns the minimum number of leading zero bits.
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getUnknownStack(MachineFunction &MF)
Stack memory without other information.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
A lightweight accessor for an operand bundle meant to be passed around by value.
This struct represents the registers (physical or virtual) that a particular set of values is assigne...
SmallVector< std::pair< Register, TypeSize >, 4 > getRegsAndSizes() const
Return a list of registers and their sizes.
SmallVector< unsigned, 4 > RegCount
This list holds the number of registers for each value.
bool isABIMangled() const
SmallVector< EVT, 4 > ValueVTs
The value types of the values, which may not be legal, and may need be promoted or synthesized from o...
SmallVector< Register, 4 > Regs
This list holds the registers assigned to the values.
void AddInlineAsmOperands(InlineAsm::Kind Code, bool HasMatching, unsigned MatchingIdx, const SDLoc &dl, SelectionDAG &DAG, std::vector< SDValue > &Ops) const
Add this value to the specified inlineasm node operand list.
SDValue getCopyFromRegs(SelectionDAG &DAG, FunctionLoweringInfo &FuncInfo, const SDLoc &dl, SDValue &Chain, SDValue *Glue, const Value *V=nullptr) const
Emit a series of CopyFromReg nodes that copies from this value and returns the result as a ValueVTs v...
SmallVector< MVT, 4 > RegVTs
The value types of the registers.
void getCopyToRegs(SDValue Val, SelectionDAG &DAG, const SDLoc &dl, SDValue &Chain, SDValue *Glue, const Value *V=nullptr, ISD::NodeType PreferredExtendType=ISD::ANY_EXTEND) const
Emit a series of CopyToReg nodes that copies the specified value into the registers specified by this...
std::optional< CallingConv::ID > CallConv
Records if this value needs to be treated in an ABI dependant manner, different to normal type legali...
bool occupiesMultipleRegs() const
Check if the total RegCount is greater than one.
These are IR-level optimization flags that may be propagated to SDNodes.
void copyFMF(const FPMathOperator &FPMO)
Propagate the fast-math-flags from an IR FPMathOperator.
void setUnpredictable(bool b)
bool hasAllowReassociation() const
void setNoUnsignedWrap(bool b)
void setNoSignedWrap(bool b)
This represents a list of ValueType's that has been intern'd by a SelectionDAG.
A MapVector that performs no allocations if smaller than a certain size.
MachineBasicBlock * Default
BranchProbability DefaultProb
MachineBasicBlock * Parent
bool FallthroughUnreachable
MachineBasicBlock * ThisBB
This structure is used to communicate between SelectionDAGBuilder and SDISel for the code generation ...
BranchProbability TrueProb
BranchProbability FalseProb
MachineBasicBlock * TrueBB
MachineBasicBlock * FalseBB
SDLoc DL
The debug location of the instruction this CaseBlock was produced from.
static CaseCluster range(const ConstantInt *Low, const ConstantInt *High, MachineBasicBlock *MBB, BranchProbability Prob)
Register Reg
The virtual register containing the index of the jump table entry to jump to.
MachineBasicBlock * Default
The MBB of the default bb, which is a successor of the range check MBB.
unsigned JTI
The JumpTableIndex for this jump table in the function.
MachineBasicBlock * MBB
The MBB into which to emit the code for the indirect jump.
std::optional< SDLoc > SL
The debug location of the instruction this JumpTable was produced from.
This contains information for each constraint that we are lowering.
TargetLowering::ConstraintType ConstraintType
Information about the constraint code, e.g.
This structure contains all information that is necessary for lowering calls.
CallLoweringInfo & setConvergent(bool Value=true)
CallLoweringInfo & setDeactivationSymbol(GlobalValue *Sym)
CallLoweringInfo & setCFIType(const ConstantInt *Type)
SmallVector< ISD::InputArg, 32 > Ins
bool IsPostTypeLegalization
SmallVector< SDValue, 4 > InVals
Type * OrigRetTy
Original unlegalized return type.
CallLoweringInfo & setDiscardResult(bool Value=true)
CallLoweringInfo & setIsPatchPoint(bool Value=true)
CallLoweringInfo & setDebugLoc(const SDLoc &dl)
CallLoweringInfo & setTailCall(bool Value=true)
CallLoweringInfo & setIsPreallocated(bool Value=true)
CallLoweringInfo & setConvergenceControlToken(SDValue Token)
SmallVector< ISD::OutputArg, 32 > Outs
SmallVector< SDValue, 32 > OutVals
Type * RetTy
Same as OrigRetTy, or partially legalized for soft float libcalls.
CallLoweringInfo & setChain(SDValue InChain)
CallLoweringInfo & setPtrAuth(PtrAuthInfo Value)
CallLoweringInfo & setCallee(CallingConv::ID CC, Type *ResultType, SDValue Target, ArgListTy &&ArgsList, AttributeSet ResultAttrs={})
This structure is used to pass arguments to makeLibCall function.
MakeLibCallOptions & setDiscardResult(bool Value=true)
This structure contains the information necessary for lowering pointer-authenticating indirect calls.
void addIPToStateRange(const InvokeInst *II, MCSymbol *InvokeBegin, MCSymbol *InvokeEnd)