24#include "llvm/IR/IntrinsicsSPIRV.h"
32#define DEBUG_TYPE "spirv-legalizer"
35 return [IsExtendedInts, TypeIdx](
const LegalityQuery &Query) {
36 const LLT Ty = Query.Types[TypeIdx];
37 return IsExtendedInts && Ty.isValid() && Ty.isScalar();
84 const unsigned PSize = ST.getPointerSize();
103 auto allPtrsScalarsAndVectors = {
104 p0, p1, p2, p3, p4, p5, p6, p7, p8,
105 p9, p10, p11, p12, p13, s1, s8, s16, s32,
106 s64, v2s1, v2s8, v2s16, v2s32, v2s64, v3s1, v3s8, v3s16,
107 v3s32, v3s64, v4s1, v4s8, v4s16, v4s32, v4s64, v8s1, v8s8,
108 v8s16, v8s32, v8s64, v16s1, v16s8, v16s16, v16s32, v16s64};
110 auto allVectors = {v2s1, v2s8, v2s16, v2s32, v2s64, v3s1, v3s8,
111 v3s16, v3s32, v3s64, v4s1, v4s8, v4s16, v4s32,
112 v4s64, v8s1, v8s8, v8s16, v8s32, v8s64, v16s1,
113 v16s8, v16s16, v16s32, v16s64};
115 auto allShaderVectors = {v2s1, v2s8, v2s16, v2s32, v2s64,
116 v3s1, v3s8, v3s16, v3s32, v3s64,
117 v4s1, v4s8, v4s16, v4s32, v4s64};
119 auto allScalars = {s1, s8, s16, s32, s64};
121 auto allScalarsAndVectors = {
122 s1, s8, s16, s32, s64, s128, v2s1, v2s8,
123 v2s16, v2s32, v2s64, v3s1, v3s8, v3s16, v3s32, v3s64,
124 v4s1, v4s8, v4s16, v4s32, v4s64, v8s1, v8s8, v8s16,
125 v8s32, v8s64, v16s1, v16s8, v16s16, v16s32, v16s64};
127 auto allIntScalarsAndVectors = {
128 s8, s16, s32, s64, s128, v2s8, v2s16, v2s32, v2s64,
129 v3s8, v3s16, v3s32, v3s64, v4s8, v4s16, v4s32, v4s64, v8s8,
130 v8s16, v8s32, v8s64, v16s8, v16s16, v16s32, v16s64};
132 auto allBoolScalarsAndVectors = {s1, v2s1, v3s1, v4s1, v8s1, v16s1};
134 auto allIntScalars = {s8, s16, s32, s64, s128};
136 auto allFloatScalarsAndF16Vector2AndVector4s = {s16, s32, s64, v2s16, v4s16};
138 auto allFloatScalarsAndVectors = {
139 s16, s32, s64, v2s16, v2s32, v2s64, v3s16, v3s32, v3s64,
140 v4s16, v4s32, v4s64, v8s16, v8s32, v8s64, v16s16, v16s32, v16s64};
142 auto allFloatAndIntScalarsAndPtrs = {s8, s16, s32, s64, p0, p1,
143 p2, p3, p4, p5, p6, p7,
144 p8, p9, p10, p11, p12, p13};
146 auto allPtrs = {p0, p1, p2, p3, p4, p5, p6, p7, p8, p9, p10, p11, p12, p13};
148 auto &allowedVectorTypes = ST.isShader() ? allShaderVectors : allVectors;
150 bool IsExtendedInts =
152 SPIRV::Extension::SPV_ALTERA_arbitrary_precision_integers) ||
153 ST.canUseExtension(SPIRV::Extension::SPV_KHR_bit_instructions) ||
154 ST.canUseExtension(SPIRV::Extension::SPV_INTEL_int4);
155 auto extendedScalarsAndVectors =
157 const LLT Ty = Query.Types[0];
158 return IsExtendedInts && Ty.isValid() && !Ty.isPointerOrPointerVector();
160 auto extendedScalarsAndVectorsProduct = [IsExtendedInts](
162 const LLT Ty1 = Query.Types[0], Ty2 = Query.Types[1];
163 return IsExtendedInts && Ty1.
isValid() && Ty2.isValid() &&
166 auto extendedPtrsScalarsAndVectors =
168 const LLT Ty = Query.Types[0];
169 return IsExtendedInts && Ty.isValid();
178 uint32_t MaxVectorSize = ST.isShader() ? 4 : 16;
183 case G_EXTRACT_VECTOR_ELT:
204 .customFor(allScalars)
213 .legalFor(allScalars)
277 {G_VECREDUCE_SMIN, G_VECREDUCE_SMAX, G_VECREDUCE_UMIN, G_VECREDUCE_UMAX,
278 G_VECREDUCE_ADD, G_VECREDUCE_MUL, G_VECREDUCE_FMUL, G_VECREDUCE_FMIN,
279 G_VECREDUCE_FMAX, G_VECREDUCE_FMINIMUM, G_VECREDUCE_FMAXIMUM,
280 G_VECREDUCE_OR, G_VECREDUCE_AND, G_VECREDUCE_XOR})
281 .legalFor(allowedVectorTypes)
311 .unsupportedIf(
typeIs(1, p9))
318 G_BITREVERSE, G_SADDSAT, G_UADDSAT, G_SSUBSAT,
319 G_USUBSAT, G_SCMP, G_UCMP})
320 .legalFor(allIntScalarsAndVectors)
321 .
legalIf(extendedScalarsAndVectors);
327 .legalForCartesianProduct(allIntScalarsAndVectors,
328 allFloatScalarsAndVectors);
331 .legalForCartesianProduct(allIntScalarsAndVectors,
332 allFloatScalarsAndVectors);
335 .legalForCartesianProduct(allFloatScalarsAndVectors,
336 allScalarsAndVectors);
340 .
legalIf(extendedScalarsAndVectorsProduct);
344 .legalForCartesianProduct(allScalarsAndVectors)
345 .
legalIf(extendedScalarsAndVectorsProduct);
349 .
legalIf(extendedPtrsScalarsAndVectors);
353 typeInSet(1, allPtrsScalarsAndVectors)));
356 .legalFor({s1, s128})
357 .legalFor(allFloatAndIntScalarsAndPtrs)
386 typeInSet(1, allPtrsScalarsAndVectors)));
390 typeInSet(1, allFloatScalarsAndVectors)));
393 G_ATOMICRMW_MAX, G_ATOMICRMW_MIN,
394 G_ATOMICRMW_SUB, G_ATOMICRMW_XOR,
395 G_ATOMICRMW_UMAX, G_ATOMICRMW_UMIN})
396 .legalForCartesianProduct(allIntScalars, allPtrs);
399 {G_ATOMICRMW_FADD, G_ATOMICRMW_FSUB, G_ATOMICRMW_FMIN, G_ATOMICRMW_FMAX})
400 .legalForCartesianProduct(allFloatScalarsAndF16Vector2AndVector4s,
411 {G_UADDO, G_SADDO, G_USUBO, G_SSUBO, G_UMULO, G_SMULO})
415 .legalForCartesianProduct(allFloatScalarsAndVectors,
416 allIntScalarsAndVectors);
420 .legalForCartesianProduct(allFloatScalarsAndVectors);
431 allFloatScalarsAndVectors, {s32, v2s32, v3s32, v4s32, v8s32, v16s32});
467 G_INTRINSIC_ROUNDEVEN})
468 .legalFor(allFloatScalarsAndVectors);
472 allFloatScalarsAndVectors);
475 allFloatScalarsAndVectors, allIntScalarsAndVectors);
477 if (ST.canUseExtInstSet(SPIRV::InstructionSet::OpenCL_std)) {
479 {G_CTTZ, G_CTTZ_ZERO_UNDEF, G_CTLZ, G_CTLZ_ZERO_UNDEF})
480 .legalForCartesianProduct(allIntScalarsAndVectors,
481 allIntScalarsAndVectors);
490 verify(*ST.getInstrInfo());
504 MI.eraseFromParent();
521 MI.eraseFromParent();
529 Register ConvReg =
MRI.createGenericVirtualRegister(ConvTy);
541 unsigned NumElements = Ty.getNumElements();
542 unsigned MaxVectorSize = ST.isShader() ? 4 : 16;
544 NumElements > MaxVectorSize;
553 LLT DstTy =
MRI.getType(DstReg);
566 LLT PtrTy =
MRI.getType(PtrReg);
569 for (
unsigned i = 0; i < NumElts; ++i) {
571 Register EltPtr =
MRI.createGenericVirtualRegister(PtrTy);
581 if (!
MI.memoperands_empty()) {
588 Register EltReg =
MRI.createGenericVirtualRegister(EltTy);
589 MIRBuilder.
buildLoad(EltReg, EltPtr, EltPtrInfo, EltAlign);
594 MI.eraseFromParent();
604 LLT ValTy =
MRI.getType(ValReg);
606 assert(ValTy.isVector() &&
"Expected vector store");
612 for (
unsigned i = 0; i < NumElts; ++i)
613 SplitRegs.
push_back(
MRI.createGenericVirtualRegister(EltTy));
617 LLT PtrTy =
MRI.getType(PtrReg);
620 for (
unsigned i = 0; i < NumElts; ++i) {
622 Register EltPtr =
MRI.createGenericVirtualRegister(PtrTy);
632 if (!
MI.memoperands_empty()) {
639 MIRBuilder.
buildStore(SplitRegs[i], EltPtr, EltPtrInfo, EltAlign);
642 MI.eraseFromParent();
650 switch (
MI.getOpcode()) {
654 case TargetOpcode::G_BITCAST:
655 return legalizeBitcast(Helper,
MI);
656 case TargetOpcode::G_EXTRACT_VECTOR_ELT:
658 case TargetOpcode::G_INSERT_VECTOR_ELT:
660 case TargetOpcode::G_INTRINSIC:
661 case TargetOpcode::G_INTRINSIC_W_SIDE_EFFECTS:
663 case TargetOpcode::G_IS_FPCLASS:
664 return legalizeIsFPClass(Helper,
MI, LocObserver);
665 case TargetOpcode::G_ICMP: {
666 assert(GR->getSPIRVTypeForVReg(
MI.getOperand(0).getReg()));
667 auto &Op0 =
MI.getOperand(2);
668 auto &Op1 =
MI.getOperand(3);
673 if ((!ST->canDirectlyComparePointers() ||
675 MRI.getType(Reg0).isPointer() &&
MRI.getType(Reg1).isPointer()) {
678 ST->getPointerSize());
679 SPIRVType *SpirvTy = GR->getOrCreateSPIRVType(
680 LLVMTy, Helper.
MIRBuilder, SPIRV::AccessQualifier::ReadWrite,
true);
686 case TargetOpcode::G_LOAD:
688 case TargetOpcode::G_STORE:
708 const Type *LLVMArrTy =
711 LLVMArrTy, MIRBuilder, SPIRV::AccessQualifier::ReadWrite,
true);
713 ArrSpvTy, MIRBuilder, SPIRV::StorageClass::Function);
715 Register StackReg = StackTemp.getReg(0);
731 LLT DstTy =
MRI.getType(DstReg);
732 LLT SrcTy =
MRI.getType(SrcReg);
740 MI.eraseFromParent();
752 LLT DstTy =
MRI.getType(DstReg);
757 LLT SrcTy =
MRI.getType(SrcReg);
762 if (IdxVal < SrcTy.getNumElements()) {
767 for (
unsigned I = 0,
E = SrcTy.getNumElements();
I <
E; ++
I) {
774 Regs[IdxVal] = ValReg;
776 MI.eraseFromParent();
787 MIRBuilder.
buildStore(SrcReg, StackTemp, PtrInfo, VecAlign);
790 LLT PtrTy =
MRI.getType(StackTemp.getReg(0));
791 Register EltPtr =
MRI.createGenericVirtualRegister(PtrTy);
796 .
addUse(StackTemp.getReg(0))
802 MIRBuilder.
buildStore(ValReg, EltPtr, EltPtrInfo, EltAlign);
804 MIRBuilder.
buildLoad(DstReg, StackTemp, PtrInfo, VecAlign);
805 MI.eraseFromParent();
818 LLT SrcTy =
MRI.getType(SrcReg);
826 if (IdxVal < SrcTy.getNumElements()) {
827 LLT DstTy =
MRI.getType(DstReg);
830 for (
unsigned I = 0,
E = SrcTy.getNumElements();
I <
E; ++
I) {
841 MI.eraseFromParent();
852 MIRBuilder.
buildStore(SrcReg, StackTemp, PtrInfo, VecAlign);
855 LLT PtrTy =
MRI.getType(StackTemp.getReg(0));
856 Register EltPtr =
MRI.createGenericVirtualRegister(PtrTy);
861 .
addUse(StackTemp.getReg(0))
867 MIRBuilder.
buildLoad(DstReg, EltPtr, EltPtrInfo, EltAlign);
869 MI.eraseFromParent();
882 LLT DstTy =
MRI.getType(DstReg);
888 if (
MI.getNumOperands() == 2) {
898 for (
unsigned i = 2; i <
MI.getNumOperands(); ++i) {
903 MI.eraseFromParent();
911 switch (IntrinsicID) {
912 case Intrinsic::spv_bitcast:
914 case Intrinsic::spv_insertelt:
916 case Intrinsic::spv_extractelt:
918 case Intrinsic::spv_const_composite:
935 MI.eraseFromParent();
942bool SPIRVLegalizerInfo::legalizeIsFPClass(
945 auto [DstReg, DstTy, SrcReg, SrcTy] =
MI.getFirst2RegLLTs();
949 auto &MF = MIRBuilder.
getMF();
950 MachineRegisterInfo &
MRI = MF.getRegInfo();
954 if (DstTy.isVector())
956 SPIRVType *SPIRVDstTy = GR->getOrCreateSPIRVType(
957 LLVMDstTy, MIRBuilder, SPIRV::AccessQualifier::ReadWrite,
960 unsigned BitSize = SrcTy.getScalarSizeInBits();
965 if (SrcTy.isVector()) {
966 IntTy =
LLT::vector(SrcTy.getElementCount(), IntTy);
969 SPIRVType *SPIRVIntTy = GR->getOrCreateSPIRVType(
970 LLVMIntTy, MIRBuilder, SPIRV::AccessQualifier::ReadWrite,
974 LLT DstTyCopy = DstTy;
975 const auto assignSPIRVTy = [&](MachineInstrBuilder &&
MI) {
979 LLT MITy =
MRI.getType(
MI.getReg(0));
980 assert((MITy == IntTy || MITy == DstTyCopy) &&
981 "Unexpected LLT type while lowering G_IS_FPCLASS");
982 auto *SPVTy = MITy == IntTy ? SPIRVIntTy : SPIRVDstTy;
983 GR->assignSPIRVTypeToVReg(SPVTy,
MI.getReg(0), MF);
988 const auto buildSPIRVConstant = [&](LLT Ty,
auto &&
C) -> MachineInstrBuilder {
992 assert((Ty == IntTy || Ty == DstTyCopy) &&
993 "Unexpected LLT type while lowering constant for G_IS_FPCLASS");
994 SPIRVType *VecEltTy = GR->getOrCreateSPIRVType(
995 (Ty == IntTy ? LLVMIntTy : LLVMDstTy)->getScalarType(), MIRBuilder,
996 SPIRV::AccessQualifier::ReadWrite,
998 GR->assignSPIRVTypeToVReg(VecEltTy, ScalarC.getReg(0), MF);
1003 MIRBuilder.
buildCopy(DstReg, buildSPIRVConstant(DstTy, 0));
1004 MI.eraseFromParent();
1008 MIRBuilder.
buildCopy(DstReg, buildSPIRVConstant(DstTy, 1));
1009 MI.eraseFromParent();
1017 Register ResVReg =
MRI.createGenericVirtualRegister(IntTy);
1018 MRI.setRegClass(ResVReg, GR->getRegClass(SPIRVIntTy));
1019 GR->assignSPIRVTypeToVReg(SPIRVIntTy, ResVReg, Helper.
MIRBuilder.
getMF());
1020 auto AsInt = MIRBuilder.
buildInstr(SPIRV::OpBitcast)
1022 .
addUse(GR->getSPIRVTypeID(SPIRVIntTy))
1024 AsInt = assignSPIRVTy(std::move(AsInt));
1030 APInt ExpMask = Inf;
1036 auto SignBitC = buildSPIRVConstant(IntTy, SignBit);
1037 auto ValueMaskC = buildSPIRVConstant(IntTy, ValueMask);
1038 auto InfC = buildSPIRVConstant(IntTy, Inf);
1039 auto ExpMaskC = buildSPIRVConstant(IntTy, ExpMask);
1040 auto ZeroC = buildSPIRVConstant(IntTy, 0);
1042 auto Abs = assignSPIRVTy(MIRBuilder.
buildAnd(IntTy, AsInt, ValueMaskC));
1043 auto Sign = assignSPIRVTy(
1046 auto Res = buildSPIRVConstant(DstTy, 0);
1048 const auto appendToRes = [&](MachineInstrBuilder &&ToAppend) {
1049 Res = assignSPIRVTy(
1050 MIRBuilder.
buildOr(DstTyCopy, Res, assignSPIRVTy(std::move(ToAppend))));
1063 Mask &= ~fcPosFinite;
1067 DstTy, Abs, ExpMaskC));
1068 appendToRes(MIRBuilder.
buildAnd(DstTy, Cmp, Sign));
1069 Mask &= ~fcNegFinite;
1077 auto ExpBits = assignSPIRVTy(MIRBuilder.
buildAnd(IntTy, AsInt, ExpMaskC));
1080 Mask &= ~PartialCheck;
1089 else if (PartialCheck ==
fcZero)
1101 auto OneC = buildSPIRVConstant(IntTy, 1);
1102 auto VMinusOne = MIRBuilder.
buildSub(IntTy, V, OneC);
1103 auto SubnormalRes = assignSPIRVTy(
1105 buildSPIRVConstant(IntTy, AllOneMantissa)));
1107 SubnormalRes = MIRBuilder.
buildAnd(DstTy, SubnormalRes, Sign);
1108 appendToRes(std::move(SubnormalRes));
1115 else if (PartialCheck ==
fcInf)
1120 auto NegInfC = buildSPIRVConstant(IntTy, NegInf);
1127 auto InfWithQnanBitC =
1128 buildSPIRVConstant(IntTy, std::move(Inf) | QNaNBitMask);
1129 if (PartialCheck ==
fcNan) {
1133 }
else if (PartialCheck ==
fcQNan) {
1140 auto IsNan = assignSPIRVTy(
1142 auto IsNotQnan = assignSPIRVTy(MIRBuilder.
buildICmp(
1144 appendToRes(MIRBuilder.
buildAnd(DstTy, IsNan, IsNotQnan));
1151 APInt ExpLSB = ExpMask & ~(ExpMask.
shl(1));
1152 auto ExpMinusOne = assignSPIRVTy(
1153 MIRBuilder.
buildSub(IntTy, Abs, buildSPIRVConstant(IntTy, ExpLSB)));
1154 APInt MaxExpMinusOne = std::move(ExpMask) - ExpLSB;
1155 auto NormalRes = assignSPIRVTy(
1157 buildSPIRVConstant(IntTy, MaxExpMinusOne)));
1159 NormalRes = MIRBuilder.
buildAnd(DstTy, NormalRes, Sign);
1161 auto PosSign = assignSPIRVTy(MIRBuilder.
buildXor(
1162 DstTy, Sign, buildSPIRVConstant(DstTy, InversionMask)));
1163 NormalRes = MIRBuilder.
buildAnd(DstTy, NormalRes, PosSign);
1165 appendToRes(std::move(NormalRes));
1169 MI.eraseFromParent();
unsigned const MachineRegisterInfo * MRI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static void scalarize(Instruction *I, SmallVectorImpl< Instruction * > &Worklist)
Declares convenience wrapper classes for interpreting MachineInstr instances as specific generic oper...
This file declares the MachineIRBuilder class.
Promote Memory to Register
const SmallVectorImpl< MachineOperand > & Cond
static bool legalizeSpvInsertElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool needsVectorLegalization(const LLT &Ty, const SPIRVSubtarget &ST)
static Register convertPtrToInt(Register Reg, LLT ConvTy, SPIRVType *SpvType, LegalizerHelper &Helper, MachineRegisterInfo &MRI, SPIRVGlobalRegistry *GR)
static MachineInstrBuilder createStackTemporaryForVector(LegalizerHelper &Helper, SPIRVGlobalRegistry *GR, Register SrcReg, LLT SrcTy, MachinePointerInfo &PtrInfo, Align &VecAlign)
LegalityPredicate typeOfExtendedScalars(unsigned TypeIdx, bool IsExtendedInts)
static bool legalizeExtractVectorElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeStore(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeSpvExtractElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeSpvBitcast(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeSpvConstComposite(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeLoad(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
static bool legalizeInsertVectorElt(LegalizerHelper &Helper, MachineInstr &MI, SPIRVGlobalRegistry *GR)
APInt bitcastToAPInt() const
static APFloat getLargest(const fltSemantics &Sem, bool Negative=false)
Returns the largest finite number in the given semantics.
static APFloat getInf(const fltSemantics &Sem, bool Negative=false)
Factory for Positive and Negative Infinity.
static APInt getAllOnes(unsigned numBits)
Return an APInt of a specified width with all bits set.
static APInt getSignMask(unsigned BitWidth)
Get the SignMask for a specific bit width.
unsigned getActiveBits() const
Compute the number of active bits in the value.
static APInt getSignedMaxValue(unsigned numBits)
Gets maximum signed value of APInt for a specific bit width.
APInt shl(unsigned shiftAmt) const
Left-shift function.
static APInt getOneBitSet(unsigned numBits, unsigned BitNo)
Return an APInt with exactly one bit set in the result.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
static LLVM_ABI ArrayType * get(Type *ElementType, uint64_t NumElements)
This static method is the primary way to construct an ArrayType.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
@ ICMP_UGE
unsigned greater or equal
@ ICMP_UGT
unsigned greater than
@ ICMP_ULT
unsigned less than
static constexpr ElementCount getFixed(ScalarTy MinVal)
static LLVM_ABI IntegerType * get(LLVMContext &C, unsigned NumBits)
This static method is the primary way of constructing an IntegerType.
static constexpr LLT vector(ElementCount EC, unsigned ScalarSizeInBits)
Get a low-level vector of some number of elements and element width.
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr bool isValid() const
constexpr uint16_t getNumElements() const
Returns the number of elements in a vector LLT.
constexpr bool isVector() const
static constexpr LLT pointer(unsigned AddressSpace, unsigned SizeInBits)
Get a low-level pointer in the given address space.
constexpr LLT getElementType() const
Returns the vector's element type. Only valid for vector types.
constexpr unsigned getAddressSpace() const
static constexpr LLT fixed_vector(unsigned NumElements, unsigned ScalarSizeInBits)
Get a low-level fixed-width vector of some number of elements and element width.
constexpr bool isPointerOrPointerVector() const
constexpr bool isFixedVector() const
Returns true if the LLT is a fixed vector.
constexpr LLT getScalarType() const
constexpr TypeSize getSizeInBytes() const
Returns the total size of the type in bytes, i.e.
LLVM_ABI void computeTables()
Compute any ancillary tables needed to quickly decide how an operation should be handled.
LegalizeRuleSet & legalFor(std::initializer_list< LLT > Types)
The instruction is legal when type index 0 is any type in the given list.
LegalizeRuleSet & fewerElementsIf(LegalityPredicate Predicate, LegalizeMutation Mutation)
Remove elements to reach the type selected by the mutation if the predicate is true.
LegalizeRuleSet & moreElementsToNextPow2(unsigned TypeIdx)
Add more elements to the vector to reach the next power of two.
LegalizeRuleSet & lower()
The instruction is lowered.
LegalizeRuleSet & scalarizeIf(LegalityPredicate Predicate, unsigned TypeIdx)
LegalizeRuleSet & lowerIf(LegalityPredicate Predicate)
The instruction is lowered if predicate is true.
LegalizeRuleSet & custom()
Unconditionally custom lower.
LegalizeRuleSet & unsupportedIf(LegalityPredicate Predicate)
LegalizeRuleSet & alwaysLegal()
LegalizeRuleSet & customIf(LegalityPredicate Predicate)
LegalizeRuleSet & scalarize(unsigned TypeIdx)
LegalizeRuleSet & legalForCartesianProduct(std::initializer_list< LLT > Types)
The instruction is legal when type indexes 0 and 1 are both in the given list.
LegalizeRuleSet & legalIf(LegalityPredicate Predicate)
The instruction is legal if predicate is true.
LegalizeRuleSet & customFor(std::initializer_list< LLT > Types)
LLVM_ABI MachineInstrBuilder createStackTemporary(TypeSize Bytes, Align Alignment, MachinePointerInfo &PtrInfo)
Create a stack temporary based on the size in bytes and the alignment.
MachineIRBuilder & MIRBuilder
Expose MIRBuilder so clients can set their own RecordInsertInstruction functions.
LLVM_ABI Align getStackTemporaryAlignment(LLT Type, Align MinAlign=Align()) const
Return the alignment to use for a stack temporary object with the given type.
LegalizeRuleSet & getActionDefinitionsBuilder(unsigned Opcode)
Get the action definition builder for the given opcode.
const LegacyLegalizerInfo & getLegacyLegalizerInfo() const
Helper class to build MachineInstr.
LLVMContext & getContext() const
MachineInstrBuilder buildUnmerge(ArrayRef< LLT > Res, const SrcOp &Op)
Build and insert Res0, ... = G_UNMERGE_VALUES Op.
MachineInstrBuilder buildAnd(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1)
Build and insert Res = G_AND Op0, Op1.
MachineInstrBuilder buildICmp(CmpInst::Predicate Pred, const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1, std::optional< unsigned > Flags=std::nullopt)
Build and insert a Res = G_ICMP Pred, Op0, Op1.
MachineInstrBuilder buildSub(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_SUB Op0, Op1.
MachineInstrBuilder buildIntrinsic(Intrinsic::ID ID, ArrayRef< Register > Res, bool HasSideEffects, bool isConvergent)
Build and insert a G_INTRINSIC instruction.
MachineInstrBuilder buildSplatBuildVector(const DstOp &Res, const SrcOp &Src)
Build and insert Res = G_BUILD_VECTOR with Src replicated to fill the number of elements.
MachineInstrBuilder buildBuildVector(const DstOp &Res, ArrayRef< Register > Ops)
Build and insert Res = G_BUILD_VECTOR Op0, ...
MachineInstrBuilder buildLoad(const DstOp &Res, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert Res = G_LOAD Addr, MMO.
MachineInstrBuilder buildStore(const SrcOp &Val, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert G_STORE Val, Addr, MMO.
MachineInstrBuilder buildInstr(unsigned Opcode)
Build and insert <empty> = Opcode <empty>.
MachineFunction & getMF()
Getter for the function we currently build.
MachineInstrBuilder buildBitcast(const DstOp &Dst, const SrcOp &Src)
Build and insert Dst = G_BITCAST Src.
MachineRegisterInfo * getMRI()
Getter for MRI.
MachineInstrBuilder buildOr(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_OR Op0, Op1.
MachineInstrBuilder buildCopy(const DstOp &Res, const SrcOp &Op)
Build and insert Res = COPY Op.
MachineInstrBuilder buildXor(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1)
Build and insert Res = G_XOR Op0, Op1.
virtual MachineInstrBuilder buildConstant(const DstOp &Res, const ConstantInt &Val)
Build and insert Res = G_CONSTANT Val.
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addDef(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register definition operand.
Representation of each machine instruction.
A description of a memory reference used in the backend.
const MachinePointerInfo & getPointerInfo() const
LLVM_ABI Align getAlign() const
Return the minimum known alignment in bytes of the actual memory reference.
MachineOperand class - Representation of each machine instruction operand.
Register getReg() const
getReg - Returns the register number.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
Wrapper class representing virtual and physical registers.
SPIRVType * getSPIRVTypeForVReg(Register VReg, const MachineFunction *MF=nullptr) const
void assignSPIRVTypeToVReg(SPIRVType *Type, Register VReg, const MachineFunction &MF)
const Type * getTypeForSPIRVType(const SPIRVType *Ty) const
SPIRVType * getOrCreateSPIRVType(const Type *Type, MachineInstr &I, SPIRV::AccessQualifier::AccessQualifier AQ, bool EmitIR)
SPIRVType * getOrCreateSPIRVPointerType(const Type *BaseType, MachineIRBuilder &MIRBuilder, SPIRV::StorageClass::StorageClass SC)
SPIRVType * getScalarOrVectorComponentType(Register VReg) const
const TargetRegisterClass * getRegClass(SPIRVType *SpvType) const
LLT getRegType(SPIRVType *SpvType) const
SPIRVLegalizerInfo(const SPIRVSubtarget &ST)
bool legalizeCustom(LegalizerHelper &Helper, MachineInstr &MI, LostDebugLocObserver &LocObserver) const override
Called for instructions with the Custom LegalizationAction.
bool legalizeIntrinsic(LegalizerHelper &Helper, MachineInstr &MI) const override
SPIRVGlobalRegistry * getSPIRVGlobalRegistry() const
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
The instances of the Type class are immutable: once they are created, they are never changed.
static LLVM_ABI VectorType * get(Type *ElementType, ElementCount EC)
This static method is the primary way to construct an VectorType.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
LLVM_ABI LegalityPredicate isScalar(unsigned TypeIdx)
True iff the specified type index is a scalar.
LLVM_ABI LegalityPredicate numElementsNotPow2(unsigned TypeIdx)
True iff the specified type index is a vector whose element count is not a power of 2.
LLVM_ABI LegalityPredicate vectorElementCountIsLessThanOrEqualTo(unsigned TypeIdx, unsigned Size)
True iff the specified type index is a vector with a number of elements that's less than or equal to ...
LLVM_ABI LegalityPredicate typeInSet(unsigned TypeIdx, std::initializer_list< LLT > TypesInit)
True iff the given type index is one of the specified types.
LLVM_ABI LegalityPredicate vectorElementCountIsGreaterThan(unsigned TypeIdx, unsigned Size)
True iff the specified type index is a vector with a number of elements that's greater than the given...
Predicate any(Predicate P0, Predicate P1)
True iff P0 or P1 are true.
LegalityPredicate typeIsNot(unsigned TypeIdx, LLT Type)
True iff the given type index is not the specified type.
Predicate all(Predicate P0, Predicate P1)
True iff P0 and P1 are true.
LLVM_ABI LegalityPredicate typeIs(unsigned TypeIdx, LLT TypesInit)
True iff the given type index is the specified type.
LLVM_ABI LegalizeMutation changeElementCountTo(unsigned TypeIdx, unsigned FromTypeIdx)
Keep the same scalar or element type as TypeIdx, but take the number of elements from FromTypeIdx.
LLVM_ABI LegalizeMutation changeElementSizeTo(unsigned TypeIdx, unsigned FromTypeIdx)
Change the scalar size or element size to have the same scalar size as type index FromIndex.
Invariant opcodes: All instruction sets have these as their low opcodes.
This is an optimization pass for GlobalISel generic memory operations.
LLVM_ABI const llvm::fltSemantics & getFltSemanticForLLT(LLT Ty)
Get the appropriate floating point arithmetic semantic based on the bit size of the given scalar LLT.
std::function< bool(const LegalityQuery &)> LegalityPredicate
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
FPClassTest
Floating-point class tests, supported by 'is_fpclass' intrinsic.
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
const MachineInstr SPIRVType
const std::set< unsigned > & getTypeFoldingSupportedOpcodes()
int64_t foldImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
This struct is a compact representation of a valid (non-zero power of two) alignment.
The LegalityQuery object bundles together all the information that's needed to decide whether a given...
This class contains a discriminated union of information about pointers in memory operands,...
MachinePointerInfo getWithOffset(int64_t O) const