2197 B.setInstrAndDebugLoc(
MI);
2198 unsigned Opc =
MI.getOpcode();
2201 case AMDGPU::G_CONSTANT:
2202 case AMDGPU::G_IMPLICIT_DEF: {
2204 LLT DstTy =
MRI.getType(DstReg);
2210 if (DstBank == &AMDGPU::VCCRegBank)
2213 if (DefRegs.
empty())
2216 B.setInsertPt(*
MI.getParent(), ++
MI.getIterator());
2219 LLVMContext &Ctx =
B.getMF().getFunction().getContext();
2221 MI.getOperand(0).setReg(NewDstReg);
2222 if (
Opc != AMDGPU::G_IMPLICIT_DEF) {
2223 uint64_t ConstVal =
MI.getOperand(1).getCImm()->getZExtValue();
2224 MI.getOperand(1).setCImm(
2228 MRI.setRegBank(NewDstReg, *DstBank);
2229 B.buildTrunc(DefRegs[0], NewDstReg);
2232 case AMDGPU::G_PHI: {
2234 LLT DstTy =
MRI.getType(DstReg);
2241 if (DstBank == &AMDGPU::VCCRegBank) {
2248 for (
unsigned I = 1, E =
MI.getNumOperands();
I != E;
I += 2) {
2252 if (SrcBank != &AMDGPU::VCCRegBank) {
2257 MRI.setRegBank(Copy.getReg(0), AMDGPU::VCCRegBank);
2258 MI.getOperand(
I).setReg(Copy.getReg(0));
2269 ApplyRegBankMapping ApplyBank(
B, *
this,
MRI, DstBank);
2270 B.setInsertPt(
B.getMBB(),
MI);
2278 case AMDGPU::G_FCMP:
2282 case AMDGPU::G_ICMP:
2283 case AMDGPU::G_UADDO:
2284 case AMDGPU::G_USUBO:
2285 case AMDGPU::G_UADDE:
2286 case AMDGPU::G_SADDE:
2287 case AMDGPU::G_USUBE:
2288 case AMDGPU::G_SSUBE: {
2289 unsigned BoolDstOp =
2290 (
Opc == AMDGPU::G_ICMP ||
Opc == AMDGPU::G_FCMP) ? 0 : 1;
2291 Register DstReg =
MI.getOperand(BoolDstOp).getReg();
2295 if (DstBank != &AMDGPU::SGPRRegBank)
2298 const bool HasCarryIn =
MI.getNumOperands() == 5;
2304 MRI.setRegBank(NewDstReg, AMDGPU::SGPRRegBank);
2305 MI.getOperand(BoolDstOp).setReg(NewDstReg);
2309 MRI.setRegBank(NewSrcReg, AMDGPU::SGPRRegBank);
2310 B.buildZExt(NewSrcReg,
MI.getOperand(4).getReg());
2311 MI.getOperand(4).setReg(NewSrcReg);
2315 B.setInsertPt(*
MBB, std::next(
MI.getIterator()));
2320 if (DefRegs.
empty())
2322 B.buildTrunc(DefRegs[0], NewDstReg);
2325 case AMDGPU::G_SELECT: {
2327 LLT DstTy =
MRI.getType(DstReg);
2330 if (CondRegs.
empty())
2337 if (CondBank == &AMDGPU::SGPRRegBank) {
2340 MRI.setRegBank(NewCondReg, AMDGPU::SGPRRegBank);
2342 MI.getOperand(1).setReg(NewCondReg);
2343 B.buildZExt(NewCondReg, CondRegs[0]);
2356 if (DefRegs.
empty()) {
2361 if (Src1Regs.
empty())
2367 if (Src2Regs.
empty())
2374 auto Flags =
MI.getFlags();
2375 B.buildSelect(DefRegs[0], CondRegs[0], Src1Regs[0], Src2Regs[0], Flags);
2376 B.buildSelect(DefRegs[1], CondRegs[0], Src1Regs[1], Src2Regs[1], Flags);
2378 MRI.setRegBank(DstReg, AMDGPU::VGPRRegBank);
2379 MI.eraseFromParent();
2382 case AMDGPU::G_BRCOND: {
2383 Register CondReg =
MI.getOperand(0).getReg();
2388 if (CondBank == &AMDGPU::SGPRRegBank) {
2391 MRI.setRegBank(NewCondReg, AMDGPU::SGPRRegBank);
2393 MI.getOperand(0).setReg(NewCondReg);
2394 B.buildZExt(NewCondReg, CondReg);
2402 case AMDGPU::G_XOR: {
2406 LLT DstTy =
MRI.getType(DstReg);
2412 if (DstBank == &AMDGPU::VCCRegBank)
2416 ApplyRegBankMapping ApplyBank(
B, *
this,
MRI, DstBank);
2425 if (DstTy.
getSizeInBits() == 16 && DstBank == &AMDGPU::SGPRRegBank) {
2429 ApplyRegBankMapping ApplySALU(
B, *
this,
MRI, &AMDGPU::SGPRRegBank);
2434 if (
MI.getOpcode() == AMDGPU::G_XOR &&
2455 if (DefRegs.
empty()) {
2462 (Src0Regs.
empty() || Src0Regs.
size() == 2));
2468 if (Src0Regs.
empty())
2473 if (Src1Regs.
empty())
2480 auto Flags =
MI.getFlags();
2481 B.buildInstr(
Opc, {DefRegs[0]}, {Src0Regs[0], Src1Regs[0]}, Flags);
2482 B.buildInstr(
Opc, {DefRegs[1]}, {Src0Regs[1], Src1Regs[1]}, Flags);
2484 MRI.setRegBank(DstReg, AMDGPU::VGPRRegBank);
2485 MI.eraseFromParent();
2488 case AMDGPU::G_ABS: {
2494 if (SrcBank && SrcBank == &AMDGPU::VGPRRegBank) {
2496 ApplyRegBankMapping Apply(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2509 case AMDGPU::G_LSHR:
2510 case AMDGPU::G_ASHR:
2511 case AMDGPU::G_SMIN:
2512 case AMDGPU::G_SMAX:
2513 case AMDGPU::G_UMIN:
2514 case AMDGPU::G_UMAX: {
2516 LLT DstTy =
MRI.getType(DstReg);
2521 if (!
Subtarget.hasVectorMulU64() &&
Opc == AMDGPU::G_MUL &&
2534 if (DstBank == &AMDGPU::VGPRRegBank)
2540 ApplyRegBankMapping ApplySALU(
B, *
this,
MRI, &AMDGPU::SGPRRegBank);
2545 std::tie(WideSrcLo, WideSrcHi) =
2547 auto Lo =
B.buildInstr(AMDGPU::G_ABS, {
S32}, {WideSrcLo});
2548 auto Hi =
B.buildInstr(AMDGPU::G_ABS, {
S32}, {WideSrcHi});
2549 B.buildBuildVectorTrunc(DstReg, {
Lo.getReg(0),
Hi.getReg(0)});
2550 MI.eraseFromParent();
2559 std::tie(WideSrc0Lo, WideSrc0Hi)
2561 std::tie(WideSrc1Lo, WideSrc1Hi)
2563 auto Lo =
B.buildInstr(
MI.getOpcode(), {S32}, {WideSrc0Lo, WideSrc1Lo});
2564 auto Hi =
B.buildInstr(
MI.getOpcode(), {S32}, {WideSrc0Hi, WideSrc1Hi});
2565 B.buildBuildVectorTrunc(DstReg, {
Lo.getReg(0),
Hi.getReg(0)});
2566 MI.eraseFromParent();
2574 if (
Opc == AMDGPU::G_SHL ||
Opc == AMDGPU::G_LSHR ||
2575 Opc == AMDGPU::G_ASHR) {
2576 B.setInsertPt(*
MBB,
MI.getIterator());
2584 case AMDGPU::G_AMDGPU_S_MUL_I64_I32:
2585 case AMDGPU::G_AMDGPU_S_MUL_U64_U32: {
2599 Register SrcReg0 =
MI.getOperand(1).getReg();
2600 Register SrcReg1 =
MI.getOperand(2).getReg();
2603 assert(
MRI.getType(DstReg) ==
S64 &&
"This is a special case for s_mul_u64 "
2604 "that handles only 64-bit operands.");
2610 if (DstBank == &AMDGPU::SGPRRegBank) {
2611 MI.setDesc(
TII->get(AMDGPU::S_MUL_U64));
2612 MRI.setRegClass(DstReg, &AMDGPU::SGPR_64RegClass);
2613 MRI.setRegClass(SrcReg0, &AMDGPU::SGPR_64RegClass);
2614 MRI.setRegClass(SrcReg1, &AMDGPU::SGPR_64RegClass);
2620 assert(
MRI.getRegBankOrNull(DstReg) == &AMDGPU::VGPRRegBank &&
2621 "The destination operand should be in vector registers.");
2624 Register Op0L =
MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass);
2625 MRI.setRegClass(Op0L, &AMDGPU::VGPR_32RegClass);
2627 B.buildTrunc(Op0L, SrcReg0);
2630 Register Op1L =
MRI.createVirtualRegister(&AMDGPU::VGPR_32RegClass);
2631 MRI.setRegClass(Op1L, &AMDGPU::VGPR_32RegClass);
2633 B.buildTrunc(Op1L, SrcReg1);
2635 unsigned NewOpc =
Opc == AMDGPU::G_AMDGPU_S_MUL_U64_U32
2636 ? AMDGPU::G_AMDGPU_MAD_U64_U32
2637 : AMDGPU::G_AMDGPU_MAD_I64_I32;
2641 MRI.setRegClass(Zero64, &AMDGPU::VReg_64RegClass);
2642 Register CarryOut =
MRI.createVirtualRegister(&AMDGPU::VReg_64RegClass);
2643 MRI.setRegClass(CarryOut, &AMDGPU::VReg_64RegClass);
2644 B.buildInstr(NewOpc, {DstReg, CarryOut}, {Op0L, Op1L, Zero64});
2645 MI.eraseFromParent();
2648 case AMDGPU::G_SEXT_INREG: {
2650 if (SrcRegs.
empty())
2654 ApplyRegBankMapping O(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2661 int Amt =
MI.getOperand(2).getImm();
2667 B.buildFreeze(DstRegs[0], SrcRegs[0]);
2669 auto Freeze =
B.buildFreeze(
S32, SrcRegs[0]);
2671 B.buildSExtInReg(DstRegs[0], Freeze, Amt);
2674 B.buildAShr(DstRegs[1], DstRegs[0],
B.buildConstant(
S32, 31));
2678 B.buildCopy(DstRegs[0], SrcRegs[0]);
2679 B.buildSExtInReg(DstRegs[1], DstRegs[0], Amt - 32);
2683 MRI.setRegBank(DstReg, AMDGPU::VGPRRegBank);
2684 MI.eraseFromParent();
2687 case AMDGPU::G_CTPOP:
2688 case AMDGPU::G_BITREVERSE: {
2691 if (DstBank == &AMDGPU::SGPRRegBank)
2696 LLT Ty =
MRI.getType(SrcReg);
2700 ApplyRegBankMapping ApplyVALU(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2709 case AMDGPU::G_AMDGPU_FFBH_U32:
2710 case AMDGPU::G_AMDGPU_FFBL_B32:
2711 case AMDGPU::G_CTLZ_ZERO_UNDEF:
2712 case AMDGPU::G_CTTZ_ZERO_UNDEF: {
2715 if (DstBank == &AMDGPU::SGPRRegBank)
2720 LLT Ty =
MRI.getType(SrcReg);
2730 ApplyRegBankMapping ApplyVALU(
B, *
this,
MRI, &AMDGPU::VGPRRegBank);
2732 unsigned NewOpc =
Opc == AMDGPU::G_CTLZ_ZERO_UNDEF
2733 ? (
unsigned)AMDGPU::G_AMDGPU_FFBH_U32
2734 :
Opc == AMDGPU::G_CTTZ_ZERO_UNDEF
2735 ? (
unsigned)AMDGPU::G_AMDGPU_FFBL_B32
2737 unsigned Idx = NewOpc == AMDGPU::G_AMDGPU_FFBH_U32;
2738 auto X =
B.buildInstr(NewOpc, {
S32}, {SrcRegs[Idx]});
2739 auto Y =
B.buildInstr(NewOpc, {
S32}, {SrcRegs[Idx ^ 1]});
2741 Opc == AMDGPU::G_CTLZ_ZERO_UNDEF ||
Opc == AMDGPU::G_CTTZ_ZERO_UNDEF
2743 : AMDGPU::G_UADDSAT;
2744 Y =
B.buildInstr(AddOpc, {
S32}, {
Y,
B.buildConstant(
S32, 32)});
2746 B.buildUMin(DstReg,
X,
Y);
2747 MI.eraseFromParent();
2750 case AMDGPU::G_SEXT:
2751 case AMDGPU::G_ZEXT:
2752 case AMDGPU::G_ANYEXT: {
2754 LLT SrcTy =
MRI.getType(SrcReg);
2755 const bool Signed =
Opc == AMDGPU::G_SEXT;
2763 LLT DstTy =
MRI.getType(DstReg);
2765 SrcBank != &AMDGPU::SGPRRegBank &&
2766 SrcBank != &AMDGPU::VCCRegBank &&
2770 SrcTy.getSizeInBits() <= 32) {
2776 B.buildSExtOrTrunc(DefRegs[0], SrcReg);
2777 }
else if (
Opc == AMDGPU::G_ZEXT) {
2778 B.buildZExtOrTrunc(DefRegs[0], SrcReg);
2780 B.buildAnyExtOrTrunc(DefRegs[0], SrcReg);
2784 MRI.setRegBank(DstReg, *SrcBank);
2785 MI.eraseFromParent();
2795 if (SrcBank == &AMDGPU::VCCRegBank) {
2802 const bool UseSel64 = DstSize > 32 &&
2803 SrcBank->
getID() == AMDGPU::SGPRRegBankID;
2807 auto True =
B.buildConstant(SelType,
Signed ? -1 : 1);
2808 auto False =
B.buildConstant(SelType, 0);
2810 MRI.setRegBank(True.getReg(0), *DstBank);
2811 MRI.setRegBank(False.getReg(0), *DstBank);
2812 MRI.setRegBank(DstReg, *DstBank);
2815 B.buildSelect(DefRegs[0], SrcReg, True, False);
2817 }
else if (DstSize < 32) {
2818 auto Sel =
B.buildSelect(SelType, SrcReg, True, False);
2819 MRI.setRegBank(Sel.getReg(0), *DstBank);
2820 B.buildTrunc(DstReg, Sel);
2822 B.buildSelect(DstReg, SrcReg, True, False);
2825 MI.eraseFromParent();
2831 case AMDGPU::G_EXTRACT_VECTOR_ELT: {
2840 LLT DstTy =
MRI.getType(DstReg);
2841 LLT SrcTy =
MRI.getType(SrcReg);
2843 if (foldExtractEltToCmpSelect(
B,
MI, OpdMapper))
2855 unsigned ConstOffset;
2856 std::tie(BaseIdxReg, ConstOffset) =
2863 bool ShouldMoveIndexIntoLoop = IdxBank != &AMDGPU::SGPRRegBank &&
2865 ConstOffset < SrcTy.getNumElements();
2868 if (ShouldMoveIndexIntoLoop)
2869 MI.getOperand(2).setReg(BaseIdxReg);
2875 const bool NeedCopyToVGPR = DstBank == &AMDGPU::VGPRRegBank &&
2876 SrcBank == &AMDGPU::SGPRRegBank;
2877 if (DstRegs.
empty()) {
2882 if (NeedCopyToVGPR) {
2884 Register TmpReg =
MRI.createGenericVirtualRegister(DstTy);
2885 MRI.setRegBank(TmpReg, AMDGPU::SGPRRegBank);
2886 MI.getOperand(0).setReg(TmpReg);
2887 B.setInsertPt(*
MI.getParent(), ++
MI.getIterator());
2894 if (ShouldMoveIndexIntoLoop)
2904 auto CastSrc =
B.buildBitcast(Vec32, SrcReg);
2905 auto One =
B.buildConstant(
S32, 1);
2916 auto IdxLo =
B.buildShl(
S32, BaseIdxReg, One);
2917 auto IdxHi =
B.buildAdd(
S32, IdxLo, One);
2919 auto Extract0 =
B.buildExtractVectorElement(DstRegs[0], CastSrc, IdxLo);
2920 auto Extract1 =
B.buildExtractVectorElement(DstRegs[1], CastSrc, IdxHi);
2922 MRI.setRegBank(DstReg, *DstBank);
2923 MRI.setRegBank(CastSrc.getReg(0), *SrcBank);
2924 MRI.setRegBank(One.getReg(0), AMDGPU::SGPRRegBank);
2925 MRI.setRegBank(IdxLo.getReg(0), AMDGPU::SGPRRegBank);
2926 MRI.setRegBank(IdxHi.getReg(0), AMDGPU::SGPRRegBank);
2930 MI.eraseFromParent();
2936 B.setInstr(*Span.
begin());
2937 MI.eraseFromParent();
2941 if (NeedCopyToVGPR) {
2945 MRI.setRegBank(TmpReg0, AMDGPU::SGPRRegBank);
2946 MRI.setRegBank(TmpReg1, AMDGPU::SGPRRegBank);
2948 Extract0->getOperand(0).setReg(TmpReg0);
2949 Extract1->getOperand(0).setReg(TmpReg1);
2957 if (ShouldMoveIndexIntoLoop)
2962 case AMDGPU::G_INSERT_VECTOR_ELT: {
2966 LLT VecTy =
MRI.getType(DstReg);
2972 MRI.setType(
MI.getOperand(1).getReg(), VecTy);
2974 if (foldInsertEltToCmpSelect(
B,
MI, OpdMapper))
2982 LLT InsTy =
MRI.getType(InsReg);
2986 unsigned ConstOffset;
2987 std::tie(BaseIdxReg, ConstOffset) =
2994 bool ShouldMoveIndexIntoLoop = IdxBank != &AMDGPU::SGPRRegBank &&
2999 if (ShouldMoveIndexIntoLoop)
3000 MI.getOperand(3).setReg(BaseIdxReg);
3003 if (InsRegs.
empty()) {
3007 if (ShouldMoveIndexIntoLoop) {
3019 auto CastSrc =
B.buildBitcast(Vec32, SrcReg);
3020 auto One =
B.buildConstant(
S32, 1);
3029 auto IdxLo =
B.buildShl(
S32, BaseIdxReg, One);
3030 auto IdxHi =
B.buildAdd(
S32, IdxLo, One);
3032 auto InsLo =
B.buildInsertVectorElement(Vec32, CastSrc, InsRegs[0], IdxLo);
3033 auto InsHi =
B.buildInsertVectorElement(Vec32, InsLo, InsRegs[1], IdxHi);
3042 MRI.setRegBank(InsReg, *InsSrcBank);
3043 MRI.setRegBank(CastSrc.getReg(0), *SrcBank);
3044 MRI.setRegBank(InsLo.getReg(0), *DstBank);
3045 MRI.setRegBank(InsHi.getReg(0), *DstBank);
3046 MRI.setRegBank(One.getReg(0), AMDGPU::SGPRRegBank);
3047 MRI.setRegBank(IdxLo.getReg(0), AMDGPU::SGPRRegBank);
3048 MRI.setRegBank(IdxHi.getReg(0), AMDGPU::SGPRRegBank);
3053 B.setInsertPt(
B.getMBB(),
MI);
3054 B.buildBitcast(DstReg, InsHi);
3055 MI.eraseFromParent();
3059 B.setInstr(*Span.
begin());
3060 MI.eraseFromParent();
3071 B.buildBitcast(DstReg, InsHi);
3074 if (ShouldMoveIndexIntoLoop)
3079 case AMDGPU::G_AMDGPU_BUFFER_LOAD:
3080 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT:
3081 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT:
3082 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE:
3083 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE:
3084 case AMDGPU::G_AMDGPU_BUFFER_LOAD_TFE:
3085 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT_TFE:
3086 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT_TFE:
3087 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE_TFE:
3088 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE_TFE:
3089 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT:
3090 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_TFE:
3091 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_D16:
3092 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT:
3093 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT_D16:
3094 case AMDGPU::G_AMDGPU_BUFFER_STORE:
3095 case AMDGPU::G_AMDGPU_BUFFER_STORE_BYTE:
3096 case AMDGPU::G_AMDGPU_BUFFER_STORE_SHORT:
3097 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT:
3098 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT_D16:
3099 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT:
3100 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT_D16: {
3105 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SWAP:
3106 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_ADD:
3107 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB:
3108 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMIN:
3109 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMIN:
3110 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMAX:
3111 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMAX:
3112 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_AND:
3113 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_OR:
3114 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_XOR:
3115 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_INC:
3116 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_DEC:
3117 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB_CLAMP_U32:
3118 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_COND_SUB_U32:
3119 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FADD:
3120 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMIN:
3121 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMAX: {
3126 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_CMPSWAP: {
3131 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD:
3132 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_UBYTE:
3133 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SBYTE:
3134 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_USHORT:
3135 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SSHORT: {
3139 case AMDGPU::G_AMDGPU_S_BUFFER_PREFETCH:
3143 case AMDGPU::G_INTRINSIC:
3144 case AMDGPU::G_INTRINSIC_CONVERGENT: {
3146 case Intrinsic::amdgcn_readlane: {
3157 case Intrinsic::amdgcn_writelane: {
3167 case Intrinsic::amdgcn_interp_p1:
3168 case Intrinsic::amdgcn_interp_p2:
3169 case Intrinsic::amdgcn_interp_mov:
3170 case Intrinsic::amdgcn_interp_p1_f16:
3171 case Intrinsic::amdgcn_interp_p2_f16:
3172 case Intrinsic::amdgcn_lds_param_load: {
3180 case Intrinsic::amdgcn_interp_inreg_p10:
3181 case Intrinsic::amdgcn_interp_inreg_p2:
3182 case Intrinsic::amdgcn_interp_inreg_p10_f16:
3183 case Intrinsic::amdgcn_interp_inreg_p2_f16:
3184 case Intrinsic::amdgcn_interp_p10_rtz_f16:
3185 case Intrinsic::amdgcn_interp_p2_rtz_f16:
3186 case Intrinsic::amdgcn_permlane16_swap:
3187 case Intrinsic::amdgcn_permlane32_swap:
3190 case Intrinsic::amdgcn_permlane16:
3191 case Intrinsic::amdgcn_permlanex16: {
3199 case Intrinsic::amdgcn_permlane_bcast:
3200 case Intrinsic::amdgcn_permlane_up:
3201 case Intrinsic::amdgcn_permlane_down:
3202 case Intrinsic::amdgcn_permlane_xor:
3207 case Intrinsic::amdgcn_permlane_idx_gen: {
3211 case Intrinsic::amdgcn_sbfe:
3214 case Intrinsic::amdgcn_ubfe:
3217 case Intrinsic::amdgcn_inverse_ballot:
3218 case Intrinsic::amdgcn_s_bitreplicate:
3219 case Intrinsic::amdgcn_s_quadmask:
3220 case Intrinsic::amdgcn_s_wqm:
3224 case Intrinsic::amdgcn_ballot:
3230 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD:
3231 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_D16:
3232 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_NORET:
3233 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE:
3234 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE_D16: {
3244 case AMDGPU::G_AMDGPU_BVH_INTERSECT_RAY:
3245 case AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY:
3246 case AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY: {
3248 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY ||
3249 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY;
3250 unsigned NumMods = IsDualOrBVH8 ? 0 : 1;
3251 unsigned LastRegOpIdx =
MI.getNumExplicitOperands() - 1 - NumMods;
3256 case AMDGPU::G_INTRINSIC_W_SIDE_EFFECTS:
3257 case AMDGPU::G_INTRINSIC_CONVERGENT_W_SIDE_EFFECTS: {
3260 case Intrinsic::amdgcn_ds_ordered_add:
3261 case Intrinsic::amdgcn_ds_ordered_swap: {
3268 case Intrinsic::amdgcn_ds_gws_init:
3269 case Intrinsic::amdgcn_ds_gws_barrier:
3270 case Intrinsic::amdgcn_ds_gws_sema_br: {
3276 case Intrinsic::amdgcn_ds_gws_sema_v:
3277 case Intrinsic::amdgcn_ds_gws_sema_p:
3278 case Intrinsic::amdgcn_ds_gws_sema_release_all: {
3283 case Intrinsic::amdgcn_ds_append:
3284 case Intrinsic::amdgcn_ds_consume: {
3288 case Intrinsic::amdgcn_s_alloc_vgpr:
3291 case Intrinsic::amdgcn_s_sendmsg:
3292 case Intrinsic::amdgcn_s_sendmsghalt: {
3297 case Intrinsic::amdgcn_s_setreg: {
3301 case Intrinsic::amdgcn_s_ttracedata:
3304 case Intrinsic::amdgcn_raw_buffer_load_lds:
3305 case Intrinsic::amdgcn_raw_buffer_load_async_lds:
3306 case Intrinsic::amdgcn_raw_ptr_buffer_load_lds:
3307 case Intrinsic::amdgcn_raw_ptr_buffer_load_async_lds: {
3314 case Intrinsic::amdgcn_struct_buffer_load_lds:
3315 case Intrinsic::amdgcn_struct_buffer_load_async_lds:
3316 case Intrinsic::amdgcn_struct_ptr_buffer_load_lds:
3317 case Intrinsic::amdgcn_struct_ptr_buffer_load_async_lds: {
3324 case Intrinsic::amdgcn_cluster_load_async_to_lds_b8:
3325 case Intrinsic::amdgcn_cluster_load_async_to_lds_b32:
3326 case Intrinsic::amdgcn_cluster_load_async_to_lds_b64:
3327 case Intrinsic::amdgcn_cluster_load_async_to_lds_b128: {
3332 case Intrinsic::amdgcn_load_to_lds:
3333 case Intrinsic::amdgcn_load_async_to_lds:
3334 case Intrinsic::amdgcn_global_load_lds:
3335 case Intrinsic::amdgcn_global_load_async_lds: {
3340 case Intrinsic::amdgcn_lds_direct_load: {
3346 case Intrinsic::amdgcn_exp_row:
3350 case Intrinsic::amdgcn_cluster_load_b32:
3351 case Intrinsic::amdgcn_cluster_load_b64:
3352 case Intrinsic::amdgcn_cluster_load_b128: {
3357 case Intrinsic::amdgcn_s_sleep_var:
3361 case Intrinsic::amdgcn_s_barrier_join:
3362 case Intrinsic::amdgcn_s_wakeup_barrier:
3365 case Intrinsic::amdgcn_s_barrier_init:
3366 case Intrinsic::amdgcn_s_barrier_signal_var:
3370 case Intrinsic::amdgcn_s_get_barrier_state:
3371 case Intrinsic::amdgcn_s_get_named_barrier_state: {
3375 case Intrinsic::amdgcn_s_prefetch_data: {
3377 unsigned AS =
MRI.getType(PtrReg).getAddressSpace();
3382 MI.eraseFromParent();
3385 case Intrinsic::amdgcn_tensor_load_to_lds:
3386 case Intrinsic::amdgcn_tensor_store_from_lds: {
3400 if (RSrcIntrin->IsImage) {
3411 case AMDGPU::G_SI_CALL: {
3422 unsigned FrameSetupOpcode = AMDGPU::ADJCALLSTACKUP;
3423 unsigned FrameDestroyOpcode = AMDGPU::ADJCALLSTACKDOWN;
3429 unsigned NonCopyInstrsLen = 0;
3435 while (Start->getOpcode() != FrameSetupOpcode) {
3437 bool IsCopy =
false;
3438 if (Start->getOpcode() == AMDGPU::COPY) {
3439 auto &Dst = Start->getOperand(0);
3442 if (Reg.isPhysical() &&
MI.readsRegister(Reg,
TRI)) {
3447 auto &Src = Start->getOperand(1);
3450 IsCopy = Info->getScratchRSrcReg() == Reg;
3458 NonCopyInstrsLen = NonCopyInstrs.
size();
3463 NonCopyInstrs.
resize(NonCopyInstrsLen);
3465 for (
auto *NonCopy :
reverse(NonCopyInstrs)) {
3466 MBB->splice(LastCopy,
MBB, NonCopy->getIterator());
3471 NonCopyInstrs.
clear();
3472 NonCopyInstrsLen = 0;
3475 while (End->getOpcode() != FrameDestroyOpcode) {
3477 bool IsCopy =
false;
3478 if (End->getOpcode() == AMDGPU::COPY) {
3479 auto &Src = End->getOperand(1);
3482 IsCopy = Reg.isPhysical() &&
MI.modifiesRegister(Reg,
TRI);
3488 NonCopyInstrsLen = NonCopyInstrs.
size();
3493 NonCopyInstrs.
resize(NonCopyInstrsLen);
3497 for (
auto *NonCopy :
reverse(NonCopyInstrs)) {
3498 MBB->splice(LastCopy,
MBB, NonCopy->getIterator());
3502 B.setInsertPt(
B.getMBB(), Start);
3506 case AMDGPU::G_AMDGPU_FLAT_LOAD_MONITOR:
3507 case AMDGPU::G_AMDGPU_GLOBAL_LOAD_MONITOR:
3508 case AMDGPU::G_LOAD:
3509 case AMDGPU::G_ZEXTLOAD:
3510 case AMDGPU::G_SEXTLOAD: {
3515 case AMDGPU::G_DYN_STACKALLOC:
3518 case AMDGPU::G_STACKRESTORE: {
3523 case AMDGPU::G_SBFX:
3526 case AMDGPU::G_UBFX:
3529 case AMDGPU::G_AMDGPU_MAD_U64_U32:
3530 case AMDGPU::G_AMDGPU_MAD_I64_I32:
3533 case AMDGPU::G_PREFETCH: {
3535 MI.eraseFromParent();
3540 if (PtrBank == AMDGPU::VGPRRegBankID &&
3541 (!
Subtarget.hasVmemPrefInsts() || !
MI.getOperand(3).getImm())) {
3543 MI.eraseFromParent();
3546 unsigned AS =
MRI.getType(PtrReg).getAddressSpace();
3551 !
MI.getOperand(3).getImm() ))) {
3552 MI.eraseFromParent();
3849 if (
MI.isCopy() ||
MI.getOpcode() == AMDGPU::G_FREEZE) {
3864 DstBank = &AMDGPU::VCCRegBank;
3867 DstBank = &AMDGPU::VCCRegBank;
3878 if (
MI.getOpcode() != AMDGPU::G_FREEZE &&
3883 unsigned OpdsMappingSize =
MI.isCopy() ? 1 : 2;
3885 OpdsMapping[0] = &ValMap;
3886 if (
MI.getOpcode() == AMDGPU::G_FREEZE)
3887 OpdsMapping[1] = &ValMap;
3894 if (
MI.isRegSequence()) {
3897 unsigned BankID = AMDGPU::SGPRRegBankID;
3899 for (
unsigned I = 1, E =
MI.getNumOperands();
I != E;
I += 2) {
3903 if (OpBank != AMDGPU::SGPRRegBankID) {
3904 BankID = AMDGPU::VGPRRegBankID;
3921 unsigned ResultBank = AMDGPU::InvalidRegBankID;
3926 ResultBank = DstBank->
getID();
3928 for (
unsigned I = 0;
I <
PHI->getNumIncomingValues(); ++
I) {
3933 if (!Bank || Bank->
getID() == AMDGPU::VGPRRegBankID) {
3934 ResultBank = AMDGPU::VGPRRegBankID;
3939 unsigned OpBank = Bank->
getID();
3943 assert(ResultBank != AMDGPU::InvalidRegBankID);
3945 unsigned Size =
MRI.getType(DstReg).getSizeInBits();
3960 switch (
MI.getOpcode()) {
3967 case AMDGPU::G_MUL: {
3968 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
3973 unsigned TargetBankID = AMDGPU::InvalidRegBankID;
3974 unsigned BankLHS = AMDGPU::InvalidRegBankID;
3975 unsigned BankRHS = AMDGPU::InvalidRegBankID;
3977 TargetBankID = DstBank->
getID();
3978 if (DstBank == &AMDGPU::VCCRegBank) {
3979 TargetBankID = AMDGPU::VCCRegBankID;
3980 BankLHS = AMDGPU::VCCRegBankID;
3981 BankRHS = AMDGPU::VCCRegBankID;
3984 AMDGPU::SGPRRegBankID);
3986 AMDGPU::SGPRRegBankID);
3990 AMDGPU::VCCRegBankID);
3992 AMDGPU::VCCRegBankID);
3995 if (BankLHS == AMDGPU::VGPRRegBankID || BankRHS == AMDGPU::VGPRRegBankID) {
3996 TargetBankID = AMDGPU::VGPRRegBankID;
3997 }
else if (BankLHS == AMDGPU::VCCRegBankID || BankRHS == AMDGPU::VCCRegBankID) {
3998 TargetBankID = AMDGPU::VCCRegBankID;
3999 BankLHS = AMDGPU::VCCRegBankID;
4000 BankRHS = AMDGPU::VCCRegBankID;
4001 }
else if (BankLHS == AMDGPU::SGPRRegBankID && BankRHS == AMDGPU::SGPRRegBankID) {
4002 TargetBankID = AMDGPU::SGPRRegBankID;
4006 OpdsMapping[0] = AMDGPU::getValueMapping(TargetBankID,
Size);
4007 OpdsMapping[1] = AMDGPU::getValueMapping(BankLHS,
Size);
4008 OpdsMapping[2] = AMDGPU::getValueMapping(BankRHS,
Size);
4015 OpdsMapping[0] = getValueMappingSGPR64Only(AMDGPU::SGPRRegBankID,
Size);
4016 OpdsMapping[1] = OpdsMapping[2] = OpdsMapping[0];
4018 if (
MI.getOpcode() == AMDGPU::G_MUL &&
Subtarget.hasVectorMulU64())
4019 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4022 getValueMappingSGPR64Only(AMDGPU::VGPRRegBankID,
Size);
4024 OpdsMapping[1] = AMDGPU::getValueMapping(Bank1,
Size);
4027 OpdsMapping[2] = AMDGPU::getValueMapping(Bank2,
Size);
4035 case AMDGPU::G_PTR_ADD:
4036 case AMDGPU::G_PTRMASK:
4040 case AMDGPU::G_LSHR:
4041 case AMDGPU::G_ASHR:
4042 case AMDGPU::G_UADDO:
4043 case AMDGPU::G_USUBO:
4044 case AMDGPU::G_UADDE:
4045 case AMDGPU::G_SADDE:
4046 case AMDGPU::G_USUBE:
4047 case AMDGPU::G_SSUBE:
4049 case AMDGPU::G_SHUFFLE_VECTOR:
4050 case AMDGPU::G_SBFX:
4051 case AMDGPU::G_UBFX:
4052 case AMDGPU::G_AMDGPU_S_MUL_I64_I32:
4053 case AMDGPU::G_AMDGPU_S_MUL_U64_U32:
4057 case AMDGPU::G_SMIN:
4058 case AMDGPU::G_SMAX:
4059 case AMDGPU::G_UMIN:
4060 case AMDGPU::G_UMAX:
4063 if (
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits() == 64 &&
4069 case AMDGPU::G_FADD:
4070 case AMDGPU::G_FSUB:
4071 case AMDGPU::G_FMUL:
4073 case AMDGPU::G_FFLOOR:
4074 case AMDGPU::G_FCEIL:
4075 case AMDGPU::G_INTRINSIC_ROUNDEVEN:
4076 case AMDGPU::G_FMINNUM:
4077 case AMDGPU::G_FMAXNUM:
4078 case AMDGPU::G_FMINIMUM:
4079 case AMDGPU::G_FMAXIMUM:
4080 case AMDGPU::G_FMINIMUMNUM:
4081 case AMDGPU::G_FMAXIMUMNUM:
4082 case AMDGPU::G_INTRINSIC_TRUNC:
4083 case AMDGPU::G_STRICT_FADD:
4084 case AMDGPU::G_STRICT_FSUB:
4085 case AMDGPU::G_STRICT_FMUL:
4086 case AMDGPU::G_STRICT_FMA: {
4087 LLT Ty =
MRI.getType(
MI.getOperand(0).getReg());
4088 unsigned Size = Ty.getSizeInBits();
4089 if (
Subtarget.hasSALUFloatInsts() && Ty.isScalar() &&
4094 case AMDGPU::G_FPTOSI:
4095 case AMDGPU::G_FPTOUI:
4096 case AMDGPU::G_FPTOSI_SAT:
4097 case AMDGPU::G_FPTOUI_SAT:
4098 case AMDGPU::G_SITOFP:
4099 case AMDGPU::G_UITOFP: {
4100 unsigned SizeDst =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4101 unsigned SizeSrc =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4102 if (
Subtarget.hasSALUFloatInsts() && SizeDst == 32 && SizeSrc == 32 &&
4107 case AMDGPU::G_FPTRUNC:
4108 case AMDGPU::G_FPEXT: {
4109 unsigned SizeDst =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4110 unsigned SizeSrc =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4111 if (
Subtarget.hasSALUFloatInsts() && SizeDst != 64 && SizeSrc != 64 &&
4116 case AMDGPU::G_FSQRT:
4117 case AMDGPU::G_FEXP2:
4118 case AMDGPU::G_FLOG2: {
4119 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4125 case AMDGPU::G_SADDSAT:
4126 case AMDGPU::G_SSUBSAT:
4127 case AMDGPU::G_UADDSAT:
4128 case AMDGPU::G_USUBSAT:
4129 case AMDGPU::G_FMAD:
4130 case AMDGPU::G_FLDEXP:
4131 case AMDGPU::G_FMINNUM_IEEE:
4132 case AMDGPU::G_FMAXNUM_IEEE:
4133 case AMDGPU::G_FCANONICALIZE:
4134 case AMDGPU::G_STRICT_FLDEXP:
4135 case AMDGPU::G_BSWAP:
4136 case AMDGPU::G_FSHR:
4137 case AMDGPU::G_AMDGPU_FMIN_LEGACY:
4138 case AMDGPU::G_AMDGPU_FMAX_LEGACY:
4139 case AMDGPU::G_AMDGPU_RCP_IFLAG:
4140 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE0:
4141 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE1:
4142 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE2:
4143 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE3:
4144 case AMDGPU::G_AMDGPU_CVT_PK_I16_I32:
4145 case AMDGPU::G_AMDGPU_SMED3:
4146 case AMDGPU::G_AMDGPU_FMED3:
4148 case AMDGPU::G_UMULH:
4149 case AMDGPU::G_SMULH: {
4154 case AMDGPU::G_AMDGPU_MAD_U64_U32:
4155 case AMDGPU::G_AMDGPU_MAD_I64_I32: {
4164 bool AllSalu =
true;
4165 bool MulSalu =
true;
4166 for (
unsigned i = 0; i < 5; ++i) {
4169 if (Bank->getID() != AMDGPU::SGPRRegBankID) {
4171 if (i == 2 || i == 3) {
4185 if (!MulSalu ||
Subtarget.hasFullRate64Ops())
4189 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 64);
4190 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4191 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4192 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4193 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 64);
4196 case AMDGPU::G_IMPLICIT_DEF: {
4197 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4198 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4201 case AMDGPU::G_FCONSTANT:
4202 case AMDGPU::G_CONSTANT:
4203 case AMDGPU::G_GLOBAL_VALUE:
4204 case AMDGPU::G_FRAME_INDEX:
4205 case AMDGPU::G_BLOCK_ADDR:
4206 case AMDGPU::G_READSTEADYCOUNTER:
4207 case AMDGPU::G_READCYCLECOUNTER: {
4208 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4209 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4212 case AMDGPU::G_DYN_STACKALLOC: {
4214 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4216 OpdsMapping[1] = AMDGPU::getValueMapping(SrcBankID, 32);
4219 case AMDGPU::G_AMDGPU_WAVE_ADDRESS: {
4224 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4225 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4228 case AMDGPU::G_INSERT: {
4233 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, DstSize);
4234 OpdsMapping[1] = AMDGPU::getValueMapping(BankID, SrcSize);
4235 OpdsMapping[2] = AMDGPU::getValueMapping(BankID, EltSize);
4236 OpdsMapping[3] =
nullptr;
4239 case AMDGPU::G_EXTRACT: {
4243 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, DstSize);
4244 OpdsMapping[1] = AMDGPU::getValueMapping(BankID, SrcSize);
4245 OpdsMapping[2] =
nullptr;
4248 case AMDGPU::G_BUILD_VECTOR:
4249 case AMDGPU::G_BUILD_VECTOR_TRUNC: {
4250 LLT DstTy =
MRI.getType(
MI.getOperand(0).getReg());
4253 unsigned SrcSize =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4256 unsigned DstBankID =
regBankUnion(Src0BankID, Src1BankID);
4258 OpdsMapping[0] = AMDGPU::getValueMapping(DstBankID, DstSize);
4259 OpdsMapping[1] = AMDGPU::getValueMapping(Src0BankID, SrcSize);
4260 OpdsMapping[2] = AMDGPU::getValueMapping(Src1BankID, SrcSize);
4266 case AMDGPU::G_MERGE_VALUES:
4267 case AMDGPU::G_CONCAT_VECTORS: {
4269 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4270 unsigned SrcSize =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4272 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, DstSize);
4274 for (
unsigned i = 1, e =
MI.getNumOperands(); i != e; ++i)
4275 OpdsMapping[i] = AMDGPU::getValueMapping(Bank, SrcSize);
4278 case AMDGPU::G_BITREVERSE:
4279 case AMDGPU::G_BITCAST:
4280 case AMDGPU::G_INTTOPTR:
4281 case AMDGPU::G_PTRTOINT:
4282 case AMDGPU::G_FABS:
4283 case AMDGPU::G_FNEG: {
4284 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4286 OpdsMapping[0] = OpdsMapping[1] = AMDGPU::getValueMapping(BankID,
Size);
4289 case AMDGPU::G_AMDGPU_FFBH_U32:
4290 case AMDGPU::G_AMDGPU_FFBL_B32:
4291 case AMDGPU::G_CTLZ_ZERO_UNDEF:
4292 case AMDGPU::G_CTTZ_ZERO_UNDEF: {
4293 unsigned Size =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4295 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, 32);
4296 OpdsMapping[1] = AMDGPU::getValueMappingSGPR64Only(BankID,
Size);
4299 case AMDGPU::G_CTPOP: {
4300 unsigned Size =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4302 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, 32);
4307 OpdsMapping[1] = AMDGPU::getValueMapping(BankID,
Size);
4310 case AMDGPU::G_TRUNC: {
4316 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, DstSize);
4317 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, SrcSize);
4320 case AMDGPU::G_ZEXT:
4321 case AMDGPU::G_SEXT:
4322 case AMDGPU::G_ANYEXT:
4323 case AMDGPU::G_SEXT_INREG: {
4332 switch (SrcBank->
getID()) {
4333 case AMDGPU::SGPRRegBankID:
4334 DstBank = AMDGPU::SGPRRegBankID;
4337 DstBank = AMDGPU::VGPRRegBankID;
4343 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(DstBank, DstSize);
4344 OpdsMapping[1] = AMDGPU::getValueMappingSGPR64Only(SrcBank->
getID(),
4348 case AMDGPU::G_IS_FPCLASS: {
4350 unsigned SrcSize =
MRI.getType(SrcReg).getSizeInBits();
4351 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4352 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, DstSize);
4353 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4356 case AMDGPU::G_STORE: {
4358 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4363 AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4364 OpdsMapping[0] = ValMapping;
4368 case AMDGPU::G_ICMP:
4369 case AMDGPU::G_FCMP: {
4370 unsigned Size =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4375 AMDGPU::SGPRRegBankID);
4379 auto canUseSCCICMP = [&]() {
4382 return Size == 32 ||
4387 auto canUseSCCFCMP = [&]() {
4391 bool isICMP =
MI.getOpcode() == AMDGPU::G_ICMP;
4392 bool CanUseSCC = DstBank == AMDGPU::SGPRRegBankID &&
4393 Op2Bank == AMDGPU::SGPRRegBankID &&
4394 Op3Bank == AMDGPU::SGPRRegBankID &&
4395 (isICMP ? canUseSCCICMP() : canUseSCCFCMP());
4397 DstBank = CanUseSCC ? AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
4398 unsigned SrcBank = CanUseSCC ? AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
4402 const unsigned ResultSize = 1;
4404 OpdsMapping[0] = AMDGPU::getValueMapping(DstBank, ResultSize);
4405 OpdsMapping[1] =
nullptr;
4406 OpdsMapping[2] = AMDGPU::getValueMapping(SrcBank,
Size);
4407 OpdsMapping[3] = AMDGPU::getValueMapping(SrcBank,
Size);
4410 case AMDGPU::G_EXTRACT_VECTOR_ELT: {
4413 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4414 unsigned SrcSize =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4415 unsigned IdxSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4417 unsigned OutputBankID =
regBankUnion(SrcBankID, IdxBank);
4419 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(OutputBankID, DstSize);
4420 OpdsMapping[1] = AMDGPU::getValueMapping(SrcBankID, SrcSize);
4423 OpdsMapping[2] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4426 case AMDGPU::G_INSERT_VECTOR_ELT: {
4428 AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
4430 unsigned VecSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4431 unsigned InsertSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4432 unsigned IdxSize =
MRI.getType(
MI.getOperand(3).getReg()).getSizeInBits();
4436 OpdsMapping[0] = AMDGPU::getValueMapping(OutputBankID, VecSize);
4437 OpdsMapping[1] = AMDGPU::getValueMapping(OutputBankID, VecSize);
4441 if (InsertSize == 64 && OutputBankID == AMDGPU::VGPRRegBankID) {
4442 OpdsMapping[2] = AMDGPU::getValueMappingSplit64(InsertEltBankID,
4445 assert(InsertSize == 32 || InsertSize == 64);
4446 OpdsMapping[2] = AMDGPU::getValueMapping(InsertEltBankID, InsertSize);
4450 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBankID, IdxSize);
4453 case AMDGPU::G_UNMERGE_VALUES: {
4458 for (
unsigned i = 0, e =
MI.getNumOperands(); i != e; ++i) {
4460 OpdsMapping[i] = AMDGPU::getValueMapping(Bank,
Size);
4464 case AMDGPU::G_AMDGPU_BUFFER_LOAD:
4465 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE:
4466 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE:
4467 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT:
4468 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT:
4469 case AMDGPU::G_AMDGPU_BUFFER_LOAD_TFE:
4470 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE_TFE:
4471 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE_TFE:
4472 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT_TFE:
4473 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT_TFE:
4474 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT:
4475 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_TFE:
4476 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_D16:
4477 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT:
4478 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT_D16:
4479 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT:
4480 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT_D16:
4481 case AMDGPU::G_AMDGPU_BUFFER_STORE:
4482 case AMDGPU::G_AMDGPU_BUFFER_STORE_BYTE:
4483 case AMDGPU::G_AMDGPU_BUFFER_STORE_SHORT:
4484 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT:
4485 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT_D16: {
4504 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SWAP:
4505 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_ADD:
4506 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB:
4507 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMIN:
4508 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMIN:
4509 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMAX:
4510 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMAX:
4511 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_AND:
4512 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_OR:
4513 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_XOR:
4514 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_INC:
4515 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_DEC:
4516 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB_CLAMP_U32:
4517 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_COND_SUB_U32:
4518 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FADD:
4519 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMIN:
4520 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMAX: {
4543 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_CMPSWAP: {
4569 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD:
4570 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_UBYTE:
4571 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SBYTE:
4572 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_USHORT:
4573 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SSHORT: {
4581 unsigned RSrcBank = OpdsMapping[1]->BreakDown[0].RegBank->getID();
4582 unsigned OffsetBank = OpdsMapping[2]->BreakDown[0].RegBank->getID();
4583 unsigned ResultBank =
regBankUnion(RSrcBank, OffsetBank);
4585 unsigned Size0 =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4586 OpdsMapping[0] = AMDGPU::getValueMapping(ResultBank, Size0);
4589 case AMDGPU::G_AMDGPU_S_BUFFER_PREFETCH:
4593 case AMDGPU::G_AMDGPU_SPONENTRY: {
4594 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4595 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4598 case AMDGPU::G_INTRINSIC:
4599 case AMDGPU::G_INTRINSIC_CONVERGENT: {
4603 case Intrinsic::amdgcn_div_fmas:
4604 case Intrinsic::amdgcn_div_fixup:
4605 case Intrinsic::amdgcn_trig_preop:
4606 case Intrinsic::amdgcn_sin:
4607 case Intrinsic::amdgcn_cos:
4608 case Intrinsic::amdgcn_log_clamp:
4609 case Intrinsic::amdgcn_rcp_legacy:
4610 case Intrinsic::amdgcn_rsq_legacy:
4611 case Intrinsic::amdgcn_rsq_clamp:
4612 case Intrinsic::amdgcn_tanh:
4613 case Intrinsic::amdgcn_fmul_legacy:
4614 case Intrinsic::amdgcn_fma_legacy:
4615 case Intrinsic::amdgcn_frexp_mant:
4616 case Intrinsic::amdgcn_frexp_exp:
4617 case Intrinsic::amdgcn_fract:
4618 case Intrinsic::amdgcn_cvt_pknorm_i16:
4619 case Intrinsic::amdgcn_cvt_pknorm_u16:
4620 case Intrinsic::amdgcn_cvt_pk_i16:
4621 case Intrinsic::amdgcn_cvt_pk_u16:
4622 case Intrinsic::amdgcn_cvt_sr_pk_f16_f32:
4623 case Intrinsic::amdgcn_cvt_sr_pk_bf16_f32:
4624 case Intrinsic::amdgcn_cvt_pk_f16_fp8:
4625 case Intrinsic::amdgcn_cvt_pk_f16_bf8:
4626 case Intrinsic::amdgcn_cvt_pk_fp8_f16:
4627 case Intrinsic::amdgcn_cvt_pk_bf8_f16:
4628 case Intrinsic::amdgcn_cvt_sr_fp8_f16:
4629 case Intrinsic::amdgcn_cvt_sr_bf8_f16:
4630 case Intrinsic::amdgcn_cvt_scale_pk8_f16_fp8:
4631 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_fp8:
4632 case Intrinsic::amdgcn_cvt_scale_pk8_f16_bf8:
4633 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_bf8:
4634 case Intrinsic::amdgcn_cvt_scale_pk8_f16_fp4:
4635 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_fp4:
4636 case Intrinsic::amdgcn_cvt_scale_pk8_f32_fp8:
4637 case Intrinsic::amdgcn_cvt_scale_pk8_f32_bf8:
4638 case Intrinsic::amdgcn_cvt_scale_pk8_f32_fp4:
4639 case Intrinsic::amdgcn_cvt_scale_pk16_f16_fp6:
4640 case Intrinsic::amdgcn_cvt_scale_pk16_bf16_fp6:
4641 case Intrinsic::amdgcn_cvt_scale_pk16_f16_bf6:
4642 case Intrinsic::amdgcn_cvt_scale_pk16_bf16_bf6:
4643 case Intrinsic::amdgcn_cvt_scale_pk16_f32_fp6:
4644 case Intrinsic::amdgcn_cvt_scale_pk16_f32_bf6:
4645 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_bf16:
4646 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_bf16:
4647 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_f16:
4648 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_f16:
4649 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_f32:
4650 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_f32:
4651 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_f32:
4652 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_f16:
4653 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_bf16:
4654 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_f32:
4655 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_f32:
4656 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_f16:
4657 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_f16:
4658 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_bf16:
4659 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_bf16:
4660 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_bf16:
4661 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_bf16:
4662 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_f16:
4663 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_f16:
4664 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_f32:
4665 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_f32:
4666 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_f32:
4667 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_f16:
4668 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_bf16:
4669 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_f32:
4670 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_f32:
4671 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_f16:
4672 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_f16:
4673 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_bf16:
4674 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_bf16:
4675 case Intrinsic::amdgcn_sat_pk4_i4_i8:
4676 case Intrinsic::amdgcn_sat_pk4_u4_u8:
4677 case Intrinsic::amdgcn_fmed3:
4678 case Intrinsic::amdgcn_cubeid:
4679 case Intrinsic::amdgcn_cubema:
4680 case Intrinsic::amdgcn_cubesc:
4681 case Intrinsic::amdgcn_cubetc:
4682 case Intrinsic::amdgcn_sffbh:
4683 case Intrinsic::amdgcn_fmad_ftz:
4684 case Intrinsic::amdgcn_mbcnt_lo:
4685 case Intrinsic::amdgcn_mbcnt_hi:
4686 case Intrinsic::amdgcn_mul_u24:
4687 case Intrinsic::amdgcn_mul_i24:
4688 case Intrinsic::amdgcn_mulhi_u24:
4689 case Intrinsic::amdgcn_mulhi_i24:
4690 case Intrinsic::amdgcn_lerp:
4691 case Intrinsic::amdgcn_sad_u8:
4692 case Intrinsic::amdgcn_msad_u8:
4693 case Intrinsic::amdgcn_sad_hi_u8:
4694 case Intrinsic::amdgcn_sad_u16:
4695 case Intrinsic::amdgcn_qsad_pk_u16_u8:
4696 case Intrinsic::amdgcn_mqsad_pk_u16_u8:
4697 case Intrinsic::amdgcn_mqsad_u32_u8:
4698 case Intrinsic::amdgcn_cvt_pk_u8_f32:
4699 case Intrinsic::amdgcn_alignbyte:
4700 case Intrinsic::amdgcn_perm:
4701 case Intrinsic::amdgcn_prng_b32:
4702 case Intrinsic::amdgcn_fdot2:
4703 case Intrinsic::amdgcn_sdot2:
4704 case Intrinsic::amdgcn_udot2:
4705 case Intrinsic::amdgcn_sdot4:
4706 case Intrinsic::amdgcn_udot4:
4707 case Intrinsic::amdgcn_sdot8:
4708 case Intrinsic::amdgcn_udot8:
4709 case Intrinsic::amdgcn_fdot2_bf16_bf16:
4710 case Intrinsic::amdgcn_fdot2_f16_f16:
4711 case Intrinsic::amdgcn_fdot2_f32_bf16:
4712 case Intrinsic::amdgcn_fdot2c_f32_bf16:
4713 case Intrinsic::amdgcn_sudot4:
4714 case Intrinsic::amdgcn_sudot8:
4715 case Intrinsic::amdgcn_dot4_f32_fp8_bf8:
4716 case Intrinsic::amdgcn_dot4_f32_bf8_fp8:
4717 case Intrinsic::amdgcn_dot4_f32_fp8_fp8:
4718 case Intrinsic::amdgcn_dot4_f32_bf8_bf8:
4719 case Intrinsic::amdgcn_cvt_f32_fp8:
4720 case Intrinsic::amdgcn_cvt_f32_fp8_e5m3:
4721 case Intrinsic::amdgcn_cvt_f32_bf8:
4722 case Intrinsic::amdgcn_cvt_off_f32_i4:
4723 case Intrinsic::amdgcn_cvt_pk_f32_fp8:
4724 case Intrinsic::amdgcn_cvt_pk_f32_bf8:
4725 case Intrinsic::amdgcn_cvt_pk_fp8_f32:
4726 case Intrinsic::amdgcn_cvt_pk_fp8_f32_e5m3:
4727 case Intrinsic::amdgcn_cvt_pk_bf8_f32:
4728 case Intrinsic::amdgcn_cvt_sr_fp8_f32:
4729 case Intrinsic::amdgcn_cvt_sr_fp8_f32_e5m3:
4730 case Intrinsic::amdgcn_cvt_sr_bf8_f32:
4731 case Intrinsic::amdgcn_cvt_sr_bf16_f32:
4732 case Intrinsic::amdgcn_cvt_sr_f16_f32:
4733 case Intrinsic::amdgcn_cvt_f16_fp8:
4734 case Intrinsic::amdgcn_cvt_f16_bf8:
4735 case Intrinsic::amdgcn_cvt_scalef32_pk32_fp6_f16:
4736 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf6_f16:
4737 case Intrinsic::amdgcn_cvt_scalef32_pk32_fp6_bf16:
4738 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf6_bf16:
4739 case Intrinsic::amdgcn_cvt_scalef32_f16_fp8:
4740 case Intrinsic::amdgcn_cvt_scalef32_f16_bf8:
4741 case Intrinsic::amdgcn_cvt_scalef32_f32_fp8:
4742 case Intrinsic::amdgcn_cvt_scalef32_f32_bf8:
4743 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_f32:
4744 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_f32:
4745 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_fp8:
4746 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_bf8:
4747 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_f16:
4748 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_bf16:
4749 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_f16:
4750 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_bf16:
4751 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_fp4:
4752 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_f32:
4753 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_fp4:
4754 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_fp4:
4755 case Intrinsic::amdgcn_cvt_scalef32_pk32_f32_fp6:
4756 case Intrinsic::amdgcn_cvt_scalef32_pk32_f32_bf6:
4757 case Intrinsic::amdgcn_cvt_scalef32_pk32_f16_bf6:
4758 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf16_bf6:
4759 case Intrinsic::amdgcn_cvt_scalef32_pk32_f16_fp6:
4760 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf16_fp6:
4761 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_bf8:
4762 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_bf8:
4763 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_fp8:
4764 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_fp8:
4765 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_f16:
4766 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_bf16:
4767 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_f16:
4768 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_bf16:
4769 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_f32:
4770 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_bf16:
4771 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_f16:
4772 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_f32:
4773 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_bf16:
4774 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_f16:
4775 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_f32:
4776 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_bf16:
4777 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_f16:
4778 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_f32:
4779 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_bf16:
4780 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_f16:
4781 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_f32:
4782 case Intrinsic::amdgcn_ashr_pk_i8_i32:
4783 case Intrinsic::amdgcn_ashr_pk_u8_i32:
4784 case Intrinsic::amdgcn_cvt_scalef32_2xpk16_fp6_f32:
4785 case Intrinsic::amdgcn_cvt_scalef32_2xpk16_bf6_f32:
4786 case Intrinsic::amdgcn_wmma_bf16_16x16x16_bf16:
4787 case Intrinsic::amdgcn_wmma_f16_16x16x16_f16:
4788 case Intrinsic::amdgcn_wmma_bf16_16x16x16_bf16_tied:
4789 case Intrinsic::amdgcn_wmma_f16_16x16x16_f16_tied:
4790 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf16:
4791 case Intrinsic::amdgcn_wmma_f32_16x16x16_f16:
4792 case Intrinsic::amdgcn_wmma_i32_16x16x16_iu4:
4793 case Intrinsic::amdgcn_wmma_i32_16x16x16_iu8:
4794 case Intrinsic::amdgcn_wmma_f32_16x16x16_fp8_fp8:
4795 case Intrinsic::amdgcn_wmma_f32_16x16x16_fp8_bf8:
4796 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf8_fp8:
4797 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf8_bf8:
4798 case Intrinsic::amdgcn_wmma_i32_16x16x32_iu4:
4799 case Intrinsic::amdgcn_swmmac_f32_16x16x32_f16:
4800 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf16:
4801 case Intrinsic::amdgcn_swmmac_f16_16x16x32_f16:
4802 case Intrinsic::amdgcn_swmmac_bf16_16x16x32_bf16:
4803 case Intrinsic::amdgcn_swmmac_i32_16x16x32_iu8:
4804 case Intrinsic::amdgcn_swmmac_i32_16x16x32_iu4:
4805 case Intrinsic::amdgcn_swmmac_i32_16x16x64_iu4:
4806 case Intrinsic::amdgcn_swmmac_f32_16x16x32_fp8_fp8:
4807 case Intrinsic::amdgcn_swmmac_f32_16x16x32_fp8_bf8:
4808 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf8_fp8:
4809 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf8_bf8:
4810 case Intrinsic::amdgcn_wmma_f32_16x16x4_f32:
4811 case Intrinsic::amdgcn_wmma_f32_16x16x32_bf16:
4812 case Intrinsic::amdgcn_wmma_f32_16x16x32_f16:
4813 case Intrinsic::amdgcn_wmma_f16_16x16x32_f16:
4814 case Intrinsic::amdgcn_wmma_bf16_16x16x32_bf16:
4815 case Intrinsic::amdgcn_wmma_bf16f32_16x16x32_bf16:
4816 case Intrinsic::amdgcn_wmma_f32_16x16x64_fp8_fp8:
4817 case Intrinsic::amdgcn_wmma_f32_16x16x64_fp8_bf8:
4818 case Intrinsic::amdgcn_wmma_f32_16x16x64_bf8_fp8:
4819 case Intrinsic::amdgcn_wmma_f32_16x16x64_bf8_bf8:
4820 case Intrinsic::amdgcn_wmma_f16_16x16x64_fp8_fp8:
4821 case Intrinsic::amdgcn_wmma_f16_16x16x64_fp8_bf8:
4822 case Intrinsic::amdgcn_wmma_f16_16x16x64_bf8_fp8:
4823 case Intrinsic::amdgcn_wmma_f16_16x16x64_bf8_bf8:
4824 case Intrinsic::amdgcn_wmma_f16_16x16x128_fp8_fp8:
4825 case Intrinsic::amdgcn_wmma_f16_16x16x128_fp8_bf8:
4826 case Intrinsic::amdgcn_wmma_f16_16x16x128_bf8_fp8:
4827 case Intrinsic::amdgcn_wmma_f16_16x16x128_bf8_bf8:
4828 case Intrinsic::amdgcn_wmma_f32_16x16x128_fp8_fp8:
4829 case Intrinsic::amdgcn_wmma_f32_16x16x128_fp8_bf8:
4830 case Intrinsic::amdgcn_wmma_f32_16x16x128_bf8_fp8:
4831 case Intrinsic::amdgcn_wmma_f32_16x16x128_bf8_bf8:
4832 case Intrinsic::amdgcn_wmma_i32_16x16x64_iu8:
4833 case Intrinsic::amdgcn_wmma_f32_16x16x128_f8f6f4:
4834 case Intrinsic::amdgcn_wmma_scale_f32_16x16x128_f8f6f4:
4835 case Intrinsic::amdgcn_wmma_scale16_f32_16x16x128_f8f6f4:
4836 case Intrinsic::amdgcn_wmma_f32_32x16x128_f4:
4837 case Intrinsic::amdgcn_wmma_scale_f32_32x16x128_f4:
4838 case Intrinsic::amdgcn_wmma_scale16_f32_32x16x128_f4:
4839 case Intrinsic::amdgcn_swmmac_f16_16x16x64_f16:
4840 case Intrinsic::amdgcn_swmmac_bf16_16x16x64_bf16:
4841 case Intrinsic::amdgcn_swmmac_f32_16x16x64_bf16:
4842 case Intrinsic::amdgcn_swmmac_bf16f32_16x16x64_bf16:
4843 case Intrinsic::amdgcn_swmmac_f32_16x16x64_f16:
4844 case Intrinsic::amdgcn_swmmac_f32_16x16x128_fp8_fp8:
4845 case Intrinsic::amdgcn_swmmac_f32_16x16x128_fp8_bf8:
4846 case Intrinsic::amdgcn_swmmac_f32_16x16x128_bf8_fp8:
4847 case Intrinsic::amdgcn_swmmac_f32_16x16x128_bf8_bf8:
4848 case Intrinsic::amdgcn_swmmac_f16_16x16x128_fp8_fp8:
4849 case Intrinsic::amdgcn_swmmac_f16_16x16x128_fp8_bf8:
4850 case Intrinsic::amdgcn_swmmac_f16_16x16x128_bf8_fp8:
4851 case Intrinsic::amdgcn_swmmac_f16_16x16x128_bf8_bf8:
4852 case Intrinsic::amdgcn_swmmac_i32_16x16x128_iu8:
4853 case Intrinsic::amdgcn_perm_pk16_b4_u4:
4854 case Intrinsic::amdgcn_perm_pk16_b6_u4:
4855 case Intrinsic::amdgcn_perm_pk16_b8_u4:
4856 case Intrinsic::amdgcn_add_max_i32:
4857 case Intrinsic::amdgcn_add_max_u32:
4858 case Intrinsic::amdgcn_add_min_i32:
4859 case Intrinsic::amdgcn_add_min_u32:
4860 case Intrinsic::amdgcn_pk_add_max_i16:
4861 case Intrinsic::amdgcn_pk_add_max_u16:
4862 case Intrinsic::amdgcn_pk_add_min_i16:
4863 case Intrinsic::amdgcn_pk_add_min_u16:
4865 case Intrinsic::amdgcn_log:
4866 case Intrinsic::amdgcn_exp2:
4867 case Intrinsic::amdgcn_rcp:
4868 case Intrinsic::amdgcn_rsq:
4869 case Intrinsic::amdgcn_sqrt: {
4870 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4876 case Intrinsic::amdgcn_sbfe:
4877 case Intrinsic::amdgcn_ubfe:
4881 case Intrinsic::amdgcn_ds_swizzle:
4882 case Intrinsic::amdgcn_ds_permute:
4883 case Intrinsic::amdgcn_ds_bpermute:
4884 case Intrinsic::amdgcn_update_dpp:
4885 case Intrinsic::amdgcn_mov_dpp8:
4886 case Intrinsic::amdgcn_mov_dpp:
4887 case Intrinsic::amdgcn_strict_wwm:
4888 case Intrinsic::amdgcn_wwm:
4889 case Intrinsic::amdgcn_strict_wqm:
4890 case Intrinsic::amdgcn_wqm:
4891 case Intrinsic::amdgcn_softwqm:
4892 case Intrinsic::amdgcn_set_inactive:
4893 case Intrinsic::amdgcn_set_inactive_chain_arg:
4894 case Intrinsic::amdgcn_permlane64:
4895 case Intrinsic::amdgcn_ds_bpermute_fi_b32:
4897 case Intrinsic::amdgcn_cvt_pkrtz:
4901 case Intrinsic::amdgcn_kernarg_segment_ptr:
4902 case Intrinsic::amdgcn_s_getpc:
4903 case Intrinsic::amdgcn_groupstaticsize:
4904 case Intrinsic::amdgcn_reloc_constant:
4905 case Intrinsic::returnaddress: {
4906 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4907 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4910 case Intrinsic::amdgcn_wqm_vote: {
4911 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4912 OpdsMapping[0] = OpdsMapping[2]
4913 = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID,
Size);
4916 case Intrinsic::amdgcn_ps_live: {
4917 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4920 case Intrinsic::amdgcn_div_scale: {
4921 unsigned Dst0Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4922 unsigned Dst1Size =
MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits();
4923 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Dst0Size);
4924 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, Dst1Size);
4926 unsigned SrcSize =
MRI.getType(
MI.getOperand(3).getReg()).getSizeInBits();
4927 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4928 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4931 case Intrinsic::amdgcn_class: {
4932 Register Src0Reg =
MI.getOperand(2).getReg();
4933 Register Src1Reg =
MI.getOperand(3).getReg();
4934 unsigned Src0Size =
MRI.getType(Src0Reg).getSizeInBits();
4935 unsigned Src1Size =
MRI.getType(Src1Reg).getSizeInBits();
4936 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4937 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, DstSize);
4938 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Src0Size);
4939 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Src1Size);
4942 case Intrinsic::amdgcn_icmp:
4943 case Intrinsic::amdgcn_fcmp: {
4944 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4946 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
4947 unsigned OpSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4948 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
4949 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
4952 case Intrinsic::amdgcn_readlane: {
4955 unsigned IdxSize =
MRI.getType(IdxReg).getSizeInBits();
4957 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4960 case Intrinsic::amdgcn_readfirstlane: {
4961 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4962 unsigned SrcSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
4963 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
4964 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4967 case Intrinsic::amdgcn_writelane: {
4968 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
4970 unsigned SrcSize =
MRI.getType(SrcReg).getSizeInBits();
4973 unsigned IdxSize =
MRI.getType(IdxReg).getSizeInBits();
4975 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
4979 OpdsMapping[2] = AMDGPU::getValueMapping(SrcBank, SrcSize);
4980 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4981 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4984 case Intrinsic::amdgcn_if_break: {
4986 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4987 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4988 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4991 case Intrinsic::amdgcn_permlane16:
4992 case Intrinsic::amdgcn_permlanex16: {
4994 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4995 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4996 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5001 case Intrinsic::amdgcn_permlane_bcast:
5002 case Intrinsic::amdgcn_permlane_up:
5003 case Intrinsic::amdgcn_permlane_down:
5004 case Intrinsic::amdgcn_permlane_xor: {
5006 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5007 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5012 case Intrinsic::amdgcn_permlane_idx_gen: {
5014 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5015 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5019 case Intrinsic::amdgcn_permlane16_var:
5020 case Intrinsic::amdgcn_permlanex16_var: {
5022 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5023 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5024 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5025 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5028 case Intrinsic::amdgcn_mfma_f32_4x4x1f32:
5029 case Intrinsic::amdgcn_mfma_f32_4x4x4f16:
5030 case Intrinsic::amdgcn_mfma_i32_4x4x4i8:
5031 case Intrinsic::amdgcn_mfma_f32_4x4x2bf16:
5032 case Intrinsic::amdgcn_mfma_f32_16x16x1f32:
5033 case Intrinsic::amdgcn_mfma_f32_16x16x4f32:
5034 case Intrinsic::amdgcn_mfma_f32_16x16x4f16:
5035 case Intrinsic::amdgcn_mfma_f32_16x16x16f16:
5036 case Intrinsic::amdgcn_mfma_i32_16x16x4i8:
5037 case Intrinsic::amdgcn_mfma_i32_16x16x16i8:
5038 case Intrinsic::amdgcn_mfma_f32_16x16x2bf16:
5039 case Intrinsic::amdgcn_mfma_f32_16x16x8bf16:
5040 case Intrinsic::amdgcn_mfma_f32_32x32x1f32:
5041 case Intrinsic::amdgcn_mfma_f32_32x32x2f32:
5042 case Intrinsic::amdgcn_mfma_f32_32x32x4f16:
5043 case Intrinsic::amdgcn_mfma_f32_32x32x8f16:
5044 case Intrinsic::amdgcn_mfma_i32_32x32x4i8:
5045 case Intrinsic::amdgcn_mfma_i32_32x32x8i8:
5046 case Intrinsic::amdgcn_mfma_f32_32x32x2bf16:
5047 case Intrinsic::amdgcn_mfma_f32_32x32x4bf16:
5048 case Intrinsic::amdgcn_mfma_f32_32x32x4bf16_1k:
5049 case Intrinsic::amdgcn_mfma_f32_16x16x4bf16_1k:
5050 case Intrinsic::amdgcn_mfma_f32_4x4x4bf16_1k:
5051 case Intrinsic::amdgcn_mfma_f32_32x32x8bf16_1k:
5052 case Intrinsic::amdgcn_mfma_f32_16x16x16bf16_1k:
5053 case Intrinsic::amdgcn_mfma_f64_16x16x4f64:
5054 case Intrinsic::amdgcn_mfma_f64_4x4x4f64:
5055 case Intrinsic::amdgcn_mfma_i32_16x16x32_i8:
5056 case Intrinsic::amdgcn_mfma_i32_32x32x16_i8:
5057 case Intrinsic::amdgcn_mfma_f32_16x16x8_xf32:
5058 case Intrinsic::amdgcn_mfma_f32_32x32x4_xf32:
5059 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf8_bf8:
5060 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf8_fp8:
5061 case Intrinsic::amdgcn_mfma_f32_16x16x32_fp8_bf8:
5062 case Intrinsic::amdgcn_mfma_f32_16x16x32_fp8_fp8:
5063 case Intrinsic::amdgcn_mfma_f32_32x32x16_bf8_bf8:
5064 case Intrinsic::amdgcn_mfma_f32_32x32x16_bf8_fp8:
5065 case Intrinsic::amdgcn_mfma_f32_32x32x16_fp8_bf8:
5066 case Intrinsic::amdgcn_mfma_f32_32x32x16_fp8_fp8:
5067 case Intrinsic::amdgcn_mfma_f32_16x16x32_f16:
5068 case Intrinsic::amdgcn_mfma_f32_32x32x16_f16:
5069 case Intrinsic::amdgcn_mfma_i32_16x16x64_i8:
5070 case Intrinsic::amdgcn_mfma_i32_32x32x32_i8:
5071 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf16: {
5072 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5073 unsigned MinNumRegsRequired = DstSize / 32;
5083 bool UseAGPRForm = !
Subtarget.hasGFX90AInsts() ||
5084 Info->selectAGPRFormMFMA(MinNumRegsRequired);
5096 case Intrinsic::amdgcn_mfma_scale_f32_16x16x128_f8f6f4:
5097 case Intrinsic::amdgcn_mfma_scale_f32_32x32x64_f8f6f4: {
5098 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5099 unsigned MinNumRegsRequired = DstSize / 32;
5118 case Intrinsic::amdgcn_smfmac_f32_16x16x32_f16:
5119 case Intrinsic::amdgcn_smfmac_f32_32x32x16_f16:
5120 case Intrinsic::amdgcn_smfmac_f32_16x16x32_bf16:
5121 case Intrinsic::amdgcn_smfmac_f32_32x32x16_bf16:
5122 case Intrinsic::amdgcn_smfmac_i32_16x16x64_i8:
5123 case Intrinsic::amdgcn_smfmac_i32_32x32x32_i8:
5124 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf8_bf8:
5125 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf8_fp8:
5126 case Intrinsic::amdgcn_smfmac_f32_16x16x64_fp8_bf8:
5127 case Intrinsic::amdgcn_smfmac_f32_16x16x64_fp8_fp8:
5128 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf8_bf8:
5129 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf8_fp8:
5130 case Intrinsic::amdgcn_smfmac_f32_32x32x32_fp8_bf8:
5131 case Intrinsic::amdgcn_smfmac_f32_32x32x32_fp8_fp8:
5132 case Intrinsic::amdgcn_smfmac_f32_16x16x64_f16:
5133 case Intrinsic::amdgcn_smfmac_f32_32x32x32_f16:
5134 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf16:
5135 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf16:
5136 case Intrinsic::amdgcn_smfmac_i32_16x16x128_i8:
5137 case Intrinsic::amdgcn_smfmac_i32_32x32x64_i8:
5138 case Intrinsic::amdgcn_smfmac_f32_16x16x128_bf8_bf8:
5139 case Intrinsic::amdgcn_smfmac_f32_16x16x128_bf8_fp8:
5140 case Intrinsic::amdgcn_smfmac_f32_16x16x128_fp8_bf8:
5141 case Intrinsic::amdgcn_smfmac_f32_16x16x128_fp8_fp8:
5142 case Intrinsic::amdgcn_smfmac_f32_32x32x64_bf8_bf8:
5143 case Intrinsic::amdgcn_smfmac_f32_32x32x64_bf8_fp8:
5144 case Intrinsic::amdgcn_smfmac_f32_32x32x64_fp8_bf8:
5145 case Intrinsic::amdgcn_smfmac_f32_32x32x64_fp8_fp8: {
5147 unsigned DstSize =
MRI.getType(DstReg).getSizeInBits();
5148 unsigned MinNumRegsRequired = DstSize / 32;
5164 case Intrinsic::amdgcn_interp_p1:
5165 case Intrinsic::amdgcn_interp_p2:
5166 case Intrinsic::amdgcn_interp_mov:
5167 case Intrinsic::amdgcn_interp_p1_f16:
5168 case Intrinsic::amdgcn_interp_p2_f16:
5169 case Intrinsic::amdgcn_lds_param_load: {
5170 const int M0Idx =
MI.getNumOperands() - 1;
5171 Register M0Reg =
MI.getOperand(M0Idx).getReg();
5173 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5175 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5176 for (
int I = 2;
I != M0Idx &&
MI.getOperand(
I).
isReg(); ++
I)
5177 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5181 OpdsMapping[M0Idx] = AMDGPU::getValueMapping(M0Bank, 32);
5184 case Intrinsic::amdgcn_interp_inreg_p10:
5185 case Intrinsic::amdgcn_interp_inreg_p2:
5186 case Intrinsic::amdgcn_interp_inreg_p10_f16:
5187 case Intrinsic::amdgcn_interp_inreg_p2_f16:
5188 case Intrinsic::amdgcn_interp_p10_rtz_f16:
5189 case Intrinsic::amdgcn_interp_p2_rtz_f16: {
5190 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5191 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5192 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5193 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5194 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5197 case Intrinsic::amdgcn_permlane16_swap:
5198 case Intrinsic::amdgcn_permlane32_swap: {
5199 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5200 OpdsMapping[0] = OpdsMapping[1] = OpdsMapping[3] = OpdsMapping[4] =
5201 AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5204 case Intrinsic::amdgcn_ballot: {
5205 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5206 unsigned SrcSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
5207 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
5208 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, SrcSize);
5211 case Intrinsic::amdgcn_inverse_ballot: {
5213 Register MaskReg =
MI.getOperand(2).getReg();
5214 unsigned MaskSize =
MRI.getType(MaskReg).getSizeInBits();
5215 unsigned MaskBank =
getRegBankID(MaskReg,
MRI, AMDGPU::SGPRRegBankID);
5216 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5217 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, MaskSize);
5220 case Intrinsic::amdgcn_bitop3: {
5222 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5223 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5224 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5225 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5228 case Intrinsic::amdgcn_s_quadmask:
5229 case Intrinsic::amdgcn_s_wqm: {
5230 Register MaskReg =
MI.getOperand(2).getReg();
5231 unsigned MaskSize =
MRI.getType(MaskReg).getSizeInBits();
5232 unsigned MaskBank =
getRegBankID(MaskReg,
MRI, AMDGPU::SGPRRegBankID);
5233 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, MaskSize);
5234 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, MaskSize);
5237 case Intrinsic::amdgcn_wave_reduce_add:
5238 case Intrinsic::amdgcn_wave_reduce_fadd:
5239 case Intrinsic::amdgcn_wave_reduce_sub:
5240 case Intrinsic::amdgcn_wave_reduce_fsub:
5241 case Intrinsic::amdgcn_wave_reduce_min:
5242 case Intrinsic::amdgcn_wave_reduce_umin:
5243 case Intrinsic::amdgcn_wave_reduce_fmin:
5244 case Intrinsic::amdgcn_wave_reduce_max:
5245 case Intrinsic::amdgcn_wave_reduce_umax:
5246 case Intrinsic::amdgcn_wave_reduce_fmax:
5247 case Intrinsic::amdgcn_wave_reduce_and:
5248 case Intrinsic::amdgcn_wave_reduce_or:
5249 case Intrinsic::amdgcn_wave_reduce_xor: {
5250 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5251 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
5252 unsigned OpSize =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
5255 OpdsMapping[2] = AMDGPU::getValueMapping(regBankID, OpSize);
5258 case Intrinsic::amdgcn_s_bitreplicate: {
5259 Register MaskReg =
MI.getOperand(2).getReg();
5260 unsigned MaskBank =
getRegBankID(MaskReg,
MRI, AMDGPU::SGPRRegBankID);
5261 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 64);
5262 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, 32);
5265 case Intrinsic::amdgcn_wave_shuffle: {
5266 unsigned OpSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5267 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5268 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5269 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5275 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD:
5276 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_D16:
5277 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_NORET:
5278 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE:
5279 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE_D16: {
5282 assert(RSrcIntrin &&
"missing RsrcIntrinsic for image intrinsic");
5289 case AMDGPU::G_AMDGPU_BVH_INTERSECT_RAY:
5290 case AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY:
5291 case AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY: {
5293 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY ||
5294 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY;
5295 unsigned NumMods = IsDualOrBVH8 ? 0 : 1;
5296 unsigned LastRegOpIdx =
MI.getNumExplicitOperands() - 1 - NumMods;
5297 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5298 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5300 OpdsMapping[1] = AMDGPU::getValueMapping(
5301 AMDGPU::VGPRRegBankID,
5302 MRI.getType(
MI.getOperand(1).getReg()).getSizeInBits());
5303 OpdsMapping[2] = AMDGPU::getValueMapping(
5304 AMDGPU::VGPRRegBankID,
5305 MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits());
5307 OpdsMapping[LastRegOpIdx] =
5309 if (LastRegOpIdx == 3) {
5311 unsigned Size =
MRI.getType(
MI.getOperand(2).getReg()).getSizeInBits();
5314 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5317 unsigned FirstSrcOpIdx = IsDualOrBVH8 ? 4 : 2;
5318 for (
unsigned I = FirstSrcOpIdx;
I < LastRegOpIdx; ++
I) {
5319 unsigned Size =
MRI.getType(
MI.getOperand(
I).getReg()).getSizeInBits();
5320 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5325 case AMDGPU::G_INTRINSIC_W_SIDE_EFFECTS:
5326 case AMDGPU::G_INTRINSIC_CONVERGENT_W_SIDE_EFFECTS: {
5329 case Intrinsic::amdgcn_s_getreg:
5330 case Intrinsic::amdgcn_s_memtime:
5331 case Intrinsic::amdgcn_s_memrealtime:
5332 case Intrinsic::amdgcn_s_get_waveid_in_workgroup:
5333 case Intrinsic::amdgcn_s_sendmsg_rtn: {
5334 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5335 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5338 case Intrinsic::amdgcn_global_atomic_fmin_num:
5339 case Intrinsic::amdgcn_global_atomic_fmax_num:
5340 case Intrinsic::amdgcn_flat_atomic_fmin_num:
5341 case Intrinsic::amdgcn_flat_atomic_fmax_num:
5342 case Intrinsic::amdgcn_global_atomic_ordered_add_b64:
5343 case Intrinsic::amdgcn_global_load_tr_b64:
5344 case Intrinsic::amdgcn_global_load_tr_b128:
5345 case Intrinsic::amdgcn_global_load_tr4_b64:
5346 case Intrinsic::amdgcn_global_load_tr6_b96:
5347 case Intrinsic::amdgcn_ds_load_tr8_b64:
5348 case Intrinsic::amdgcn_ds_load_tr16_b128:
5349 case Intrinsic::amdgcn_ds_load_tr4_b64:
5350 case Intrinsic::amdgcn_ds_load_tr6_b96:
5351 case Intrinsic::amdgcn_ds_read_tr4_b64:
5352 case Intrinsic::amdgcn_ds_read_tr6_b96:
5353 case Intrinsic::amdgcn_ds_read_tr8_b64:
5354 case Intrinsic::amdgcn_ds_read_tr16_b64:
5355 case Intrinsic::amdgcn_ds_atomic_async_barrier_arrive_b64:
5356 case Intrinsic::amdgcn_ds_atomic_barrier_arrive_rtn_b64:
5358 case Intrinsic::amdgcn_ds_ordered_add:
5359 case Intrinsic::amdgcn_ds_ordered_swap: {
5360 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5361 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5363 AMDGPU::SGPRRegBankID);
5364 OpdsMapping[2] = AMDGPU::getValueMapping(M0Bank, 32);
5365 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5368 case Intrinsic::amdgcn_ds_append:
5369 case Intrinsic::amdgcn_ds_consume: {
5370 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5371 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5375 case Intrinsic::amdgcn_exp_compr:
5376 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5377 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5379 case Intrinsic::amdgcn_exp:
5381 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5382 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5383 OpdsMapping[5] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5384 OpdsMapping[6] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5386 case Intrinsic::amdgcn_exp_row:
5387 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5388 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5389 OpdsMapping[5] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5390 OpdsMapping[6] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5393 case Intrinsic::amdgcn_s_alloc_vgpr:
5394 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 1);
5395 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
5397 case Intrinsic::amdgcn_s_sendmsg:
5398 case Intrinsic::amdgcn_s_sendmsghalt: {
5401 AMDGPU::SGPRRegBankID);
5402 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5405 case Intrinsic::amdgcn_s_setreg: {
5408 AMDGPU::SGPRRegBankID);
5409 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5412 case Intrinsic::amdgcn_s_ttracedata: {
5416 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, 32);
5419 case Intrinsic::amdgcn_end_cf: {
5421 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5424 case Intrinsic::amdgcn_else: {
5426 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5427 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, WaveSize);
5428 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, WaveSize);
5431 case Intrinsic::amdgcn_init_whole_wave:
5432 case Intrinsic::amdgcn_live_mask: {
5433 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5436 case Intrinsic::amdgcn_wqm_demote:
5437 case Intrinsic::amdgcn_kill: {
5438 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5441 case Intrinsic::amdgcn_raw_buffer_load:
5442 case Intrinsic::amdgcn_raw_ptr_buffer_load:
5443 case Intrinsic::amdgcn_raw_atomic_buffer_load:
5444 case Intrinsic::amdgcn_raw_ptr_atomic_buffer_load:
5445 case Intrinsic::amdgcn_raw_tbuffer_load:
5446 case Intrinsic::amdgcn_raw_ptr_tbuffer_load: {
5455 case Intrinsic::amdgcn_raw_buffer_load_lds:
5456 case Intrinsic::amdgcn_raw_buffer_load_async_lds:
5457 case Intrinsic::amdgcn_raw_ptr_buffer_load_lds:
5458 case Intrinsic::amdgcn_raw_ptr_buffer_load_async_lds: {
5465 case Intrinsic::amdgcn_raw_buffer_store:
5466 case Intrinsic::amdgcn_raw_ptr_buffer_store:
5467 case Intrinsic::amdgcn_raw_buffer_store_format:
5468 case Intrinsic::amdgcn_raw_ptr_buffer_store_format:
5469 case Intrinsic::amdgcn_raw_tbuffer_store:
5470 case Intrinsic::amdgcn_raw_ptr_tbuffer_store: {
5477 case Intrinsic::amdgcn_struct_buffer_load:
5478 case Intrinsic::amdgcn_struct_ptr_buffer_load:
5479 case Intrinsic::amdgcn_struct_tbuffer_load:
5480 case Intrinsic::amdgcn_struct_ptr_tbuffer_load:
5481 case Intrinsic::amdgcn_struct_atomic_buffer_load:
5482 case Intrinsic::amdgcn_struct_ptr_atomic_buffer_load: {
5490 case Intrinsic::amdgcn_struct_buffer_load_lds:
5491 case Intrinsic::amdgcn_struct_buffer_load_async_lds:
5492 case Intrinsic::amdgcn_struct_ptr_buffer_load_lds:
5493 case Intrinsic::amdgcn_struct_ptr_buffer_load_async_lds: {
5501 case Intrinsic::amdgcn_struct_buffer_store:
5502 case Intrinsic::amdgcn_struct_ptr_buffer_store:
5503 case Intrinsic::amdgcn_struct_tbuffer_store:
5504 case Intrinsic::amdgcn_struct_ptr_tbuffer_store: {
5512 case Intrinsic::amdgcn_init_exec_from_input: {
5514 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5517 case Intrinsic::amdgcn_ds_gws_init:
5518 case Intrinsic::amdgcn_ds_gws_barrier:
5519 case Intrinsic::amdgcn_ds_gws_sema_br: {
5520 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5524 AMDGPU::SGPRRegBankID);
5525 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5528 case Intrinsic::amdgcn_ds_gws_sema_v:
5529 case Intrinsic::amdgcn_ds_gws_sema_p:
5530 case Intrinsic::amdgcn_ds_gws_sema_release_all: {
5533 AMDGPU::SGPRRegBankID);
5534 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, 32);
5537 case Intrinsic::amdgcn_cluster_load_b32:
5538 case Intrinsic::amdgcn_cluster_load_b64:
5539 case Intrinsic::amdgcn_cluster_load_b128: {
5544 OpdsMapping[4] = AMDGPU::getValueMapping(M0Bank, 32);
5547 case Intrinsic::amdgcn_cluster_load_async_to_lds_b8:
5548 case Intrinsic::amdgcn_cluster_load_async_to_lds_b32:
5549 case Intrinsic::amdgcn_cluster_load_async_to_lds_b64:
5550 case Intrinsic::amdgcn_cluster_load_async_to_lds_b128: {
5555 OpdsMapping[5] = AMDGPU::getValueMapping(M0Bank, 32);
5558 case Intrinsic::amdgcn_global_store_async_from_lds_b8:
5559 case Intrinsic::amdgcn_global_store_async_from_lds_b32:
5560 case Intrinsic::amdgcn_global_store_async_from_lds_b64:
5561 case Intrinsic::amdgcn_global_store_async_from_lds_b128:
5562 case Intrinsic::amdgcn_global_load_async_to_lds_b8:
5563 case Intrinsic::amdgcn_global_load_async_to_lds_b32:
5564 case Intrinsic::amdgcn_global_load_async_to_lds_b64:
5565 case Intrinsic::amdgcn_global_load_async_to_lds_b128:
5566 case Intrinsic::amdgcn_load_to_lds:
5567 case Intrinsic::amdgcn_global_load_lds: {
5572 case Intrinsic::amdgcn_lds_direct_load: {
5573 const int M0Idx =
MI.getNumOperands() - 1;
5574 Register M0Reg =
MI.getOperand(M0Idx).getReg();
5576 unsigned DstSize =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5578 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5579 for (
int I = 2;
I != M0Idx &&
MI.getOperand(
I).
isReg(); ++
I)
5580 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5584 OpdsMapping[M0Idx] = AMDGPU::getValueMapping(M0Bank, 32);
5587 case Intrinsic::amdgcn_ds_add_gs_reg_rtn:
5588 case Intrinsic::amdgcn_ds_sub_gs_reg_rtn:
5592 case Intrinsic::amdgcn_ds_bvh_stack_rtn:
5593 case Intrinsic::amdgcn_ds_bvh_stack_push4_pop1_rtn:
5594 case Intrinsic::amdgcn_ds_bvh_stack_push8_pop1_rtn:
5595 case Intrinsic::amdgcn_ds_bvh_stack_push8_pop2_rtn: {
5608 case Intrinsic::amdgcn_s_sleep_var:
5611 case Intrinsic::amdgcn_s_barrier_join:
5612 case Intrinsic::amdgcn_s_wakeup_barrier:
5615 case Intrinsic::amdgcn_s_barrier_init:
5616 case Intrinsic::amdgcn_s_barrier_signal_var:
5620 case Intrinsic::amdgcn_s_barrier_signal_isfirst: {
5621 const unsigned ResultSize = 1;
5623 AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, ResultSize);
5626 case Intrinsic::amdgcn_s_get_barrier_state:
5627 case Intrinsic::amdgcn_s_get_named_barrier_state: {
5632 case Intrinsic::amdgcn_pops_exiting_wave_id:
5634 case Intrinsic::amdgcn_tensor_load_to_lds:
5635 case Intrinsic::amdgcn_tensor_store_from_lds: {
5638 for (
unsigned I = 1;
I <
MI.getNumOperands(); ++
I) {
5639 if (
MI.getOperand(
I).isReg()) {
5643 OpdsMapping[
I] = AMDGPU::getValueMapping(OpBank,
Size);
5648 case Intrinsic::amdgcn_s_prefetch_data: {
5653 case Intrinsic::amdgcn_flat_prefetch:
5654 case Intrinsic::amdgcn_global_prefetch:
5661 case AMDGPU::G_SELECT: {
5662 unsigned Size =
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits();
5664 AMDGPU::SGPRRegBankID);
5666 AMDGPU::SGPRRegBankID);
5667 bool SGPRSrcs = Op2Bank == AMDGPU::SGPRRegBankID &&
5668 Op3Bank == AMDGPU::SGPRRegBankID;
5670 unsigned CondBankDefault = SGPRSrcs ?
5671 AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
5674 if (CondBank == AMDGPU::SGPRRegBankID)
5675 CondBank = SGPRSrcs ? AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
5676 else if (CondBank == AMDGPU::VGPRRegBankID)
5677 CondBank = AMDGPU::VCCRegBankID;
5679 unsigned Bank = SGPRSrcs && CondBank == AMDGPU::SGPRRegBankID ?
5680 AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
5682 assert(CondBank == AMDGPU::VCCRegBankID || CondBank == AMDGPU::SGPRRegBankID);
5686 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5687 OpdsMapping[1] = AMDGPU::getValueMapping(CondBank, 1);
5688 OpdsMapping[2] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5689 OpdsMapping[3] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5691 OpdsMapping[0] = AMDGPU::getValueMapping(Bank,
Size);
5692 OpdsMapping[1] = AMDGPU::getValueMapping(CondBank, 1);
5693 OpdsMapping[2] = AMDGPU::getValueMapping(Bank,
Size);
5694 OpdsMapping[3] = AMDGPU::getValueMapping(Bank,
Size);
5700 case AMDGPU::G_SI_CALL: {
5701 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 64);
5707 for (
unsigned I = 4;
I <
MI.getNumOperands(); ++
I) {
5708 if (
MI.getOperand(
I).isReg()) {
5712 OpdsMapping[
I] = AMDGPU::getValueMapping(OpBank,
Size);
5717 case AMDGPU::G_LOAD:
5718 case AMDGPU::G_ZEXTLOAD:
5719 case AMDGPU::G_SEXTLOAD:
5722 case AMDGPU::G_ATOMICRMW_XCHG:
5723 case AMDGPU::G_ATOMICRMW_ADD:
5724 case AMDGPU::G_ATOMICRMW_SUB:
5725 case AMDGPU::G_ATOMICRMW_AND:
5726 case AMDGPU::G_ATOMICRMW_OR:
5727 case AMDGPU::G_ATOMICRMW_XOR:
5728 case AMDGPU::G_ATOMICRMW_MAX:
5729 case AMDGPU::G_ATOMICRMW_MIN:
5730 case AMDGPU::G_ATOMICRMW_UMAX:
5731 case AMDGPU::G_ATOMICRMW_UMIN:
5732 case AMDGPU::G_ATOMICRMW_FADD:
5733 case AMDGPU::G_ATOMICRMW_FMIN:
5734 case AMDGPU::G_ATOMICRMW_FMAX:
5735 case AMDGPU::G_ATOMICRMW_UINC_WRAP:
5736 case AMDGPU::G_ATOMICRMW_UDEC_WRAP:
5737 case AMDGPU::G_ATOMICRMW_USUB_COND:
5738 case AMDGPU::G_ATOMICRMW_USUB_SAT:
5739 case AMDGPU::G_AMDGPU_ATOMIC_CMPXCHG: {
5745 case AMDGPU::G_ATOMIC_CMPXCHG: {
5752 case AMDGPU::G_BRCOND: {
5754 AMDGPU::SGPRRegBankID);
5755 assert(
MRI.getType(
MI.getOperand(0).getReg()).getSizeInBits() == 1);
5756 if (Bank != AMDGPU::SGPRRegBankID)
5757 Bank = AMDGPU::VCCRegBankID;
5759 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, 1);
5762 case AMDGPU::G_INTRINSIC_FPTRUNC_ROUND:
5764 case AMDGPU::G_PREFETCH:
5767 case AMDGPU::G_AMDGPU_WHOLE_WAVE_FUNC_SETUP:
5768 case AMDGPU::G_AMDGPU_WHOLE_WAVE_FUNC_RETURN:
5769 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5771 case AMDGPU::G_AMDGPU_FLAT_LOAD_MONITOR:
5772 case AMDGPU::G_AMDGPU_GLOBAL_LOAD_MONITOR: {
5775 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5776 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, PtrSize);
5783 MI.getNumOperands());