2196 B.setInstrAndDebugLoc(
MI);
2197 unsigned Opc =
MI.getOpcode();
2200 case AMDGPU::G_CONSTANT:
2201 case AMDGPU::G_IMPLICIT_DEF: {
2209 if (DstBank == &AMDGPU::VCCRegBank)
2212 if (DefRegs.
empty())
2215 B.setInsertPt(*
MI.getParent(), ++
MI.getIterator());
2218 LLVMContext &Ctx =
B.getMF().getFunction().getContext();
2220 MI.getOperand(0).setReg(NewDstReg);
2221 if (
Opc != AMDGPU::G_IMPLICIT_DEF) {
2222 uint64_t ConstVal =
MI.getOperand(1).getCImm()->getZExtValue();
2223 MI.getOperand(1).setCImm(
2228 B.buildTrunc(DefRegs[0], NewDstReg);
2231 case AMDGPU::G_PHI: {
2240 if (DstBank == &AMDGPU::VCCRegBank) {
2247 for (
unsigned I = 1, E =
MI.getNumOperands();
I != E;
I += 2) {
2251 if (SrcBank != &AMDGPU::VCCRegBank) {
2256 MRI.
setRegBank(Copy.getReg(0), AMDGPU::VCCRegBank);
2257 MI.getOperand(
I).setReg(Copy.getReg(0));
2268 ApplyRegBankMapping ApplyBank(
B, *
this, MRI, DstBank);
2269 B.setInsertPt(
B.getMBB(),
MI);
2277 case AMDGPU::G_FCMP:
2281 case AMDGPU::G_ICMP:
2282 case AMDGPU::G_UADDO:
2283 case AMDGPU::G_USUBO:
2284 case AMDGPU::G_UADDE:
2285 case AMDGPU::G_SADDE:
2286 case AMDGPU::G_USUBE:
2287 case AMDGPU::G_SSUBE: {
2288 unsigned BoolDstOp =
2289 (
Opc == AMDGPU::G_ICMP ||
Opc == AMDGPU::G_FCMP) ? 0 : 1;
2290 Register DstReg =
MI.getOperand(BoolDstOp).getReg();
2294 if (DstBank != &AMDGPU::SGPRRegBank)
2297 const bool HasCarryIn =
MI.getNumOperands() == 5;
2303 MRI.
setRegBank(NewDstReg, AMDGPU::SGPRRegBank);
2304 MI.getOperand(BoolDstOp).setReg(NewDstReg);
2308 MRI.
setRegBank(NewSrcReg, AMDGPU::SGPRRegBank);
2309 B.buildZExt(NewSrcReg,
MI.getOperand(4).getReg());
2310 MI.getOperand(4).setReg(NewSrcReg);
2314 B.setInsertPt(*
MBB, std::next(
MI.getIterator()));
2319 if (DefRegs.
empty())
2321 B.buildTrunc(DefRegs[0], NewDstReg);
2324 case AMDGPU::G_SELECT: {
2329 if (CondRegs.
empty())
2336 if (CondBank == &AMDGPU::SGPRRegBank) {
2339 MRI.
setRegBank(NewCondReg, AMDGPU::SGPRRegBank);
2341 MI.getOperand(1).setReg(NewCondReg);
2342 B.buildZExt(NewCondReg, CondRegs[0]);
2355 if (DefRegs.
empty()) {
2360 if (Src1Regs.
empty())
2366 if (Src2Regs.
empty())
2373 auto Flags =
MI.getFlags();
2374 B.buildSelect(DefRegs[0], CondRegs[0], Src1Regs[0], Src2Regs[0], Flags);
2375 B.buildSelect(DefRegs[1], CondRegs[0], Src1Regs[1], Src2Regs[1], Flags);
2378 MI.eraseFromParent();
2381 case AMDGPU::G_BRCOND: {
2382 Register CondReg =
MI.getOperand(0).getReg();
2387 if (CondBank == &AMDGPU::SGPRRegBank) {
2390 MRI.
setRegBank(NewCondReg, AMDGPU::SGPRRegBank);
2392 MI.getOperand(0).setReg(NewCondReg);
2393 B.buildZExt(NewCondReg, CondReg);
2401 case AMDGPU::G_XOR: {
2411 if (DstBank == &AMDGPU::VCCRegBank)
2415 ApplyRegBankMapping ApplyBank(
B, *
this, MRI, DstBank);
2424 if (DstTy.
getSizeInBits() == 16 && DstBank == &AMDGPU::SGPRRegBank) {
2428 ApplyRegBankMapping ApplySALU(
B, *
this, MRI, &AMDGPU::SGPRRegBank);
2433 if (
MI.getOpcode() == AMDGPU::G_XOR &&
2454 if (DefRegs.
empty()) {
2461 (Src0Regs.
empty() || Src0Regs.
size() == 2));
2467 if (Src0Regs.
empty())
2472 if (Src1Regs.
empty())
2479 auto Flags =
MI.getFlags();
2480 B.buildInstr(
Opc, {DefRegs[0]}, {Src0Regs[0], Src1Regs[0]}, Flags);
2481 B.buildInstr(
Opc, {DefRegs[1]}, {Src0Regs[1], Src1Regs[1]}, Flags);
2484 MI.eraseFromParent();
2487 case AMDGPU::G_ABS: {
2493 if (SrcBank && SrcBank == &AMDGPU::VGPRRegBank) {
2495 ApplyRegBankMapping Apply(
B, *
this, MRI, &AMDGPU::VGPRRegBank);
2508 case AMDGPU::G_LSHR:
2509 case AMDGPU::G_ASHR:
2510 case AMDGPU::G_SMIN:
2511 case AMDGPU::G_SMAX:
2512 case AMDGPU::G_UMIN:
2513 case AMDGPU::G_UMAX: {
2520 if (!
Subtarget.hasVectorMulU64() &&
Opc == AMDGPU::G_MUL &&
2533 if (DstBank == &AMDGPU::VGPRRegBank)
2539 ApplyRegBankMapping ApplySALU(
B, *
this, MRI, &AMDGPU::SGPRRegBank);
2544 std::tie(WideSrcLo, WideSrcHi) =
2546 auto Lo =
B.buildInstr(AMDGPU::G_ABS, {
S32}, {WideSrcLo});
2547 auto Hi =
B.buildInstr(AMDGPU::G_ABS, {
S32}, {WideSrcHi});
2548 B.buildBuildVectorTrunc(DstReg, {
Lo.getReg(0),
Hi.getReg(0)});
2549 MI.eraseFromParent();
2558 std::tie(WideSrc0Lo, WideSrc0Hi)
2560 std::tie(WideSrc1Lo, WideSrc1Hi)
2562 auto Lo =
B.buildInstr(
MI.getOpcode(), {S32}, {WideSrc0Lo, WideSrc1Lo});
2563 auto Hi =
B.buildInstr(
MI.getOpcode(), {S32}, {WideSrc0Hi, WideSrc1Hi});
2564 B.buildBuildVectorTrunc(DstReg, {
Lo.getReg(0),
Hi.getReg(0)});
2565 MI.eraseFromParent();
2573 if (
Opc == AMDGPU::G_SHL ||
Opc == AMDGPU::G_LSHR ||
2574 Opc == AMDGPU::G_ASHR) {
2575 B.setInsertPt(*
MBB,
MI.getIterator());
2583 case AMDGPU::G_AMDGPU_S_MUL_I64_I32:
2584 case AMDGPU::G_AMDGPU_S_MUL_U64_U32: {
2598 Register SrcReg0 =
MI.getOperand(1).getReg();
2599 Register SrcReg1 =
MI.getOperand(2).getReg();
2602 assert(MRI.
getType(DstReg) ==
S64 &&
"This is a special case for s_mul_u64 "
2603 "that handles only 64-bit operands.");
2609 if (DstBank == &AMDGPU::SGPRRegBank) {
2610 MI.setDesc(
TII->get(AMDGPU::S_MUL_U64));
2611 MRI.
setRegClass(DstReg, &AMDGPU::SGPR_64RegClass);
2612 MRI.
setRegClass(SrcReg0, &AMDGPU::SGPR_64RegClass);
2613 MRI.
setRegClass(SrcReg1, &AMDGPU::SGPR_64RegClass);
2620 "The destination operand should be in vector registers.");
2626 B.buildTrunc(Op0L, SrcReg0);
2632 B.buildTrunc(Op1L, SrcReg1);
2634 unsigned NewOpc =
Opc == AMDGPU::G_AMDGPU_S_MUL_U64_U32
2635 ? AMDGPU::G_AMDGPU_MAD_U64_U32
2636 : AMDGPU::G_AMDGPU_MAD_I64_I32;
2640 MRI.
setRegClass(Zero64, &AMDGPU::VReg_64RegClass);
2642 MRI.
setRegClass(CarryOut, &AMDGPU::VReg_64RegClass);
2643 B.buildInstr(NewOpc, {DstReg, CarryOut}, {Op0L, Op1L, Zero64});
2644 MI.eraseFromParent();
2647 case AMDGPU::G_SEXT_INREG: {
2649 if (SrcRegs.
empty())
2653 ApplyRegBankMapping O(
B, *
this, MRI, &AMDGPU::VGPRRegBank);
2660 int Amt =
MI.getOperand(2).getImm();
2666 B.buildFreeze(DstRegs[0], SrcRegs[0]);
2668 auto Freeze =
B.buildFreeze(
S32, SrcRegs[0]);
2670 B.buildSExtInReg(DstRegs[0], Freeze, Amt);
2673 B.buildAShr(DstRegs[1], DstRegs[0],
B.buildConstant(
S32, 31));
2677 B.buildCopy(DstRegs[0], SrcRegs[0]);
2678 B.buildSExtInReg(DstRegs[1], DstRegs[0], Amt - 32);
2683 MI.eraseFromParent();
2686 case AMDGPU::G_CTPOP:
2687 case AMDGPU::G_BITREVERSE: {
2690 if (DstBank == &AMDGPU::SGPRRegBank)
2699 ApplyRegBankMapping ApplyVALU(
B, *
this, MRI, &AMDGPU::VGPRRegBank);
2708 case AMDGPU::G_AMDGPU_FFBH_U32:
2709 case AMDGPU::G_AMDGPU_FFBL_B32:
2710 case AMDGPU::G_CTLZ_ZERO_UNDEF:
2711 case AMDGPU::G_CTTZ_ZERO_UNDEF: {
2714 if (DstBank == &AMDGPU::SGPRRegBank)
2729 ApplyRegBankMapping ApplyVALU(
B, *
this, MRI, &AMDGPU::VGPRRegBank);
2731 unsigned NewOpc =
Opc == AMDGPU::G_CTLZ_ZERO_UNDEF
2732 ? (
unsigned)AMDGPU::G_AMDGPU_FFBH_U32
2733 :
Opc == AMDGPU::G_CTTZ_ZERO_UNDEF
2734 ? (
unsigned)AMDGPU::G_AMDGPU_FFBL_B32
2736 unsigned Idx = NewOpc == AMDGPU::G_AMDGPU_FFBH_U32;
2737 auto X =
B.buildInstr(NewOpc, {
S32}, {SrcRegs[Idx]});
2738 auto Y =
B.buildInstr(NewOpc, {
S32}, {SrcRegs[Idx ^ 1]});
2740 Opc == AMDGPU::G_CTLZ_ZERO_UNDEF ||
Opc == AMDGPU::G_CTTZ_ZERO_UNDEF
2742 : AMDGPU::G_UADDSAT;
2743 Y =
B.buildInstr(AddOpc, {
S32}, {
Y,
B.buildConstant(
S32, 32)});
2745 B.buildUMin(DstReg,
X,
Y);
2746 MI.eraseFromParent();
2749 case AMDGPU::G_SEXT:
2750 case AMDGPU::G_ZEXT:
2751 case AMDGPU::G_ANYEXT: {
2754 const bool Signed =
Opc == AMDGPU::G_SEXT;
2764 SrcBank != &AMDGPU::SGPRRegBank &&
2765 SrcBank != &AMDGPU::VCCRegBank &&
2769 SrcTy.getSizeInBits() <= 32) {
2775 B.buildSExtOrTrunc(DefRegs[0], SrcReg);
2776 }
else if (
Opc == AMDGPU::G_ZEXT) {
2777 B.buildZExtOrTrunc(DefRegs[0], SrcReg);
2779 B.buildAnyExtOrTrunc(DefRegs[0], SrcReg);
2784 MI.eraseFromParent();
2794 if (SrcBank == &AMDGPU::VCCRegBank) {
2801 const bool UseSel64 = DstSize > 32 &&
2802 SrcBank->
getID() == AMDGPU::SGPRRegBankID;
2806 auto True =
B.buildConstant(SelType,
Signed ? -1 : 1);
2807 auto False =
B.buildConstant(SelType, 0);
2814 B.buildSelect(DefRegs[0], SrcReg, True, False);
2816 }
else if (DstSize < 32) {
2817 auto Sel =
B.buildSelect(SelType, SrcReg, True, False);
2819 B.buildTrunc(DstReg, Sel);
2821 B.buildSelect(DstReg, SrcReg, True, False);
2824 MI.eraseFromParent();
2830 case AMDGPU::G_EXTRACT_VECTOR_ELT: {
2842 if (foldExtractEltToCmpSelect(
B,
MI, OpdMapper))
2854 unsigned ConstOffset;
2855 std::tie(BaseIdxReg, ConstOffset) =
2862 bool ShouldMoveIndexIntoLoop = IdxBank != &AMDGPU::SGPRRegBank &&
2864 ConstOffset < SrcTy.getNumElements();
2867 if (ShouldMoveIndexIntoLoop)
2868 MI.getOperand(2).setReg(BaseIdxReg);
2874 const bool NeedCopyToVGPR = DstBank == &AMDGPU::VGPRRegBank &&
2875 SrcBank == &AMDGPU::SGPRRegBank;
2876 if (DstRegs.
empty()) {
2881 if (NeedCopyToVGPR) {
2885 MI.getOperand(0).setReg(TmpReg);
2886 B.setInsertPt(*
MI.getParent(), ++
MI.getIterator());
2893 if (ShouldMoveIndexIntoLoop)
2903 auto CastSrc =
B.buildBitcast(Vec32, SrcReg);
2904 auto One =
B.buildConstant(
S32, 1);
2915 auto IdxLo =
B.buildShl(
S32, BaseIdxReg, One);
2916 auto IdxHi =
B.buildAdd(
S32, IdxLo, One);
2918 auto Extract0 =
B.buildExtractVectorElement(DstRegs[0], CastSrc, IdxLo);
2919 auto Extract1 =
B.buildExtractVectorElement(DstRegs[1], CastSrc, IdxHi);
2923 MRI.
setRegBank(One.getReg(0), AMDGPU::SGPRRegBank);
2924 MRI.
setRegBank(IdxLo.getReg(0), AMDGPU::SGPRRegBank);
2925 MRI.
setRegBank(IdxHi.getReg(0), AMDGPU::SGPRRegBank);
2929 MI.eraseFromParent();
2935 B.setInstr(*Span.
begin());
2936 MI.eraseFromParent();
2940 if (NeedCopyToVGPR) {
2944 MRI.
setRegBank(TmpReg0, AMDGPU::SGPRRegBank);
2945 MRI.
setRegBank(TmpReg1, AMDGPU::SGPRRegBank);
2947 Extract0->getOperand(0).setReg(TmpReg0);
2948 Extract1->getOperand(0).setReg(TmpReg1);
2956 if (ShouldMoveIndexIntoLoop)
2961 case AMDGPU::G_INSERT_VECTOR_ELT: {
2971 MRI.
setType(
MI.getOperand(1).getReg(), VecTy);
2973 if (foldInsertEltToCmpSelect(
B,
MI, OpdMapper))
2985 unsigned ConstOffset;
2986 std::tie(BaseIdxReg, ConstOffset) =
2993 bool ShouldMoveIndexIntoLoop = IdxBank != &AMDGPU::SGPRRegBank &&
2998 if (ShouldMoveIndexIntoLoop)
2999 MI.getOperand(3).setReg(BaseIdxReg);
3002 if (InsRegs.
empty()) {
3006 if (ShouldMoveIndexIntoLoop) {
3018 auto CastSrc =
B.buildBitcast(Vec32, SrcReg);
3019 auto One =
B.buildConstant(
S32, 1);
3028 auto IdxLo =
B.buildShl(
S32, BaseIdxReg, One);
3029 auto IdxHi =
B.buildAdd(
S32, IdxLo, One);
3031 auto InsLo =
B.buildInsertVectorElement(Vec32, CastSrc, InsRegs[0], IdxLo);
3032 auto InsHi =
B.buildInsertVectorElement(Vec32, InsLo, InsRegs[1], IdxHi);
3045 MRI.
setRegBank(One.getReg(0), AMDGPU::SGPRRegBank);
3046 MRI.
setRegBank(IdxLo.getReg(0), AMDGPU::SGPRRegBank);
3047 MRI.
setRegBank(IdxHi.getReg(0), AMDGPU::SGPRRegBank);
3052 B.setInsertPt(
B.getMBB(),
MI);
3053 B.buildBitcast(DstReg, InsHi);
3054 MI.eraseFromParent();
3058 B.setInstr(*Span.
begin());
3059 MI.eraseFromParent();
3070 B.buildBitcast(DstReg, InsHi);
3073 if (ShouldMoveIndexIntoLoop)
3078 case AMDGPU::G_AMDGPU_BUFFER_LOAD:
3079 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT:
3080 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT:
3081 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE:
3082 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE:
3083 case AMDGPU::G_AMDGPU_BUFFER_LOAD_TFE:
3084 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT_TFE:
3085 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT_TFE:
3086 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE_TFE:
3087 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE_TFE:
3088 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT:
3089 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_TFE:
3090 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_D16:
3091 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT:
3092 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT_D16:
3093 case AMDGPU::G_AMDGPU_BUFFER_STORE:
3094 case AMDGPU::G_AMDGPU_BUFFER_STORE_BYTE:
3095 case AMDGPU::G_AMDGPU_BUFFER_STORE_SHORT:
3096 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT:
3097 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT_D16:
3098 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT:
3099 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT_D16: {
3104 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SWAP:
3105 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_ADD:
3106 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB:
3107 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMIN:
3108 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMIN:
3109 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMAX:
3110 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMAX:
3111 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_AND:
3112 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_OR:
3113 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_XOR:
3114 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_INC:
3115 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_DEC:
3116 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB_CLAMP_U32:
3117 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_COND_SUB_U32:
3118 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FADD:
3119 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMIN:
3120 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMAX: {
3125 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_CMPSWAP: {
3130 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD:
3131 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_UBYTE:
3132 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SBYTE:
3133 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_USHORT:
3134 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SSHORT: {
3138 case AMDGPU::G_AMDGPU_S_BUFFER_PREFETCH:
3142 case AMDGPU::G_INTRINSIC:
3143 case AMDGPU::G_INTRINSIC_CONVERGENT: {
3145 case Intrinsic::amdgcn_readlane: {
3156 case Intrinsic::amdgcn_writelane: {
3166 case Intrinsic::amdgcn_interp_p1:
3167 case Intrinsic::amdgcn_interp_p2:
3168 case Intrinsic::amdgcn_interp_mov:
3169 case Intrinsic::amdgcn_interp_p1_f16:
3170 case Intrinsic::amdgcn_interp_p2_f16:
3171 case Intrinsic::amdgcn_lds_param_load: {
3179 case Intrinsic::amdgcn_interp_inreg_p10:
3180 case Intrinsic::amdgcn_interp_inreg_p2:
3181 case Intrinsic::amdgcn_interp_inreg_p10_f16:
3182 case Intrinsic::amdgcn_interp_inreg_p2_f16:
3183 case Intrinsic::amdgcn_interp_p10_rtz_f16:
3184 case Intrinsic::amdgcn_interp_p2_rtz_f16:
3185 case Intrinsic::amdgcn_permlane16_swap:
3186 case Intrinsic::amdgcn_permlane32_swap:
3189 case Intrinsic::amdgcn_permlane16:
3190 case Intrinsic::amdgcn_permlanex16: {
3198 case Intrinsic::amdgcn_permlane_bcast:
3199 case Intrinsic::amdgcn_permlane_up:
3200 case Intrinsic::amdgcn_permlane_down:
3201 case Intrinsic::amdgcn_permlane_xor:
3206 case Intrinsic::amdgcn_permlane_idx_gen: {
3210 case Intrinsic::amdgcn_sbfe:
3213 case Intrinsic::amdgcn_ubfe:
3216 case Intrinsic::amdgcn_inverse_ballot:
3217 case Intrinsic::amdgcn_s_bitreplicate:
3218 case Intrinsic::amdgcn_s_quadmask:
3219 case Intrinsic::amdgcn_s_wqm:
3223 case Intrinsic::amdgcn_ballot:
3229 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD:
3230 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_D16:
3231 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_NORET:
3232 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE:
3233 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE_D16: {
3243 case AMDGPU::G_AMDGPU_BVH_INTERSECT_RAY:
3244 case AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY:
3245 case AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY: {
3247 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY ||
3248 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY;
3249 unsigned NumMods = IsDualOrBVH8 ? 0 : 1;
3250 unsigned LastRegOpIdx =
MI.getNumExplicitOperands() - 1 - NumMods;
3255 case AMDGPU::G_INTRINSIC_W_SIDE_EFFECTS:
3256 case AMDGPU::G_INTRINSIC_CONVERGENT_W_SIDE_EFFECTS: {
3259 case Intrinsic::amdgcn_ds_ordered_add:
3260 case Intrinsic::amdgcn_ds_ordered_swap: {
3267 case Intrinsic::amdgcn_ds_gws_init:
3268 case Intrinsic::amdgcn_ds_gws_barrier:
3269 case Intrinsic::amdgcn_ds_gws_sema_br: {
3275 case Intrinsic::amdgcn_ds_gws_sema_v:
3276 case Intrinsic::amdgcn_ds_gws_sema_p:
3277 case Intrinsic::amdgcn_ds_gws_sema_release_all: {
3282 case Intrinsic::amdgcn_ds_append:
3283 case Intrinsic::amdgcn_ds_consume: {
3287 case Intrinsic::amdgcn_s_alloc_vgpr:
3290 case Intrinsic::amdgcn_s_sendmsg:
3291 case Intrinsic::amdgcn_s_sendmsghalt: {
3296 case Intrinsic::amdgcn_s_setreg: {
3300 case Intrinsic::amdgcn_s_ttracedata:
3303 case Intrinsic::amdgcn_raw_buffer_load_lds:
3304 case Intrinsic::amdgcn_raw_buffer_load_async_lds:
3305 case Intrinsic::amdgcn_raw_ptr_buffer_load_lds:
3306 case Intrinsic::amdgcn_raw_ptr_buffer_load_async_lds: {
3313 case Intrinsic::amdgcn_struct_buffer_load_lds:
3314 case Intrinsic::amdgcn_struct_buffer_load_async_lds:
3315 case Intrinsic::amdgcn_struct_ptr_buffer_load_lds:
3316 case Intrinsic::amdgcn_struct_ptr_buffer_load_async_lds: {
3323 case Intrinsic::amdgcn_cluster_load_async_to_lds_b8:
3324 case Intrinsic::amdgcn_cluster_load_async_to_lds_b32:
3325 case Intrinsic::amdgcn_cluster_load_async_to_lds_b64:
3326 case Intrinsic::amdgcn_cluster_load_async_to_lds_b128: {
3331 case Intrinsic::amdgcn_load_to_lds:
3332 case Intrinsic::amdgcn_load_async_to_lds:
3333 case Intrinsic::amdgcn_global_load_lds:
3334 case Intrinsic::amdgcn_global_load_async_lds: {
3339 case Intrinsic::amdgcn_lds_direct_load: {
3345 case Intrinsic::amdgcn_exp_row:
3349 case Intrinsic::amdgcn_cluster_load_b32:
3350 case Intrinsic::amdgcn_cluster_load_b64:
3351 case Intrinsic::amdgcn_cluster_load_b128: {
3356 case Intrinsic::amdgcn_s_sleep_var:
3360 case Intrinsic::amdgcn_s_barrier_join:
3361 case Intrinsic::amdgcn_s_wakeup_barrier:
3364 case Intrinsic::amdgcn_s_barrier_init:
3365 case Intrinsic::amdgcn_s_barrier_signal_var:
3369 case Intrinsic::amdgcn_s_get_barrier_state:
3370 case Intrinsic::amdgcn_s_get_named_barrier_state: {
3374 case Intrinsic::amdgcn_s_prefetch_data: {
3381 MI.eraseFromParent();
3384 case Intrinsic::amdgcn_tensor_load_to_lds:
3385 case Intrinsic::amdgcn_tensor_store_from_lds: {
3399 if (RSrcIntrin->IsImage) {
3410 case AMDGPU::G_SI_CALL: {
3421 unsigned FrameSetupOpcode = AMDGPU::ADJCALLSTACKUP;
3422 unsigned FrameDestroyOpcode = AMDGPU::ADJCALLSTACKDOWN;
3428 unsigned NonCopyInstrsLen = 0;
3434 while (Start->getOpcode() != FrameSetupOpcode) {
3436 bool IsCopy =
false;
3437 if (Start->getOpcode() == AMDGPU::COPY) {
3438 auto &Dst = Start->getOperand(0);
3441 if (Reg.isPhysical() &&
MI.readsRegister(Reg,
TRI)) {
3446 auto &Src = Start->getOperand(1);
3449 IsCopy = Info->getScratchRSrcReg() == Reg;
3457 NonCopyInstrsLen = NonCopyInstrs.
size();
3462 NonCopyInstrs.
resize(NonCopyInstrsLen);
3464 for (
auto *NonCopy :
reverse(NonCopyInstrs)) {
3465 MBB->splice(LastCopy,
MBB, NonCopy->getIterator());
3470 NonCopyInstrs.
clear();
3471 NonCopyInstrsLen = 0;
3474 while (End->getOpcode() != FrameDestroyOpcode) {
3476 bool IsCopy =
false;
3477 if (End->getOpcode() == AMDGPU::COPY) {
3478 auto &Src = End->getOperand(1);
3481 IsCopy = Reg.isPhysical() &&
MI.modifiesRegister(Reg,
TRI);
3487 NonCopyInstrsLen = NonCopyInstrs.
size();
3492 NonCopyInstrs.
resize(NonCopyInstrsLen);
3496 for (
auto *NonCopy :
reverse(NonCopyInstrs)) {
3497 MBB->splice(LastCopy,
MBB, NonCopy->getIterator());
3501 B.setInsertPt(
B.getMBB(), Start);
3505 case AMDGPU::G_AMDGPU_FLAT_LOAD_MONITOR:
3506 case AMDGPU::G_AMDGPU_GLOBAL_LOAD_MONITOR:
3507 case AMDGPU::G_LOAD:
3508 case AMDGPU::G_ZEXTLOAD:
3509 case AMDGPU::G_SEXTLOAD: {
3514 case AMDGPU::G_DYN_STACKALLOC:
3517 case AMDGPU::G_STACKRESTORE: {
3522 case AMDGPU::G_SBFX:
3525 case AMDGPU::G_UBFX:
3528 case AMDGPU::G_AMDGPU_MAD_U64_U32:
3529 case AMDGPU::G_AMDGPU_MAD_I64_I32:
3532 case AMDGPU::G_PREFETCH: {
3534 MI.eraseFromParent();
3538 unsigned PtrBank =
getRegBankID(PtrReg, MRI, AMDGPU::SGPRRegBankID);
3539 if (PtrBank == AMDGPU::VGPRRegBankID &&
3540 (!
Subtarget.hasVmemPrefInsts() || !
MI.getOperand(3).getImm())) {
3542 MI.eraseFromParent();
3550 !
MI.getOperand(3).getImm() ))) {
3551 MI.eraseFromParent();
3848 if (
MI.isCopy() ||
MI.getOpcode() == AMDGPU::G_FREEZE) {
3863 DstBank = &AMDGPU::VCCRegBank;
3866 DstBank = &AMDGPU::VCCRegBank;
3877 if (
MI.getOpcode() != AMDGPU::G_FREEZE &&
3882 unsigned OpdsMappingSize =
MI.isCopy() ? 1 : 2;
3884 OpdsMapping[0] = &ValMap;
3885 if (
MI.getOpcode() == AMDGPU::G_FREEZE)
3886 OpdsMapping[1] = &ValMap;
3893 if (
MI.isRegSequence()) {
3896 unsigned BankID = AMDGPU::SGPRRegBankID;
3898 for (
unsigned I = 1, E =
MI.getNumOperands();
I != E;
I += 2) {
3902 if (OpBank != AMDGPU::SGPRRegBankID) {
3903 BankID = AMDGPU::VGPRRegBankID;
3920 unsigned ResultBank = AMDGPU::InvalidRegBankID;
3925 ResultBank = DstBank->
getID();
3927 for (
unsigned I = 0;
I <
PHI->getNumIncomingValues(); ++
I) {
3932 if (!Bank || Bank->
getID() == AMDGPU::VGPRRegBankID) {
3933 ResultBank = AMDGPU::VGPRRegBankID;
3938 unsigned OpBank = Bank->
getID();
3942 assert(ResultBank != AMDGPU::InvalidRegBankID);
3959 switch (
MI.getOpcode()) {
3966 case AMDGPU::G_MUL: {
3972 unsigned TargetBankID = AMDGPU::InvalidRegBankID;
3973 unsigned BankLHS = AMDGPU::InvalidRegBankID;
3974 unsigned BankRHS = AMDGPU::InvalidRegBankID;
3976 TargetBankID = DstBank->
getID();
3977 if (DstBank == &AMDGPU::VCCRegBank) {
3978 TargetBankID = AMDGPU::VCCRegBankID;
3979 BankLHS = AMDGPU::VCCRegBankID;
3980 BankRHS = AMDGPU::VCCRegBankID;
3983 AMDGPU::SGPRRegBankID);
3985 AMDGPU::SGPRRegBankID);
3989 AMDGPU::VCCRegBankID);
3991 AMDGPU::VCCRegBankID);
3994 if (BankLHS == AMDGPU::VGPRRegBankID || BankRHS == AMDGPU::VGPRRegBankID) {
3995 TargetBankID = AMDGPU::VGPRRegBankID;
3996 }
else if (BankLHS == AMDGPU::VCCRegBankID || BankRHS == AMDGPU::VCCRegBankID) {
3997 TargetBankID = AMDGPU::VCCRegBankID;
3998 BankLHS = AMDGPU::VCCRegBankID;
3999 BankRHS = AMDGPU::VCCRegBankID;
4000 }
else if (BankLHS == AMDGPU::SGPRRegBankID && BankRHS == AMDGPU::SGPRRegBankID) {
4001 TargetBankID = AMDGPU::SGPRRegBankID;
4005 OpdsMapping[0] = AMDGPU::getValueMapping(TargetBankID,
Size);
4006 OpdsMapping[1] = AMDGPU::getValueMapping(BankLHS,
Size);
4007 OpdsMapping[2] = AMDGPU::getValueMapping(BankRHS,
Size);
4014 OpdsMapping[0] = getValueMappingSGPR64Only(AMDGPU::SGPRRegBankID,
Size);
4015 OpdsMapping[1] = OpdsMapping[2] = OpdsMapping[0];
4017 if (
MI.getOpcode() == AMDGPU::G_MUL &&
Subtarget.hasVectorMulU64())
4018 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4021 getValueMappingSGPR64Only(AMDGPU::VGPRRegBankID,
Size);
4023 OpdsMapping[1] = AMDGPU::getValueMapping(Bank1,
Size);
4026 OpdsMapping[2] = AMDGPU::getValueMapping(Bank2,
Size);
4034 case AMDGPU::G_PTR_ADD:
4035 case AMDGPU::G_PTRMASK:
4039 case AMDGPU::G_LSHR:
4040 case AMDGPU::G_ASHR:
4041 case AMDGPU::G_UADDO:
4042 case AMDGPU::G_USUBO:
4043 case AMDGPU::G_UADDE:
4044 case AMDGPU::G_SADDE:
4045 case AMDGPU::G_USUBE:
4046 case AMDGPU::G_SSUBE:
4048 case AMDGPU::G_SHUFFLE_VECTOR:
4049 case AMDGPU::G_SBFX:
4050 case AMDGPU::G_UBFX:
4051 case AMDGPU::G_AMDGPU_S_MUL_I64_I32:
4052 case AMDGPU::G_AMDGPU_S_MUL_U64_U32:
4056 case AMDGPU::G_SMIN:
4057 case AMDGPU::G_SMAX:
4058 case AMDGPU::G_UMIN:
4059 case AMDGPU::G_UMAX:
4068 case AMDGPU::G_FADD:
4069 case AMDGPU::G_FSUB:
4070 case AMDGPU::G_FMUL:
4072 case AMDGPU::G_FFLOOR:
4073 case AMDGPU::G_FCEIL:
4074 case AMDGPU::G_INTRINSIC_ROUNDEVEN:
4075 case AMDGPU::G_FMINNUM:
4076 case AMDGPU::G_FMAXNUM:
4077 case AMDGPU::G_FMINIMUMNUM:
4078 case AMDGPU::G_FMAXIMUMNUM:
4079 case AMDGPU::G_INTRINSIC_TRUNC:
4080 case AMDGPU::G_STRICT_FADD:
4081 case AMDGPU::G_STRICT_FSUB:
4082 case AMDGPU::G_STRICT_FMUL:
4083 case AMDGPU::G_STRICT_FMA: {
4085 unsigned Size = Ty.getSizeInBits();
4086 if (
Subtarget.hasSALUFloatInsts() && Ty.isScalar() &&
4091 case AMDGPU::G_FMINIMUM:
4092 case AMDGPU::G_FMAXIMUM: {
4094 unsigned Size = Ty.getSizeInBits();
4095 if (
Subtarget.hasSALUMinimumMaximumInsts() && Ty.isScalar() &&
4100 case AMDGPU::G_FPTOSI:
4101 case AMDGPU::G_FPTOUI:
4102 case AMDGPU::G_FPTOSI_SAT:
4103 case AMDGPU::G_FPTOUI_SAT:
4104 case AMDGPU::G_SITOFP:
4105 case AMDGPU::G_UITOFP: {
4108 if (
Subtarget.hasSALUFloatInsts() && SizeDst == 32 && SizeSrc == 32 &&
4113 case AMDGPU::G_FPTRUNC:
4114 case AMDGPU::G_FPEXT: {
4117 if (
Subtarget.hasSALUFloatInsts() && SizeDst != 64 && SizeSrc != 64 &&
4122 case AMDGPU::G_FSQRT:
4123 case AMDGPU::G_FEXP2:
4124 case AMDGPU::G_FLOG2: {
4131 case AMDGPU::G_SADDSAT:
4132 case AMDGPU::G_SSUBSAT:
4133 case AMDGPU::G_UADDSAT:
4134 case AMDGPU::G_USUBSAT:
4135 case AMDGPU::G_FMAD:
4136 case AMDGPU::G_FLDEXP:
4137 case AMDGPU::G_FMINNUM_IEEE:
4138 case AMDGPU::G_FMAXNUM_IEEE:
4139 case AMDGPU::G_FCANONICALIZE:
4140 case AMDGPU::G_STRICT_FLDEXP:
4141 case AMDGPU::G_BSWAP:
4142 case AMDGPU::G_FSHR:
4143 case AMDGPU::G_AMDGPU_FMIN_LEGACY:
4144 case AMDGPU::G_AMDGPU_FMAX_LEGACY:
4145 case AMDGPU::G_AMDGPU_RCP_IFLAG:
4146 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE0:
4147 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE1:
4148 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE2:
4149 case AMDGPU::G_AMDGPU_CVT_F32_UBYTE3:
4150 case AMDGPU::G_AMDGPU_CVT_PK_I16_I32:
4151 case AMDGPU::G_AMDGPU_SMED3:
4152 case AMDGPU::G_AMDGPU_FMED3:
4154 case AMDGPU::G_UMULH:
4155 case AMDGPU::G_SMULH: {
4160 case AMDGPU::G_AMDGPU_MAD_U64_U32:
4161 case AMDGPU::G_AMDGPU_MAD_I64_I32: {
4170 bool AllSalu =
true;
4171 bool MulSalu =
true;
4172 for (
unsigned i = 0; i < 5; ++i) {
4175 if (Bank->getID() != AMDGPU::SGPRRegBankID) {
4177 if (i == 2 || i == 3) {
4191 if (!MulSalu ||
Subtarget.hasFullRate64Ops())
4195 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 64);
4196 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4197 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4198 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4199 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 64);
4202 case AMDGPU::G_IMPLICIT_DEF: {
4204 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4207 case AMDGPU::G_FCONSTANT:
4208 case AMDGPU::G_CONSTANT:
4209 case AMDGPU::G_GLOBAL_VALUE:
4210 case AMDGPU::G_FRAME_INDEX:
4211 case AMDGPU::G_BLOCK_ADDR:
4212 case AMDGPU::G_READSTEADYCOUNTER:
4213 case AMDGPU::G_READCYCLECOUNTER: {
4215 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4218 case AMDGPU::G_DYN_STACKALLOC: {
4220 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4221 unsigned SrcBankID =
getRegBankID(
MI.getOperand(1).getReg(), MRI);
4222 OpdsMapping[1] = AMDGPU::getValueMapping(SrcBankID, 32);
4225 case AMDGPU::G_AMDGPU_WAVE_ADDRESS: {
4230 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4231 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
4234 case AMDGPU::G_INSERT: {
4239 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, DstSize);
4240 OpdsMapping[1] = AMDGPU::getValueMapping(BankID, SrcSize);
4241 OpdsMapping[2] = AMDGPU::getValueMapping(BankID, EltSize);
4242 OpdsMapping[3] =
nullptr;
4245 case AMDGPU::G_EXTRACT: {
4249 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, DstSize);
4250 OpdsMapping[1] = AMDGPU::getValueMapping(BankID, SrcSize);
4251 OpdsMapping[2] =
nullptr;
4254 case AMDGPU::G_BUILD_VECTOR:
4255 case AMDGPU::G_BUILD_VECTOR_TRUNC: {
4260 unsigned Src0BankID =
getRegBankID(
MI.getOperand(1).getReg(), MRI);
4261 unsigned Src1BankID =
getRegBankID(
MI.getOperand(2).getReg(), MRI);
4262 unsigned DstBankID =
regBankUnion(Src0BankID, Src1BankID);
4264 OpdsMapping[0] = AMDGPU::getValueMapping(DstBankID, DstSize);
4265 OpdsMapping[1] = AMDGPU::getValueMapping(Src0BankID, SrcSize);
4266 OpdsMapping[2] = AMDGPU::getValueMapping(Src1BankID, SrcSize);
4272 case AMDGPU::G_MERGE_VALUES:
4273 case AMDGPU::G_CONCAT_VECTORS: {
4278 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, DstSize);
4280 for (
unsigned i = 1, e =
MI.getNumOperands(); i != e; ++i)
4281 OpdsMapping[i] = AMDGPU::getValueMapping(Bank, SrcSize);
4284 case AMDGPU::G_BITREVERSE:
4285 case AMDGPU::G_BITCAST:
4286 case AMDGPU::G_INTTOPTR:
4287 case AMDGPU::G_PTRTOINT:
4288 case AMDGPU::G_FABS:
4289 case AMDGPU::G_FNEG: {
4292 OpdsMapping[0] = OpdsMapping[1] = AMDGPU::getValueMapping(BankID,
Size);
4295 case AMDGPU::G_AMDGPU_FFBH_U32:
4296 case AMDGPU::G_AMDGPU_FFBL_B32:
4297 case AMDGPU::G_CTLZ_ZERO_UNDEF:
4298 case AMDGPU::G_CTTZ_ZERO_UNDEF: {
4301 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, 32);
4302 OpdsMapping[1] = AMDGPU::getValueMappingSGPR64Only(BankID,
Size);
4305 case AMDGPU::G_CTPOP: {
4308 OpdsMapping[0] = AMDGPU::getValueMapping(BankID, 32);
4313 OpdsMapping[1] = AMDGPU::getValueMapping(BankID,
Size);
4316 case AMDGPU::G_TRUNC: {
4322 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, DstSize);
4323 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, SrcSize);
4326 case AMDGPU::G_ZEXT:
4327 case AMDGPU::G_SEXT:
4328 case AMDGPU::G_ANYEXT:
4329 case AMDGPU::G_SEXT_INREG: {
4338 switch (SrcBank->
getID()) {
4339 case AMDGPU::SGPRRegBankID:
4340 DstBank = AMDGPU::SGPRRegBankID;
4343 DstBank = AMDGPU::VGPRRegBankID;
4349 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(DstBank, DstSize);
4350 OpdsMapping[1] = AMDGPU::getValueMappingSGPR64Only(SrcBank->
getID(),
4354 case AMDGPU::G_IS_FPCLASS: {
4358 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, DstSize);
4359 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4362 case AMDGPU::G_STORE: {
4369 AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
4370 OpdsMapping[0] = ValMapping;
4374 case AMDGPU::G_ICMP:
4375 case AMDGPU::G_FCMP: {
4381 AMDGPU::SGPRRegBankID);
4385 auto canUseSCCICMP = [&]() {
4388 return Size == 32 ||
4393 auto canUseSCCFCMP = [&]() {
4397 bool isICMP =
MI.getOpcode() == AMDGPU::G_ICMP;
4398 bool CanUseSCC = DstBank == AMDGPU::SGPRRegBankID &&
4399 Op2Bank == AMDGPU::SGPRRegBankID &&
4400 Op3Bank == AMDGPU::SGPRRegBankID &&
4401 (isICMP ? canUseSCCICMP() : canUseSCCFCMP());
4403 DstBank = CanUseSCC ? AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
4404 unsigned SrcBank = CanUseSCC ? AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
4408 const unsigned ResultSize = 1;
4410 OpdsMapping[0] = AMDGPU::getValueMapping(DstBank, ResultSize);
4411 OpdsMapping[1] =
nullptr;
4412 OpdsMapping[2] = AMDGPU::getValueMapping(SrcBank,
Size);
4413 OpdsMapping[3] = AMDGPU::getValueMapping(SrcBank,
Size);
4416 case AMDGPU::G_EXTRACT_VECTOR_ELT: {
4418 unsigned SrcBankID =
getRegBankID(
MI.getOperand(1).getReg(), MRI);
4423 unsigned OutputBankID =
regBankUnion(SrcBankID, IdxBank);
4425 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(OutputBankID, DstSize);
4426 OpdsMapping[1] = AMDGPU::getValueMapping(SrcBankID, SrcSize);
4429 OpdsMapping[2] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4432 case AMDGPU::G_INSERT_VECTOR_ELT: {
4434 AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
4439 unsigned InsertEltBankID =
getRegBankID(
MI.getOperand(2).getReg(), MRI);
4440 unsigned IdxBankID =
getRegBankID(
MI.getOperand(3).getReg(), MRI);
4442 OpdsMapping[0] = AMDGPU::getValueMapping(OutputBankID, VecSize);
4443 OpdsMapping[1] = AMDGPU::getValueMapping(OutputBankID, VecSize);
4447 if (InsertSize == 64 && OutputBankID == AMDGPU::VGPRRegBankID) {
4448 OpdsMapping[2] = AMDGPU::getValueMappingSplit64(InsertEltBankID,
4451 assert(InsertSize == 32 || InsertSize == 64);
4452 OpdsMapping[2] = AMDGPU::getValueMapping(InsertEltBankID, InsertSize);
4456 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBankID, IdxSize);
4459 case AMDGPU::G_UNMERGE_VALUES: {
4464 for (
unsigned i = 0, e =
MI.getNumOperands(); i != e; ++i) {
4466 OpdsMapping[i] = AMDGPU::getValueMapping(Bank,
Size);
4470 case AMDGPU::G_AMDGPU_BUFFER_LOAD:
4471 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE:
4472 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE:
4473 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT:
4474 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT:
4475 case AMDGPU::G_AMDGPU_BUFFER_LOAD_TFE:
4476 case AMDGPU::G_AMDGPU_BUFFER_LOAD_UBYTE_TFE:
4477 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SBYTE_TFE:
4478 case AMDGPU::G_AMDGPU_BUFFER_LOAD_USHORT_TFE:
4479 case AMDGPU::G_AMDGPU_BUFFER_LOAD_SSHORT_TFE:
4480 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT:
4481 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_TFE:
4482 case AMDGPU::G_AMDGPU_BUFFER_LOAD_FORMAT_D16:
4483 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT:
4484 case AMDGPU::G_AMDGPU_TBUFFER_LOAD_FORMAT_D16:
4485 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT:
4486 case AMDGPU::G_AMDGPU_TBUFFER_STORE_FORMAT_D16:
4487 case AMDGPU::G_AMDGPU_BUFFER_STORE:
4488 case AMDGPU::G_AMDGPU_BUFFER_STORE_BYTE:
4489 case AMDGPU::G_AMDGPU_BUFFER_STORE_SHORT:
4490 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT:
4491 case AMDGPU::G_AMDGPU_BUFFER_STORE_FORMAT_D16: {
4510 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SWAP:
4511 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_ADD:
4512 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB:
4513 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMIN:
4514 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMIN:
4515 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SMAX:
4516 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_UMAX:
4517 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_AND:
4518 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_OR:
4519 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_XOR:
4520 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_INC:
4521 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_DEC:
4522 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_SUB_CLAMP_U32:
4523 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_COND_SUB_U32:
4524 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FADD:
4525 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMIN:
4526 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_FMAX: {
4549 case AMDGPU::G_AMDGPU_BUFFER_ATOMIC_CMPSWAP: {
4575 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD:
4576 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_UBYTE:
4577 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SBYTE:
4578 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_USHORT:
4579 case AMDGPU::G_AMDGPU_S_BUFFER_LOAD_SSHORT: {
4587 unsigned RSrcBank = OpdsMapping[1]->BreakDown[0].RegBank->getID();
4588 unsigned OffsetBank = OpdsMapping[2]->BreakDown[0].RegBank->getID();
4589 unsigned ResultBank =
regBankUnion(RSrcBank, OffsetBank);
4592 OpdsMapping[0] = AMDGPU::getValueMapping(ResultBank, Size0);
4595 case AMDGPU::G_AMDGPU_S_BUFFER_PREFETCH:
4599 case AMDGPU::G_AMDGPU_SPONENTRY: {
4601 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4604 case AMDGPU::G_INTRINSIC:
4605 case AMDGPU::G_INTRINSIC_CONVERGENT: {
4609 case Intrinsic::amdgcn_div_fmas:
4610 case Intrinsic::amdgcn_div_fixup:
4611 case Intrinsic::amdgcn_trig_preop:
4612 case Intrinsic::amdgcn_sin:
4613 case Intrinsic::amdgcn_cos:
4614 case Intrinsic::amdgcn_log_clamp:
4615 case Intrinsic::amdgcn_rcp_legacy:
4616 case Intrinsic::amdgcn_rsq_legacy:
4617 case Intrinsic::amdgcn_rsq_clamp:
4618 case Intrinsic::amdgcn_tanh:
4619 case Intrinsic::amdgcn_fmul_legacy:
4620 case Intrinsic::amdgcn_fma_legacy:
4621 case Intrinsic::amdgcn_frexp_mant:
4622 case Intrinsic::amdgcn_frexp_exp:
4623 case Intrinsic::amdgcn_fract:
4624 case Intrinsic::amdgcn_cvt_pknorm_i16:
4625 case Intrinsic::amdgcn_cvt_pknorm_u16:
4626 case Intrinsic::amdgcn_cvt_pk_i16:
4627 case Intrinsic::amdgcn_cvt_pk_u16:
4628 case Intrinsic::amdgcn_cvt_sr_pk_f16_f32:
4629 case Intrinsic::amdgcn_cvt_sr_pk_bf16_f32:
4630 case Intrinsic::amdgcn_cvt_pk_f16_fp8:
4631 case Intrinsic::amdgcn_cvt_pk_f16_bf8:
4632 case Intrinsic::amdgcn_cvt_pk_fp8_f16:
4633 case Intrinsic::amdgcn_cvt_pk_bf8_f16:
4634 case Intrinsic::amdgcn_cvt_sr_fp8_f16:
4635 case Intrinsic::amdgcn_cvt_sr_bf8_f16:
4636 case Intrinsic::amdgcn_cvt_scale_pk8_f16_fp8:
4637 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_fp8:
4638 case Intrinsic::amdgcn_cvt_scale_pk8_f16_bf8:
4639 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_bf8:
4640 case Intrinsic::amdgcn_cvt_scale_pk8_f16_fp4:
4641 case Intrinsic::amdgcn_cvt_scale_pk8_bf16_fp4:
4642 case Intrinsic::amdgcn_cvt_scale_pk8_f32_fp8:
4643 case Intrinsic::amdgcn_cvt_scale_pk8_f32_bf8:
4644 case Intrinsic::amdgcn_cvt_scale_pk8_f32_fp4:
4645 case Intrinsic::amdgcn_cvt_scale_pk16_f16_fp6:
4646 case Intrinsic::amdgcn_cvt_scale_pk16_bf16_fp6:
4647 case Intrinsic::amdgcn_cvt_scale_pk16_f16_bf6:
4648 case Intrinsic::amdgcn_cvt_scale_pk16_bf16_bf6:
4649 case Intrinsic::amdgcn_cvt_scale_pk16_f32_fp6:
4650 case Intrinsic::amdgcn_cvt_scale_pk16_f32_bf6:
4651 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_bf16:
4652 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_bf16:
4653 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_f16:
4654 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_f16:
4655 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp8_f32:
4656 case Intrinsic::amdgcn_cvt_scalef32_pk8_bf8_f32:
4657 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_f32:
4658 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_f16:
4659 case Intrinsic::amdgcn_cvt_scalef32_pk8_fp4_bf16:
4660 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_f32:
4661 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_f32:
4662 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_f16:
4663 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_f16:
4664 case Intrinsic::amdgcn_cvt_scalef32_pk16_fp6_bf16:
4665 case Intrinsic::amdgcn_cvt_scalef32_pk16_bf6_bf16:
4666 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_bf16:
4667 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_bf16:
4668 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_f16:
4669 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_f16:
4670 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp8_f32:
4671 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_bf8_f32:
4672 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_f32:
4673 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_f16:
4674 case Intrinsic::amdgcn_cvt_scalef32_sr_pk8_fp4_bf16:
4675 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_f32:
4676 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_f32:
4677 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_f16:
4678 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_f16:
4679 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_fp6_bf16:
4680 case Intrinsic::amdgcn_cvt_scalef32_sr_pk16_bf6_bf16:
4681 case Intrinsic::amdgcn_sat_pk4_i4_i8:
4682 case Intrinsic::amdgcn_sat_pk4_u4_u8:
4683 case Intrinsic::amdgcn_fmed3:
4684 case Intrinsic::amdgcn_cubeid:
4685 case Intrinsic::amdgcn_cubema:
4686 case Intrinsic::amdgcn_cubesc:
4687 case Intrinsic::amdgcn_cubetc:
4688 case Intrinsic::amdgcn_sffbh:
4689 case Intrinsic::amdgcn_fmad_ftz:
4690 case Intrinsic::amdgcn_mbcnt_lo:
4691 case Intrinsic::amdgcn_mbcnt_hi:
4692 case Intrinsic::amdgcn_mul_u24:
4693 case Intrinsic::amdgcn_mul_i24:
4694 case Intrinsic::amdgcn_mulhi_u24:
4695 case Intrinsic::amdgcn_mulhi_i24:
4696 case Intrinsic::amdgcn_lerp:
4697 case Intrinsic::amdgcn_sad_u8:
4698 case Intrinsic::amdgcn_msad_u8:
4699 case Intrinsic::amdgcn_sad_hi_u8:
4700 case Intrinsic::amdgcn_sad_u16:
4701 case Intrinsic::amdgcn_qsad_pk_u16_u8:
4702 case Intrinsic::amdgcn_mqsad_pk_u16_u8:
4703 case Intrinsic::amdgcn_mqsad_u32_u8:
4704 case Intrinsic::amdgcn_cvt_pk_u8_f32:
4705 case Intrinsic::amdgcn_alignbyte:
4706 case Intrinsic::amdgcn_perm:
4707 case Intrinsic::amdgcn_prng_b32:
4708 case Intrinsic::amdgcn_fdot2:
4709 case Intrinsic::amdgcn_sdot2:
4710 case Intrinsic::amdgcn_udot2:
4711 case Intrinsic::amdgcn_sdot4:
4712 case Intrinsic::amdgcn_udot4:
4713 case Intrinsic::amdgcn_sdot8:
4714 case Intrinsic::amdgcn_udot8:
4715 case Intrinsic::amdgcn_fdot2_bf16_bf16:
4716 case Intrinsic::amdgcn_fdot2_f16_f16:
4717 case Intrinsic::amdgcn_fdot2_f32_bf16:
4718 case Intrinsic::amdgcn_fdot2c_f32_bf16:
4719 case Intrinsic::amdgcn_sudot4:
4720 case Intrinsic::amdgcn_sudot8:
4721 case Intrinsic::amdgcn_dot4_f32_fp8_bf8:
4722 case Intrinsic::amdgcn_dot4_f32_bf8_fp8:
4723 case Intrinsic::amdgcn_dot4_f32_fp8_fp8:
4724 case Intrinsic::amdgcn_dot4_f32_bf8_bf8:
4725 case Intrinsic::amdgcn_cvt_f32_fp8:
4726 case Intrinsic::amdgcn_cvt_f32_fp8_e5m3:
4727 case Intrinsic::amdgcn_cvt_f32_bf8:
4728 case Intrinsic::amdgcn_cvt_off_f32_i4:
4729 case Intrinsic::amdgcn_cvt_pk_f32_fp8:
4730 case Intrinsic::amdgcn_cvt_pk_f32_bf8:
4731 case Intrinsic::amdgcn_cvt_pk_fp8_f32:
4732 case Intrinsic::amdgcn_cvt_pk_fp8_f32_e5m3:
4733 case Intrinsic::amdgcn_cvt_pk_bf8_f32:
4734 case Intrinsic::amdgcn_cvt_sr_fp8_f32:
4735 case Intrinsic::amdgcn_cvt_sr_fp8_f32_e5m3:
4736 case Intrinsic::amdgcn_cvt_sr_bf8_f32:
4737 case Intrinsic::amdgcn_cvt_sr_bf16_f32:
4738 case Intrinsic::amdgcn_cvt_sr_f16_f32:
4739 case Intrinsic::amdgcn_cvt_f16_fp8:
4740 case Intrinsic::amdgcn_cvt_f16_bf8:
4741 case Intrinsic::amdgcn_cvt_scalef32_pk32_fp6_f16:
4742 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf6_f16:
4743 case Intrinsic::amdgcn_cvt_scalef32_pk32_fp6_bf16:
4744 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf6_bf16:
4745 case Intrinsic::amdgcn_cvt_scalef32_f16_fp8:
4746 case Intrinsic::amdgcn_cvt_scalef32_f16_bf8:
4747 case Intrinsic::amdgcn_cvt_scalef32_f32_fp8:
4748 case Intrinsic::amdgcn_cvt_scalef32_f32_bf8:
4749 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_f32:
4750 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_f32:
4751 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_fp8:
4752 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_bf8:
4753 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_f16:
4754 case Intrinsic::amdgcn_cvt_scalef32_pk_fp8_bf16:
4755 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_f16:
4756 case Intrinsic::amdgcn_cvt_scalef32_pk_bf8_bf16:
4757 case Intrinsic::amdgcn_cvt_scalef32_pk_f32_fp4:
4758 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_f32:
4759 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_fp4:
4760 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_fp4:
4761 case Intrinsic::amdgcn_cvt_scalef32_pk32_f32_fp6:
4762 case Intrinsic::amdgcn_cvt_scalef32_pk32_f32_bf6:
4763 case Intrinsic::amdgcn_cvt_scalef32_pk32_f16_bf6:
4764 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf16_bf6:
4765 case Intrinsic::amdgcn_cvt_scalef32_pk32_f16_fp6:
4766 case Intrinsic::amdgcn_cvt_scalef32_pk32_bf16_fp6:
4767 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_bf8:
4768 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_bf8:
4769 case Intrinsic::amdgcn_cvt_scalef32_pk_f16_fp8:
4770 case Intrinsic::amdgcn_cvt_scalef32_pk_bf16_fp8:
4771 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_f16:
4772 case Intrinsic::amdgcn_cvt_scalef32_pk_fp4_bf16:
4773 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_f16:
4774 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_bf16:
4775 case Intrinsic::amdgcn_cvt_scalef32_sr_pk_fp4_f32:
4776 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_bf16:
4777 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_f16:
4778 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_bf6_f32:
4779 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_bf16:
4780 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_f16:
4781 case Intrinsic::amdgcn_cvt_scalef32_sr_pk32_fp6_f32:
4782 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_bf16:
4783 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_f16:
4784 case Intrinsic::amdgcn_cvt_scalef32_sr_bf8_f32:
4785 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_bf16:
4786 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_f16:
4787 case Intrinsic::amdgcn_cvt_scalef32_sr_fp8_f32:
4788 case Intrinsic::amdgcn_ashr_pk_i8_i32:
4789 case Intrinsic::amdgcn_ashr_pk_u8_i32:
4790 case Intrinsic::amdgcn_cvt_scalef32_2xpk16_fp6_f32:
4791 case Intrinsic::amdgcn_cvt_scalef32_2xpk16_bf6_f32:
4792 case Intrinsic::amdgcn_wmma_bf16_16x16x16_bf16:
4793 case Intrinsic::amdgcn_wmma_f16_16x16x16_f16:
4794 case Intrinsic::amdgcn_wmma_bf16_16x16x16_bf16_tied:
4795 case Intrinsic::amdgcn_wmma_f16_16x16x16_f16_tied:
4796 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf16:
4797 case Intrinsic::amdgcn_wmma_f32_16x16x16_f16:
4798 case Intrinsic::amdgcn_wmma_i32_16x16x16_iu4:
4799 case Intrinsic::amdgcn_wmma_i32_16x16x16_iu8:
4800 case Intrinsic::amdgcn_wmma_f32_16x16x16_fp8_fp8:
4801 case Intrinsic::amdgcn_wmma_f32_16x16x16_fp8_bf8:
4802 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf8_fp8:
4803 case Intrinsic::amdgcn_wmma_f32_16x16x16_bf8_bf8:
4804 case Intrinsic::amdgcn_wmma_i32_16x16x32_iu4:
4805 case Intrinsic::amdgcn_swmmac_f32_16x16x32_f16:
4806 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf16:
4807 case Intrinsic::amdgcn_swmmac_f16_16x16x32_f16:
4808 case Intrinsic::amdgcn_swmmac_bf16_16x16x32_bf16:
4809 case Intrinsic::amdgcn_swmmac_i32_16x16x32_iu8:
4810 case Intrinsic::amdgcn_swmmac_i32_16x16x32_iu4:
4811 case Intrinsic::amdgcn_swmmac_i32_16x16x64_iu4:
4812 case Intrinsic::amdgcn_swmmac_f32_16x16x32_fp8_fp8:
4813 case Intrinsic::amdgcn_swmmac_f32_16x16x32_fp8_bf8:
4814 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf8_fp8:
4815 case Intrinsic::amdgcn_swmmac_f32_16x16x32_bf8_bf8:
4816 case Intrinsic::amdgcn_wmma_f32_16x16x4_f32:
4817 case Intrinsic::amdgcn_wmma_f32_16x16x32_bf16:
4818 case Intrinsic::amdgcn_wmma_f32_16x16x32_f16:
4819 case Intrinsic::amdgcn_wmma_f16_16x16x32_f16:
4820 case Intrinsic::amdgcn_wmma_bf16_16x16x32_bf16:
4821 case Intrinsic::amdgcn_wmma_bf16f32_16x16x32_bf16:
4822 case Intrinsic::amdgcn_wmma_f32_16x16x64_fp8_fp8:
4823 case Intrinsic::amdgcn_wmma_f32_16x16x64_fp8_bf8:
4824 case Intrinsic::amdgcn_wmma_f32_16x16x64_bf8_fp8:
4825 case Intrinsic::amdgcn_wmma_f32_16x16x64_bf8_bf8:
4826 case Intrinsic::amdgcn_wmma_f16_16x16x64_fp8_fp8:
4827 case Intrinsic::amdgcn_wmma_f16_16x16x64_fp8_bf8:
4828 case Intrinsic::amdgcn_wmma_f16_16x16x64_bf8_fp8:
4829 case Intrinsic::amdgcn_wmma_f16_16x16x64_bf8_bf8:
4830 case Intrinsic::amdgcn_wmma_f16_16x16x128_fp8_fp8:
4831 case Intrinsic::amdgcn_wmma_f16_16x16x128_fp8_bf8:
4832 case Intrinsic::amdgcn_wmma_f16_16x16x128_bf8_fp8:
4833 case Intrinsic::amdgcn_wmma_f16_16x16x128_bf8_bf8:
4834 case Intrinsic::amdgcn_wmma_f32_16x16x128_fp8_fp8:
4835 case Intrinsic::amdgcn_wmma_f32_16x16x128_fp8_bf8:
4836 case Intrinsic::amdgcn_wmma_f32_16x16x128_bf8_fp8:
4837 case Intrinsic::amdgcn_wmma_f32_16x16x128_bf8_bf8:
4838 case Intrinsic::amdgcn_wmma_i32_16x16x64_iu8:
4839 case Intrinsic::amdgcn_wmma_f32_16x16x128_f8f6f4:
4840 case Intrinsic::amdgcn_wmma_scale_f32_16x16x128_f8f6f4:
4841 case Intrinsic::amdgcn_wmma_scale16_f32_16x16x128_f8f6f4:
4842 case Intrinsic::amdgcn_wmma_f32_32x16x128_f4:
4843 case Intrinsic::amdgcn_wmma_scale_f32_32x16x128_f4:
4844 case Intrinsic::amdgcn_wmma_scale16_f32_32x16x128_f4:
4845 case Intrinsic::amdgcn_swmmac_f16_16x16x64_f16:
4846 case Intrinsic::amdgcn_swmmac_bf16_16x16x64_bf16:
4847 case Intrinsic::amdgcn_swmmac_f32_16x16x64_bf16:
4848 case Intrinsic::amdgcn_swmmac_bf16f32_16x16x64_bf16:
4849 case Intrinsic::amdgcn_swmmac_f32_16x16x64_f16:
4850 case Intrinsic::amdgcn_swmmac_f32_16x16x128_fp8_fp8:
4851 case Intrinsic::amdgcn_swmmac_f32_16x16x128_fp8_bf8:
4852 case Intrinsic::amdgcn_swmmac_f32_16x16x128_bf8_fp8:
4853 case Intrinsic::amdgcn_swmmac_f32_16x16x128_bf8_bf8:
4854 case Intrinsic::amdgcn_swmmac_f16_16x16x128_fp8_fp8:
4855 case Intrinsic::amdgcn_swmmac_f16_16x16x128_fp8_bf8:
4856 case Intrinsic::amdgcn_swmmac_f16_16x16x128_bf8_fp8:
4857 case Intrinsic::amdgcn_swmmac_f16_16x16x128_bf8_bf8:
4858 case Intrinsic::amdgcn_swmmac_i32_16x16x128_iu8:
4859 case Intrinsic::amdgcn_perm_pk16_b4_u4:
4860 case Intrinsic::amdgcn_perm_pk16_b6_u4:
4861 case Intrinsic::amdgcn_perm_pk16_b8_u4:
4862 case Intrinsic::amdgcn_add_max_i32:
4863 case Intrinsic::amdgcn_add_max_u32:
4864 case Intrinsic::amdgcn_add_min_i32:
4865 case Intrinsic::amdgcn_add_min_u32:
4866 case Intrinsic::amdgcn_pk_add_max_i16:
4867 case Intrinsic::amdgcn_pk_add_max_u16:
4868 case Intrinsic::amdgcn_pk_add_min_i16:
4869 case Intrinsic::amdgcn_pk_add_min_u16:
4871 case Intrinsic::amdgcn_log:
4872 case Intrinsic::amdgcn_exp2:
4873 case Intrinsic::amdgcn_rcp:
4874 case Intrinsic::amdgcn_rsq:
4875 case Intrinsic::amdgcn_sqrt: {
4882 case Intrinsic::amdgcn_sbfe:
4883 case Intrinsic::amdgcn_ubfe:
4887 case Intrinsic::amdgcn_ds_swizzle:
4888 case Intrinsic::amdgcn_ds_permute:
4889 case Intrinsic::amdgcn_ds_bpermute:
4890 case Intrinsic::amdgcn_update_dpp:
4891 case Intrinsic::amdgcn_mov_dpp8:
4892 case Intrinsic::amdgcn_mov_dpp:
4893 case Intrinsic::amdgcn_strict_wwm:
4894 case Intrinsic::amdgcn_wwm:
4895 case Intrinsic::amdgcn_strict_wqm:
4896 case Intrinsic::amdgcn_wqm:
4897 case Intrinsic::amdgcn_softwqm:
4898 case Intrinsic::amdgcn_set_inactive:
4899 case Intrinsic::amdgcn_set_inactive_chain_arg:
4900 case Intrinsic::amdgcn_permlane64:
4901 case Intrinsic::amdgcn_ds_bpermute_fi_b32:
4903 case Intrinsic::amdgcn_cvt_pkrtz:
4907 case Intrinsic::amdgcn_kernarg_segment_ptr:
4908 case Intrinsic::amdgcn_s_getpc:
4909 case Intrinsic::amdgcn_groupstaticsize:
4910 case Intrinsic::amdgcn_reloc_constant:
4911 case Intrinsic::returnaddress: {
4913 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4916 case Intrinsic::amdgcn_wqm_vote: {
4918 OpdsMapping[0] = OpdsMapping[2]
4919 = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID,
Size);
4922 case Intrinsic::amdgcn_ps_live: {
4923 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4926 case Intrinsic::amdgcn_div_scale: {
4929 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Dst0Size);
4930 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, Dst1Size);
4933 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4934 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4937 case Intrinsic::amdgcn_class: {
4938 Register Src0Reg =
MI.getOperand(2).getReg();
4939 Register Src1Reg =
MI.getOperand(3).getReg();
4943 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, DstSize);
4944 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Src0Size);
4945 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, Src1Size);
4948 case Intrinsic::amdgcn_icmp:
4949 case Intrinsic::amdgcn_fcmp: {
4952 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
4954 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
4955 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
4958 case Intrinsic::amdgcn_readlane: {
4962 unsigned IdxBank =
getRegBankID(IdxReg, MRI, AMDGPU::SGPRRegBankID);
4963 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4966 case Intrinsic::amdgcn_readfirstlane: {
4969 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
4970 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4973 case Intrinsic::amdgcn_writelane: {
4977 unsigned SrcBank =
getRegBankID(SrcReg, MRI, AMDGPU::SGPRRegBankID);
4980 unsigned IdxBank =
getRegBankID(IdxReg, MRI, AMDGPU::SGPRRegBankID);
4981 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
4985 OpdsMapping[2] = AMDGPU::getValueMapping(SrcBank, SrcSize);
4986 OpdsMapping[3] = AMDGPU::getValueMapping(IdxBank, IdxSize);
4987 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, SrcSize);
4990 case Intrinsic::amdgcn_if_break: {
4992 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4993 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
4994 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
4997 case Intrinsic::amdgcn_permlane16:
4998 case Intrinsic::amdgcn_permlanex16: {
5000 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5001 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5002 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5007 case Intrinsic::amdgcn_permlane_bcast:
5008 case Intrinsic::amdgcn_permlane_up:
5009 case Intrinsic::amdgcn_permlane_down:
5010 case Intrinsic::amdgcn_permlane_xor: {
5012 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5013 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5018 case Intrinsic::amdgcn_permlane_idx_gen: {
5020 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5021 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5025 case Intrinsic::amdgcn_permlane16_var:
5026 case Intrinsic::amdgcn_permlanex16_var: {
5028 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5029 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5030 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5031 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5034 case Intrinsic::amdgcn_mfma_f32_4x4x1f32:
5035 case Intrinsic::amdgcn_mfma_f32_4x4x4f16:
5036 case Intrinsic::amdgcn_mfma_i32_4x4x4i8:
5037 case Intrinsic::amdgcn_mfma_f32_4x4x2bf16:
5038 case Intrinsic::amdgcn_mfma_f32_16x16x1f32:
5039 case Intrinsic::amdgcn_mfma_f32_16x16x4f32:
5040 case Intrinsic::amdgcn_mfma_f32_16x16x4f16:
5041 case Intrinsic::amdgcn_mfma_f32_16x16x16f16:
5042 case Intrinsic::amdgcn_mfma_i32_16x16x4i8:
5043 case Intrinsic::amdgcn_mfma_i32_16x16x16i8:
5044 case Intrinsic::amdgcn_mfma_f32_16x16x2bf16:
5045 case Intrinsic::amdgcn_mfma_f32_16x16x8bf16:
5046 case Intrinsic::amdgcn_mfma_f32_32x32x1f32:
5047 case Intrinsic::amdgcn_mfma_f32_32x32x2f32:
5048 case Intrinsic::amdgcn_mfma_f32_32x32x4f16:
5049 case Intrinsic::amdgcn_mfma_f32_32x32x8f16:
5050 case Intrinsic::amdgcn_mfma_i32_32x32x4i8:
5051 case Intrinsic::amdgcn_mfma_i32_32x32x8i8:
5052 case Intrinsic::amdgcn_mfma_f32_32x32x2bf16:
5053 case Intrinsic::amdgcn_mfma_f32_32x32x4bf16:
5054 case Intrinsic::amdgcn_mfma_f32_32x32x4bf16_1k:
5055 case Intrinsic::amdgcn_mfma_f32_16x16x4bf16_1k:
5056 case Intrinsic::amdgcn_mfma_f32_4x4x4bf16_1k:
5057 case Intrinsic::amdgcn_mfma_f32_32x32x8bf16_1k:
5058 case Intrinsic::amdgcn_mfma_f32_16x16x16bf16_1k:
5059 case Intrinsic::amdgcn_mfma_f64_16x16x4f64:
5060 case Intrinsic::amdgcn_mfma_f64_4x4x4f64:
5061 case Intrinsic::amdgcn_mfma_i32_16x16x32_i8:
5062 case Intrinsic::amdgcn_mfma_i32_32x32x16_i8:
5063 case Intrinsic::amdgcn_mfma_f32_16x16x8_xf32:
5064 case Intrinsic::amdgcn_mfma_f32_32x32x4_xf32:
5065 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf8_bf8:
5066 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf8_fp8:
5067 case Intrinsic::amdgcn_mfma_f32_16x16x32_fp8_bf8:
5068 case Intrinsic::amdgcn_mfma_f32_16x16x32_fp8_fp8:
5069 case Intrinsic::amdgcn_mfma_f32_32x32x16_bf8_bf8:
5070 case Intrinsic::amdgcn_mfma_f32_32x32x16_bf8_fp8:
5071 case Intrinsic::amdgcn_mfma_f32_32x32x16_fp8_bf8:
5072 case Intrinsic::amdgcn_mfma_f32_32x32x16_fp8_fp8:
5073 case Intrinsic::amdgcn_mfma_f32_16x16x32_f16:
5074 case Intrinsic::amdgcn_mfma_f32_32x32x16_f16:
5075 case Intrinsic::amdgcn_mfma_i32_16x16x64_i8:
5076 case Intrinsic::amdgcn_mfma_i32_32x32x32_i8:
5077 case Intrinsic::amdgcn_mfma_f32_16x16x32_bf16: {
5079 unsigned MinNumRegsRequired = DstSize / 32;
5089 bool UseAGPRForm = !
Subtarget.hasGFX90AInsts() ||
5090 Info->selectAGPRFormMFMA(MinNumRegsRequired);
5102 case Intrinsic::amdgcn_mfma_scale_f32_16x16x128_f8f6f4:
5103 case Intrinsic::amdgcn_mfma_scale_f32_32x32x64_f8f6f4: {
5105 unsigned MinNumRegsRequired = DstSize / 32;
5124 case Intrinsic::amdgcn_smfmac_f32_16x16x32_f16:
5125 case Intrinsic::amdgcn_smfmac_f32_32x32x16_f16:
5126 case Intrinsic::amdgcn_smfmac_f32_16x16x32_bf16:
5127 case Intrinsic::amdgcn_smfmac_f32_32x32x16_bf16:
5128 case Intrinsic::amdgcn_smfmac_i32_16x16x64_i8:
5129 case Intrinsic::amdgcn_smfmac_i32_32x32x32_i8:
5130 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf8_bf8:
5131 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf8_fp8:
5132 case Intrinsic::amdgcn_smfmac_f32_16x16x64_fp8_bf8:
5133 case Intrinsic::amdgcn_smfmac_f32_16x16x64_fp8_fp8:
5134 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf8_bf8:
5135 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf8_fp8:
5136 case Intrinsic::amdgcn_smfmac_f32_32x32x32_fp8_bf8:
5137 case Intrinsic::amdgcn_smfmac_f32_32x32x32_fp8_fp8:
5138 case Intrinsic::amdgcn_smfmac_f32_16x16x64_f16:
5139 case Intrinsic::amdgcn_smfmac_f32_32x32x32_f16:
5140 case Intrinsic::amdgcn_smfmac_f32_16x16x64_bf16:
5141 case Intrinsic::amdgcn_smfmac_f32_32x32x32_bf16:
5142 case Intrinsic::amdgcn_smfmac_i32_16x16x128_i8:
5143 case Intrinsic::amdgcn_smfmac_i32_32x32x64_i8:
5144 case Intrinsic::amdgcn_smfmac_f32_16x16x128_bf8_bf8:
5145 case Intrinsic::amdgcn_smfmac_f32_16x16x128_bf8_fp8:
5146 case Intrinsic::amdgcn_smfmac_f32_16x16x128_fp8_bf8:
5147 case Intrinsic::amdgcn_smfmac_f32_16x16x128_fp8_fp8:
5148 case Intrinsic::amdgcn_smfmac_f32_32x32x64_bf8_bf8:
5149 case Intrinsic::amdgcn_smfmac_f32_32x32x64_bf8_fp8:
5150 case Intrinsic::amdgcn_smfmac_f32_32x32x64_fp8_bf8:
5151 case Intrinsic::amdgcn_smfmac_f32_32x32x64_fp8_fp8: {
5154 unsigned MinNumRegsRequired = DstSize / 32;
5170 case Intrinsic::amdgcn_interp_p1:
5171 case Intrinsic::amdgcn_interp_p2:
5172 case Intrinsic::amdgcn_interp_mov:
5173 case Intrinsic::amdgcn_interp_p1_f16:
5174 case Intrinsic::amdgcn_interp_p2_f16:
5175 case Intrinsic::amdgcn_lds_param_load: {
5176 const int M0Idx =
MI.getNumOperands() - 1;
5177 Register M0Reg =
MI.getOperand(M0Idx).getReg();
5178 unsigned M0Bank =
getRegBankID(M0Reg, MRI, AMDGPU::SGPRRegBankID);
5181 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5182 for (
int I = 2;
I != M0Idx &&
MI.getOperand(
I).
isReg(); ++
I)
5183 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5187 OpdsMapping[M0Idx] = AMDGPU::getValueMapping(M0Bank, 32);
5190 case Intrinsic::amdgcn_interp_inreg_p10:
5191 case Intrinsic::amdgcn_interp_inreg_p2:
5192 case Intrinsic::amdgcn_interp_inreg_p10_f16:
5193 case Intrinsic::amdgcn_interp_inreg_p2_f16:
5194 case Intrinsic::amdgcn_interp_p10_rtz_f16:
5195 case Intrinsic::amdgcn_interp_p2_rtz_f16: {
5197 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5198 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5199 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5200 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5203 case Intrinsic::amdgcn_permlane16_swap:
5204 case Intrinsic::amdgcn_permlane32_swap: {
5206 OpdsMapping[0] = OpdsMapping[1] = OpdsMapping[3] = OpdsMapping[4] =
5207 AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5210 case Intrinsic::amdgcn_ballot: {
5213 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
5214 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, SrcSize);
5217 case Intrinsic::amdgcn_inverse_ballot: {
5219 Register MaskReg =
MI.getOperand(2).getReg();
5221 unsigned MaskBank =
getRegBankID(MaskReg, MRI, AMDGPU::SGPRRegBankID);
5222 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5223 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, MaskSize);
5226 case Intrinsic::amdgcn_bitop3: {
5228 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5229 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5230 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5231 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5234 case Intrinsic::amdgcn_s_quadmask:
5235 case Intrinsic::amdgcn_s_wqm: {
5236 Register MaskReg =
MI.getOperand(2).getReg();
5238 unsigned MaskBank =
getRegBankID(MaskReg, MRI, AMDGPU::SGPRRegBankID);
5239 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, MaskSize);
5240 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, MaskSize);
5243 case Intrinsic::amdgcn_wave_reduce_add:
5244 case Intrinsic::amdgcn_wave_reduce_fadd:
5245 case Intrinsic::amdgcn_wave_reduce_sub:
5246 case Intrinsic::amdgcn_wave_reduce_fsub:
5247 case Intrinsic::amdgcn_wave_reduce_min:
5248 case Intrinsic::amdgcn_wave_reduce_umin:
5249 case Intrinsic::amdgcn_wave_reduce_fmin:
5250 case Intrinsic::amdgcn_wave_reduce_max:
5251 case Intrinsic::amdgcn_wave_reduce_umax:
5252 case Intrinsic::amdgcn_wave_reduce_fmax:
5253 case Intrinsic::amdgcn_wave_reduce_and:
5254 case Intrinsic::amdgcn_wave_reduce_or:
5255 case Intrinsic::amdgcn_wave_reduce_xor: {
5257 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, DstSize);
5261 OpdsMapping[2] = AMDGPU::getValueMapping(regBankID, OpSize);
5264 case Intrinsic::amdgcn_s_bitreplicate: {
5265 Register MaskReg =
MI.getOperand(2).getReg();
5266 unsigned MaskBank =
getRegBankID(MaskReg, MRI, AMDGPU::SGPRRegBankID);
5267 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 64);
5268 OpdsMapping[2] = AMDGPU::getValueMapping(MaskBank, 32);
5271 case Intrinsic::amdgcn_wave_shuffle: {
5273 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5274 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5275 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, OpSize);
5281 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD:
5282 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_D16:
5283 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_LOAD_NORET:
5284 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE:
5285 case AMDGPU::G_AMDGPU_INTRIN_IMAGE_STORE_D16: {
5288 assert(RSrcIntrin &&
"missing RsrcIntrinsic for image intrinsic");
5295 case AMDGPU::G_AMDGPU_BVH_INTERSECT_RAY:
5296 case AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY:
5297 case AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY: {
5299 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH_DUAL_INTERSECT_RAY ||
5300 MI.getOpcode() == AMDGPU::G_AMDGPU_BVH8_INTERSECT_RAY;
5301 unsigned NumMods = IsDualOrBVH8 ? 0 : 1;
5302 unsigned LastRegOpIdx =
MI.getNumExplicitOperands() - 1 - NumMods;
5304 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5306 OpdsMapping[1] = AMDGPU::getValueMapping(
5307 AMDGPU::VGPRRegBankID,
5309 OpdsMapping[2] = AMDGPU::getValueMapping(
5310 AMDGPU::VGPRRegBankID,
5313 OpdsMapping[LastRegOpIdx] =
5315 if (LastRegOpIdx == 3) {
5320 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5323 unsigned FirstSrcOpIdx = IsDualOrBVH8 ? 4 : 2;
5324 for (
unsigned I = FirstSrcOpIdx;
I < LastRegOpIdx; ++
I) {
5326 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5331 case AMDGPU::G_INTRINSIC_W_SIDE_EFFECTS:
5332 case AMDGPU::G_INTRINSIC_CONVERGENT_W_SIDE_EFFECTS: {
5335 case Intrinsic::amdgcn_s_getreg:
5336 case Intrinsic::amdgcn_s_memtime:
5337 case Intrinsic::amdgcn_s_memrealtime:
5338 case Intrinsic::amdgcn_s_get_waveid_in_workgroup:
5339 case Intrinsic::amdgcn_s_sendmsg_rtn: {
5341 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5344 case Intrinsic::amdgcn_global_atomic_fmin_num:
5345 case Intrinsic::amdgcn_global_atomic_fmax_num:
5346 case Intrinsic::amdgcn_flat_atomic_fmin_num:
5347 case Intrinsic::amdgcn_flat_atomic_fmax_num:
5348 case Intrinsic::amdgcn_global_atomic_ordered_add_b64:
5349 case Intrinsic::amdgcn_global_load_tr_b64:
5350 case Intrinsic::amdgcn_global_load_tr_b128:
5351 case Intrinsic::amdgcn_global_load_tr4_b64:
5352 case Intrinsic::amdgcn_global_load_tr6_b96:
5353 case Intrinsic::amdgcn_ds_load_tr8_b64:
5354 case Intrinsic::amdgcn_ds_load_tr16_b128:
5355 case Intrinsic::amdgcn_ds_load_tr4_b64:
5356 case Intrinsic::amdgcn_ds_load_tr6_b96:
5357 case Intrinsic::amdgcn_ds_read_tr4_b64:
5358 case Intrinsic::amdgcn_ds_read_tr6_b96:
5359 case Intrinsic::amdgcn_ds_read_tr8_b64:
5360 case Intrinsic::amdgcn_ds_read_tr16_b64:
5361 case Intrinsic::amdgcn_ds_atomic_async_barrier_arrive_b64:
5362 case Intrinsic::amdgcn_ds_atomic_barrier_arrive_rtn_b64:
5364 case Intrinsic::amdgcn_ds_ordered_add:
5365 case Intrinsic::amdgcn_ds_ordered_swap: {
5367 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5369 AMDGPU::SGPRRegBankID);
5370 OpdsMapping[2] = AMDGPU::getValueMapping(M0Bank, 32);
5371 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5374 case Intrinsic::amdgcn_ds_append:
5375 case Intrinsic::amdgcn_ds_consume: {
5377 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5381 case Intrinsic::amdgcn_exp_compr:
5382 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5383 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5385 case Intrinsic::amdgcn_exp:
5387 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5388 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5389 OpdsMapping[5] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5390 OpdsMapping[6] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5392 case Intrinsic::amdgcn_exp_row:
5393 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5394 OpdsMapping[4] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5395 OpdsMapping[5] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5396 OpdsMapping[6] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5399 case Intrinsic::amdgcn_s_alloc_vgpr:
5400 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 1);
5401 OpdsMapping[2] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 32);
5403 case Intrinsic::amdgcn_s_sendmsg:
5404 case Intrinsic::amdgcn_s_sendmsghalt: {
5407 AMDGPU::SGPRRegBankID);
5408 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5411 case Intrinsic::amdgcn_s_setreg: {
5414 AMDGPU::SGPRRegBankID);
5415 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5418 case Intrinsic::amdgcn_s_ttracedata: {
5421 getRegBankID(
MI.getOperand(1).getReg(), MRI, AMDGPU::SGPRRegBankID);
5422 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, 32);
5425 case Intrinsic::amdgcn_end_cf: {
5427 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5430 case Intrinsic::amdgcn_else: {
5432 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5433 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, WaveSize);
5434 OpdsMapping[3] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, WaveSize);
5437 case Intrinsic::amdgcn_init_whole_wave:
5438 case Intrinsic::amdgcn_live_mask: {
5439 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5442 case Intrinsic::amdgcn_wqm_demote:
5443 case Intrinsic::amdgcn_kill: {
5444 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5447 case Intrinsic::amdgcn_raw_buffer_load:
5448 case Intrinsic::amdgcn_raw_ptr_buffer_load:
5449 case Intrinsic::amdgcn_raw_atomic_buffer_load:
5450 case Intrinsic::amdgcn_raw_ptr_atomic_buffer_load:
5451 case Intrinsic::amdgcn_raw_tbuffer_load:
5452 case Intrinsic::amdgcn_raw_ptr_tbuffer_load: {
5461 case Intrinsic::amdgcn_raw_buffer_load_lds:
5462 case Intrinsic::amdgcn_raw_buffer_load_async_lds:
5463 case Intrinsic::amdgcn_raw_ptr_buffer_load_lds:
5464 case Intrinsic::amdgcn_raw_ptr_buffer_load_async_lds: {
5471 case Intrinsic::amdgcn_raw_buffer_store:
5472 case Intrinsic::amdgcn_raw_ptr_buffer_store:
5473 case Intrinsic::amdgcn_raw_buffer_store_format:
5474 case Intrinsic::amdgcn_raw_ptr_buffer_store_format:
5475 case Intrinsic::amdgcn_raw_tbuffer_store:
5476 case Intrinsic::amdgcn_raw_ptr_tbuffer_store: {
5483 case Intrinsic::amdgcn_struct_buffer_load:
5484 case Intrinsic::amdgcn_struct_ptr_buffer_load:
5485 case Intrinsic::amdgcn_struct_tbuffer_load:
5486 case Intrinsic::amdgcn_struct_ptr_tbuffer_load:
5487 case Intrinsic::amdgcn_struct_atomic_buffer_load:
5488 case Intrinsic::amdgcn_struct_ptr_atomic_buffer_load: {
5496 case Intrinsic::amdgcn_struct_buffer_load_lds:
5497 case Intrinsic::amdgcn_struct_buffer_load_async_lds:
5498 case Intrinsic::amdgcn_struct_ptr_buffer_load_lds:
5499 case Intrinsic::amdgcn_struct_ptr_buffer_load_async_lds: {
5507 case Intrinsic::amdgcn_struct_buffer_store:
5508 case Intrinsic::amdgcn_struct_ptr_buffer_store:
5509 case Intrinsic::amdgcn_struct_tbuffer_store:
5510 case Intrinsic::amdgcn_struct_ptr_tbuffer_store: {
5518 case Intrinsic::amdgcn_init_exec_from_input: {
5520 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID,
Size);
5523 case Intrinsic::amdgcn_ds_gws_init:
5524 case Intrinsic::amdgcn_ds_gws_barrier:
5525 case Intrinsic::amdgcn_ds_gws_sema_br: {
5526 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5530 AMDGPU::SGPRRegBankID);
5531 OpdsMapping[2] = AMDGPU::getValueMapping(Bank, 32);
5534 case Intrinsic::amdgcn_ds_gws_sema_v:
5535 case Intrinsic::amdgcn_ds_gws_sema_p:
5536 case Intrinsic::amdgcn_ds_gws_sema_release_all: {
5539 AMDGPU::SGPRRegBankID);
5540 OpdsMapping[1] = AMDGPU::getValueMapping(Bank, 32);
5543 case Intrinsic::amdgcn_cluster_load_b32:
5544 case Intrinsic::amdgcn_cluster_load_b64:
5545 case Intrinsic::amdgcn_cluster_load_b128: {
5549 getRegBankID(
MI.getOperand(4).getReg(), MRI, AMDGPU::SGPRRegBankID);
5550 OpdsMapping[4] = AMDGPU::getValueMapping(M0Bank, 32);
5553 case Intrinsic::amdgcn_cluster_load_async_to_lds_b8:
5554 case Intrinsic::amdgcn_cluster_load_async_to_lds_b32:
5555 case Intrinsic::amdgcn_cluster_load_async_to_lds_b64:
5556 case Intrinsic::amdgcn_cluster_load_async_to_lds_b128: {
5561 getRegBankID(
MI.getOperand(5).getReg(), MRI, AMDGPU::SGPRRegBankID);
5562 OpdsMapping[5] = AMDGPU::getValueMapping(M0Bank, 32);
5565 case Intrinsic::amdgcn_global_store_async_from_lds_b8:
5566 case Intrinsic::amdgcn_global_store_async_from_lds_b32:
5567 case Intrinsic::amdgcn_global_store_async_from_lds_b64:
5568 case Intrinsic::amdgcn_global_store_async_from_lds_b128:
5569 case Intrinsic::amdgcn_global_load_async_to_lds_b8:
5570 case Intrinsic::amdgcn_global_load_async_to_lds_b32:
5571 case Intrinsic::amdgcn_global_load_async_to_lds_b64:
5572 case Intrinsic::amdgcn_global_load_async_to_lds_b128: {
5578 case Intrinsic::amdgcn_load_to_lds:
5579 case Intrinsic::amdgcn_load_async_to_lds:
5580 case Intrinsic::amdgcn_global_load_lds:
5581 case Intrinsic::amdgcn_global_load_async_lds: {
5587 case Intrinsic::amdgcn_lds_direct_load: {
5588 const int M0Idx =
MI.getNumOperands() - 1;
5589 Register M0Reg =
MI.getOperand(M0Idx).getReg();
5590 unsigned M0Bank =
getRegBankID(M0Reg, MRI, AMDGPU::SGPRRegBankID);
5593 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, DstSize);
5594 for (
int I = 2;
I != M0Idx &&
MI.getOperand(
I).
isReg(); ++
I)
5595 OpdsMapping[
I] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, 32);
5599 OpdsMapping[M0Idx] = AMDGPU::getValueMapping(M0Bank, 32);
5602 case Intrinsic::amdgcn_ds_add_gs_reg_rtn:
5603 case Intrinsic::amdgcn_ds_sub_gs_reg_rtn:
5607 case Intrinsic::amdgcn_ds_bvh_stack_rtn:
5608 case Intrinsic::amdgcn_ds_bvh_stack_push4_pop1_rtn:
5609 case Intrinsic::amdgcn_ds_bvh_stack_push8_pop1_rtn:
5610 case Intrinsic::amdgcn_ds_bvh_stack_push8_pop2_rtn: {
5623 case Intrinsic::amdgcn_s_sleep_var:
5626 case Intrinsic::amdgcn_s_barrier_join:
5627 case Intrinsic::amdgcn_s_wakeup_barrier:
5630 case Intrinsic::amdgcn_s_barrier_init:
5631 case Intrinsic::amdgcn_s_barrier_signal_var:
5635 case Intrinsic::amdgcn_s_barrier_signal_isfirst: {
5636 const unsigned ResultSize = 1;
5638 AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, ResultSize);
5641 case Intrinsic::amdgcn_s_get_barrier_state:
5642 case Intrinsic::amdgcn_s_get_named_barrier_state: {
5647 case Intrinsic::amdgcn_pops_exiting_wave_id:
5649 case Intrinsic::amdgcn_tensor_load_to_lds:
5650 case Intrinsic::amdgcn_tensor_store_from_lds: {
5653 for (
unsigned I = 1;
I <
MI.getNumOperands(); ++
I) {
5654 if (
MI.getOperand(
I).isReg()) {
5658 OpdsMapping[
I] = AMDGPU::getValueMapping(OpBank,
Size);
5663 case Intrinsic::amdgcn_s_prefetch_data: {
5668 case Intrinsic::amdgcn_flat_prefetch:
5669 case Intrinsic::amdgcn_global_prefetch:
5676 case AMDGPU::G_SELECT: {
5679 AMDGPU::SGPRRegBankID);
5681 AMDGPU::SGPRRegBankID);
5682 bool SGPRSrcs = Op2Bank == AMDGPU::SGPRRegBankID &&
5683 Op3Bank == AMDGPU::SGPRRegBankID;
5685 unsigned CondBankDefault = SGPRSrcs ?
5686 AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
5689 if (CondBank == AMDGPU::SGPRRegBankID)
5690 CondBank = SGPRSrcs ? AMDGPU::SGPRRegBankID : AMDGPU::VCCRegBankID;
5691 else if (CondBank == AMDGPU::VGPRRegBankID)
5692 CondBank = AMDGPU::VCCRegBankID;
5694 unsigned Bank = SGPRSrcs && CondBank == AMDGPU::SGPRRegBankID ?
5695 AMDGPU::SGPRRegBankID : AMDGPU::VGPRRegBankID;
5697 assert(CondBank == AMDGPU::VCCRegBankID || CondBank == AMDGPU::SGPRRegBankID);
5701 OpdsMapping[0] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5702 OpdsMapping[1] = AMDGPU::getValueMapping(CondBank, 1);
5703 OpdsMapping[2] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5704 OpdsMapping[3] = AMDGPU::getValueMappingSGPR64Only(Bank,
Size);
5706 OpdsMapping[0] = AMDGPU::getValueMapping(Bank,
Size);
5707 OpdsMapping[1] = AMDGPU::getValueMapping(CondBank, 1);
5708 OpdsMapping[2] = AMDGPU::getValueMapping(Bank,
Size);
5709 OpdsMapping[3] = AMDGPU::getValueMapping(Bank,
Size);
5715 case AMDGPU::G_SI_CALL: {
5716 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::SGPRRegBankID, 64);
5722 for (
unsigned I = 4;
I <
MI.getNumOperands(); ++
I) {
5723 if (
MI.getOperand(
I).isReg()) {
5727 OpdsMapping[
I] = AMDGPU::getValueMapping(OpBank,
Size);
5732 case AMDGPU::G_LOAD:
5733 case AMDGPU::G_ZEXTLOAD:
5734 case AMDGPU::G_SEXTLOAD:
5737 case AMDGPU::G_ATOMICRMW_XCHG:
5738 case AMDGPU::G_ATOMICRMW_ADD:
5739 case AMDGPU::G_ATOMICRMW_SUB:
5740 case AMDGPU::G_ATOMICRMW_AND:
5741 case AMDGPU::G_ATOMICRMW_OR:
5742 case AMDGPU::G_ATOMICRMW_XOR:
5743 case AMDGPU::G_ATOMICRMW_MAX:
5744 case AMDGPU::G_ATOMICRMW_MIN:
5745 case AMDGPU::G_ATOMICRMW_UMAX:
5746 case AMDGPU::G_ATOMICRMW_UMIN:
5747 case AMDGPU::G_ATOMICRMW_FADD:
5748 case AMDGPU::G_ATOMICRMW_FMIN:
5749 case AMDGPU::G_ATOMICRMW_FMAX:
5750 case AMDGPU::G_ATOMICRMW_UINC_WRAP:
5751 case AMDGPU::G_ATOMICRMW_UDEC_WRAP:
5752 case AMDGPU::G_ATOMICRMW_USUB_COND:
5753 case AMDGPU::G_ATOMICRMW_USUB_SAT:
5754 case AMDGPU::G_AMDGPU_ATOMIC_CMPXCHG: {
5760 case AMDGPU::G_ATOMIC_CMPXCHG: {
5767 case AMDGPU::G_BRCOND: {
5769 AMDGPU::SGPRRegBankID);
5771 if (Bank != AMDGPU::SGPRRegBankID)
5772 Bank = AMDGPU::VCCRegBankID;
5774 OpdsMapping[0] = AMDGPU::getValueMapping(Bank, 1);
5777 case AMDGPU::G_INTRINSIC_FPTRUNC_ROUND:
5779 case AMDGPU::G_PREFETCH:
5782 case AMDGPU::G_AMDGPU_WHOLE_WAVE_FUNC_SETUP:
5783 case AMDGPU::G_AMDGPU_WHOLE_WAVE_FUNC_RETURN:
5784 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VCCRegBankID, 1);
5786 case AMDGPU::G_AMDGPU_FLAT_LOAD_MONITOR:
5787 case AMDGPU::G_AMDGPU_GLOBAL_LOAD_MONITOR: {
5790 OpdsMapping[0] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID,
Size);
5791 OpdsMapping[1] = AMDGPU::getValueMapping(AMDGPU::VGPRRegBankID, PtrSize);
5798 MI.getNumOperands());