65#define GET_INSTRINFO_CTOR_DTOR
66#include "AArch64GenInstrInfo.inc"
70 cl::desc(
"Restrict range of CB instructions (DEBUG)"));
74 cl::desc(
"Restrict range of TB[N]Z instructions (DEBUG)"));
78 cl::desc(
"Restrict range of CB[N]Z instructions (DEBUG)"));
82 cl::desc(
"Restrict range of Bcc instructions (DEBUG)"));
86 cl::desc(
"Restrict range of B instructions (DEBUG)"));
90 cl::desc(
"Restrict range of instructions to search for the "
91 "machine-combiner gather pattern optimization"));
96 RI(STI.getTargetTriple(), STI.getHwMode()), Subtarget(STI) {}
107 auto Op =
MI.getOpcode();
108 if (
Op == AArch64::INLINEASM ||
Op == AArch64::INLINEASM_BR)
109 return getInlineAsmLength(
MI.getOperand(0).getSymbolName(), *MAI);
113 if (
MI.isMetaInstruction())
118 unsigned NumBytes = 0;
122 NumBytes =
Desc.getSize() ?
Desc.getSize() : 4;
125 if (!MFI->shouldSignReturnAddress(*MF))
129 auto Method = STI.getAuthenticatedLRCheckMethod(*MF);
137 switch (
Desc.getOpcode()) {
140 return Desc.getSize();
147 case TargetOpcode::STACKMAP:
150 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
152 case TargetOpcode::PATCHPOINT:
155 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
157 case TargetOpcode::STATEPOINT:
159 assert(NumBytes % 4 == 0 &&
"Invalid number of NOP bytes requested!");
164 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
169 F.getFnAttributeAsParsedInteger(
"patchable-function-entry", 9) * 4;
171 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
172 case TargetOpcode::PATCHABLE_TAIL_CALL:
173 case TargetOpcode::PATCHABLE_TYPED_EVENT_CALL:
177 case TargetOpcode::PATCHABLE_EVENT_CALL:
183 NumBytes =
MI.getOperand(1).getImm();
185 case TargetOpcode::BUNDLE:
186 NumBytes = getInstBundleLength(
MI);
193unsigned AArch64InstrInfo::getInstBundleLength(
const MachineInstr &
MI)
const {
197 while (++
I != E &&
I->isInsideBundle()) {
198 assert(!
I->isBundle() &&
"No nested bundle!");
233 case AArch64::CBWPri:
234 case AArch64::CBXPri:
235 case AArch64::CBWPrr:
236 case AArch64::CBXPrr:
244 case AArch64::CBBAssertExt:
245 case AArch64::CBHAssertExt:
276 case AArch64::CBWPri:
277 case AArch64::CBXPri:
278 case AArch64::CBBAssertExt:
279 case AArch64::CBHAssertExt:
280 case AArch64::CBWPrr:
281 case AArch64::CBXPrr:
287 int64_t BrOffset)
const {
289 assert(Bits >= 3 &&
"max branch displacement must be enough to jump"
290 "over conditional branch expansion");
291 return isIntN(Bits, BrOffset / 4);
296 switch (
MI.getOpcode()) {
300 return MI.getOperand(0).getMBB();
305 return MI.getOperand(2).getMBB();
311 return MI.getOperand(1).getMBB();
312 case AArch64::CBWPri:
313 case AArch64::CBXPri:
314 case AArch64::CBBAssertExt:
315 case AArch64::CBHAssertExt:
316 case AArch64::CBWPrr:
317 case AArch64::CBXPrr:
318 return MI.getOperand(3).getMBB();
328 assert(RS &&
"RegScavenger required for long branching");
330 "new block should be inserted for expanding unconditional branch");
333 "restore block should be inserted for restoring clobbered registers");
340 "Branch offsets outside of the signed 33-bit range not supported");
351 RS->enterBasicBlockEnd(
MBB);
354 constexpr Register Reg = AArch64::X16;
355 if (!RS->isRegUsed(Reg)) {
356 insertUnconditionalBranch(
MBB, &NewDestBB,
DL);
363 Register Scavenged = RS->FindUnusedReg(&AArch64::GPR64RegClass);
364 if (Scavenged != AArch64::NoRegister &&
366 buildIndirectBranch(Scavenged, NewDestBB);
367 RS->setRegUsed(Scavenged);
376 "Unable to insert indirect branch inside function that has red zone");
399 bool AllowModify)
const {
406 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
407 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
411 if (!isUnpredicatedTerminator(*
I))
418 unsigned LastOpc = LastInst->
getOpcode();
419 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
434 unsigned SecondLastOpc = SecondLastInst->
getOpcode();
441 LastInst = SecondLastInst;
443 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
448 SecondLastInst = &*
I;
449 SecondLastOpc = SecondLastInst->
getOpcode();
460 LastInst = SecondLastInst;
462 if (
I ==
MBB.begin() || !isUnpredicatedTerminator(*--
I)) {
464 "unreachable unconditional branches removed above");
473 SecondLastInst = &*
I;
474 SecondLastOpc = SecondLastInst->
getOpcode();
478 if (SecondLastInst &&
I !=
MBB.begin() && isUnpredicatedTerminator(*--
I))
494 I->eraseFromParent();
503 I->eraseFromParent();
512 MachineBranchPredicate &MBP,
513 bool AllowModify)
const {
523 if (
I->getOpcode() == AArch64::SpeculationBarrierISBDSBEndBB ||
524 I->getOpcode() == AArch64::SpeculationBarrierSBEndBB) {
528 if (!isUnpredicatedTerminator(*
I))
533 unsigned LastOpc = LastInst->
getOpcode();
548 assert(MBP.TrueDest &&
"expected!");
549 MBP.FalseDest =
MBB.getNextNode();
551 MBP.ConditionDef =
nullptr;
552 MBP.SingleUseCondition =
false;
556 MBP.Predicate = (LastOpc == AArch64::CBNZX || LastOpc == AArch64::CBNZW)
557 ? MachineBranchPredicate::PRED_NE
558 : MachineBranchPredicate::PRED_EQ;
574 Cond[1].setImm(AArch64::CBNZW);
577 Cond[1].setImm(AArch64::CBZW);
580 Cond[1].setImm(AArch64::CBNZX);
583 Cond[1].setImm(AArch64::CBZX);
586 Cond[1].setImm(AArch64::TBNZW);
589 Cond[1].setImm(AArch64::TBZW);
592 Cond[1].setImm(AArch64::TBNZX);
595 Cond[1].setImm(AArch64::TBZX);
599 case AArch64::CBWPri:
600 case AArch64::CBXPri:
601 case AArch64::CBBAssertExt:
602 case AArch64::CBHAssertExt:
603 case AArch64::CBWPrr:
604 case AArch64::CBXPrr: {
617 int *BytesRemoved)
const {
627 I->eraseFromParent();
631 if (
I ==
MBB.begin()) {
644 I->eraseFromParent();
651void AArch64InstrInfo::instantiateCondBranch(
676 if (
Cond.size() > 5) {
687 assert(
TBB &&
"insertBranch must not be told to insert a fallthrough");
715 if (!
DefMI->isFullCopy())
717 VReg =
DefMI->getOperand(1).getReg();
726 unsigned *NewReg =
nullptr) {
731 bool Is64Bit = AArch64::GPR64allRegClass.hasSubClassEq(
MRI.getRegClass(VReg));
735 switch (
DefMI->getOpcode()) {
736 case AArch64::SUBREG_TO_REG:
740 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 0)
742 if (!
DefMI->getOperand(2).isReg())
744 if (!
DefMI->getOperand(3).isImm() ||
745 DefMI->getOperand(3).getImm() != AArch64::sub_32)
748 if (
DefMI->getOpcode() != AArch64::MOVi32imm)
750 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
753 SrcReg = AArch64::XZR;
754 Opc = AArch64::CSINCXr;
757 case AArch64::MOVi32imm:
758 case AArch64::MOVi64imm:
759 if (!
DefMI->getOperand(1).isImm() ||
DefMI->getOperand(1).getImm() != 1)
761 SrcReg = Is64Bit ? AArch64::XZR : AArch64::WZR;
762 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
765 case AArch64::ADDSXri:
766 case AArch64::ADDSWri:
768 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
773 case AArch64::ADDXri:
774 case AArch64::ADDWri:
776 if (!
DefMI->getOperand(2).isImm() ||
DefMI->getOperand(2).getImm() != 1 ||
777 DefMI->getOperand(3).getImm() != 0)
779 SrcReg =
DefMI->getOperand(1).getReg();
780 Opc = Is64Bit ? AArch64::CSINCXr : AArch64::CSINCWr;
783 case AArch64::ORNXrr:
784 case AArch64::ORNWrr: {
787 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
789 SrcReg =
DefMI->getOperand(2).getReg();
790 Opc = Is64Bit ? AArch64::CSINVXr : AArch64::CSINVWr;
794 case AArch64::SUBSXrr:
795 case AArch64::SUBSWrr:
797 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
802 case AArch64::SUBXrr:
803 case AArch64::SUBWrr: {
806 if (ZReg != AArch64::XZR && ZReg != AArch64::WZR)
808 SrcReg =
DefMI->getOperand(2).getReg();
809 Opc = Is64Bit ? AArch64::CSNEGXr : AArch64::CSNEGWr;
815 assert(
Opc && SrcReg &&
"Missing parameters");
827 int &FalseCycles)
const {
831 RI.getCommonSubClass(
MRI.getRegClass(TrueReg),
MRI.getRegClass(FalseReg));
838 if (!RI.getCommonSubClass(RC,
MRI.getRegClass(DstReg)))
842 unsigned ExtraCondLat =
Cond.size() != 1;
846 if (AArch64::GPR64allRegClass.hasSubClassEq(RC) ||
847 AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
849 CondCycles = 1 + ExtraCondLat;
850 TrueCycles = FalseCycles = 1;
860 if (AArch64::FPR64RegClass.hasSubClassEq(RC) ||
861 AArch64::FPR32RegClass.hasSubClassEq(RC)) {
862 CondCycles = 5 + ExtraCondLat;
863 TrueCycles = FalseCycles = 2;
880 switch (
Cond.size()) {
912 MRI.constrainRegClass(SrcReg, &AArch64::GPR64spRegClass);
918 MRI.constrainRegClass(SrcReg, &AArch64::GPR32spRegClass);
958 unsigned SubsOpc, SubsDestReg;
964 case AArch64::CBWPri:
965 SubsOpc = AArch64::SUBSWri;
966 SubsDestReg = AArch64::WZR;
969 case AArch64::CBXPri:
970 SubsOpc = AArch64::SUBSXri;
971 SubsDestReg = AArch64::XZR;
974 case AArch64::CBWPrr:
975 SubsOpc = AArch64::SUBSWrr;
976 SubsDestReg = AArch64::WZR;
979 case AArch64::CBXPrr:
980 SubsOpc = AArch64::SUBSXrr;
981 SubsDestReg = AArch64::XZR;
1010 switch (ExtendType) {
1016 "Unexpected compare-and-branch instruction for SXTB shift-extend");
1017 ExtOpc = AArch64::SBFMWri;
1023 "Unexpected compare-and-branch instruction for SXTH shift-extend");
1024 ExtOpc = AArch64::SBFMWri;
1030 "Unexpected compare-and-branch instruction for UXTB shift-extend");
1031 ExtOpc = AArch64::ANDWri;
1037 "Unexpected compare-and-branch instruction for UXTH shift-extend");
1038 ExtOpc = AArch64::ANDWri;
1044 Reg =
MRI.createVirtualRegister(&AArch64::GPR32spRegClass);
1047 if (ExtOpc != AArch64::ANDWri)
1049 MBBI.addImm(ExtBits);
1056 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1057 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1064 MRI.constrainRegClass(Reg,
MRI.getRegClass(
Cond[3].getReg()));
1065 MRI.constrainRegClass(
Cond[3].
getReg(), &AArch64::GPR32spRegClass);
1077 bool TryFold =
false;
1078 if (
MRI.constrainRegClass(DstReg, &AArch64::GPR64RegClass)) {
1079 RC = &AArch64::GPR64RegClass;
1080 Opc = AArch64::CSELXr;
1082 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::GPR32RegClass)) {
1083 RC = &AArch64::GPR32RegClass;
1084 Opc = AArch64::CSELWr;
1086 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR64RegClass)) {
1087 RC = &AArch64::FPR64RegClass;
1088 Opc = AArch64::FCSELDrrr;
1089 }
else if (
MRI.constrainRegClass(DstReg, &AArch64::FPR32RegClass)) {
1090 RC = &AArch64::FPR32RegClass;
1091 Opc = AArch64::FCSELSrrr;
1093 assert(RC &&
"Unsupported regclass");
1097 unsigned NewReg = 0;
1112 MRI.clearKillFlags(NewReg);
1117 MRI.constrainRegClass(TrueReg, RC);
1120 (FalseReg.
isVirtual() || FalseReg == AArch64::WZR ||
1121 FalseReg == AArch64::XZR) &&
1122 "FalseReg was folded into a non-virtual register other than WZR or XZR");
1124 MRI.constrainRegClass(FalseReg, RC);
1139 assert(BitSize == 64 &&
"Only bit sizes of 32 or 64 allowed");
1144 return Is.
size() <= 2;
1149 assert(
MI.isCopy() &&
"Expected COPY instruction");
1155 if (
Reg.isVirtual())
1156 return MRI.getRegClass(
Reg);
1157 if (
Reg.isPhysical())
1158 return RI.getMinimalPhysRegClass(
Reg);
1163 if (DstRC && SrcRC && !RI.getCommonSubClass(DstRC, SrcRC))
1166 return MI.isAsCheapAsAMove();
1172 if (Subtarget.hasExynosCheapAsMoveHandling()) {
1173 if (isExynosCheapAsMove(
MI))
1175 return MI.isAsCheapAsAMove();
1178 switch (
MI.getOpcode()) {
1180 return MI.isAsCheapAsAMove();
1182 case TargetOpcode::COPY:
1185 case AArch64::ADDWrs:
1186 case AArch64::ADDXrs:
1187 case AArch64::SUBWrs:
1188 case AArch64::SUBXrs:
1189 return Subtarget.hasALULSLFast() &&
MI.getOperand(3).getImm() <= 4;
1194 case AArch64::MOVi32imm:
1196 case AArch64::MOVi64imm:
1201bool AArch64InstrInfo::isFalkorShiftExtFast(
const MachineInstr &
MI) {
1202 switch (
MI.getOpcode()) {
1206 case AArch64::ADDWrs:
1207 case AArch64::ADDXrs:
1208 case AArch64::ADDSWrs:
1209 case AArch64::ADDSXrs: {
1210 unsigned Imm =
MI.getOperand(3).getImm();
1217 case AArch64::ADDWrx:
1218 case AArch64::ADDXrx:
1219 case AArch64::ADDXrx64:
1220 case AArch64::ADDSWrx:
1221 case AArch64::ADDSXrx:
1222 case AArch64::ADDSXrx64: {
1223 unsigned Imm =
MI.getOperand(3).getImm();
1235 case AArch64::SUBWrs:
1236 case AArch64::SUBSWrs: {
1237 unsigned Imm =
MI.getOperand(3).getImm();
1239 return ShiftVal == 0 ||
1243 case AArch64::SUBXrs:
1244 case AArch64::SUBSXrs: {
1245 unsigned Imm =
MI.getOperand(3).getImm();
1247 return ShiftVal == 0 ||
1251 case AArch64::SUBWrx:
1252 case AArch64::SUBXrx:
1253 case AArch64::SUBXrx64:
1254 case AArch64::SUBSWrx:
1255 case AArch64::SUBSXrx:
1256 case AArch64::SUBSXrx64: {
1257 unsigned Imm =
MI.getOperand(3).getImm();
1269 case AArch64::LDRBBroW:
1270 case AArch64::LDRBBroX:
1271 case AArch64::LDRBroW:
1272 case AArch64::LDRBroX:
1273 case AArch64::LDRDroW:
1274 case AArch64::LDRDroX:
1275 case AArch64::LDRHHroW:
1276 case AArch64::LDRHHroX:
1277 case AArch64::LDRHroW:
1278 case AArch64::LDRHroX:
1279 case AArch64::LDRQroW:
1280 case AArch64::LDRQroX:
1281 case AArch64::LDRSBWroW:
1282 case AArch64::LDRSBWroX:
1283 case AArch64::LDRSBXroW:
1284 case AArch64::LDRSBXroX:
1285 case AArch64::LDRSHWroW:
1286 case AArch64::LDRSHWroX:
1287 case AArch64::LDRSHXroW:
1288 case AArch64::LDRSHXroX:
1289 case AArch64::LDRSWroW:
1290 case AArch64::LDRSWroX:
1291 case AArch64::LDRSroW:
1292 case AArch64::LDRSroX:
1293 case AArch64::LDRWroW:
1294 case AArch64::LDRWroX:
1295 case AArch64::LDRXroW:
1296 case AArch64::LDRXroX:
1297 case AArch64::PRFMroW:
1298 case AArch64::PRFMroX:
1299 case AArch64::STRBBroW:
1300 case AArch64::STRBBroX:
1301 case AArch64::STRBroW:
1302 case AArch64::STRBroX:
1303 case AArch64::STRDroW:
1304 case AArch64::STRDroX:
1305 case AArch64::STRHHroW:
1306 case AArch64::STRHHroX:
1307 case AArch64::STRHroW:
1308 case AArch64::STRHroX:
1309 case AArch64::STRQroW:
1310 case AArch64::STRQroX:
1311 case AArch64::STRSroW:
1312 case AArch64::STRSroX:
1313 case AArch64::STRWroW:
1314 case AArch64::STRWroX:
1315 case AArch64::STRXroW:
1316 case AArch64::STRXroX: {
1317 unsigned IsSigned =
MI.getOperand(3).getImm();
1324 unsigned Opc =
MI.getOpcode();
1328 case AArch64::SEH_StackAlloc:
1329 case AArch64::SEH_SaveFPLR:
1330 case AArch64::SEH_SaveFPLR_X:
1331 case AArch64::SEH_SaveReg:
1332 case AArch64::SEH_SaveReg_X:
1333 case AArch64::SEH_SaveRegP:
1334 case AArch64::SEH_SaveRegP_X:
1335 case AArch64::SEH_SaveFReg:
1336 case AArch64::SEH_SaveFReg_X:
1337 case AArch64::SEH_SaveFRegP:
1338 case AArch64::SEH_SaveFRegP_X:
1339 case AArch64::SEH_SetFP:
1340 case AArch64::SEH_AddFP:
1341 case AArch64::SEH_Nop:
1342 case AArch64::SEH_PrologEnd:
1343 case AArch64::SEH_EpilogStart:
1344 case AArch64::SEH_EpilogEnd:
1345 case AArch64::SEH_PACSignLR:
1346 case AArch64::SEH_SaveAnyRegI:
1347 case AArch64::SEH_SaveAnyRegIP:
1348 case AArch64::SEH_SaveAnyRegQP:
1349 case AArch64::SEH_SaveAnyRegQPX:
1350 case AArch64::SEH_AllocZ:
1351 case AArch64::SEH_SaveZReg:
1352 case AArch64::SEH_SavePReg:
1359 unsigned &SubIdx)
const {
1360 switch (
MI.getOpcode()) {
1363 case AArch64::SBFMXri:
1364 case AArch64::UBFMXri:
1367 if (
MI.getOperand(2).getImm() != 0 ||
MI.getOperand(3).getImm() != 31)
1370 SrcReg =
MI.getOperand(1).getReg();
1371 DstReg =
MI.getOperand(0).getReg();
1372 SubIdx = AArch64::sub_32;
1381 int64_t OffsetA = 0, OffsetB = 0;
1382 TypeSize WidthA(0,
false), WidthB(0,
false);
1383 bool OffsetAIsScalable =
false, OffsetBIsScalable =
false;
1404 OffsetAIsScalable == OffsetBIsScalable) {
1405 int LowOffset = OffsetA < OffsetB ? OffsetA : OffsetB;
1406 int HighOffset = OffsetA < OffsetB ? OffsetB : OffsetA;
1407 TypeSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
1408 if (LowWidth.
isScalable() == OffsetAIsScalable &&
1426 switch (
MI.getOpcode()) {
1429 if (
MI.getOperand(0).getImm() == 0x14)
1436 case AArch64::MSRpstatesvcrImm1:
1443 auto Next = std::next(
MI.getIterator());
1444 return Next !=
MBB->end() &&
Next->isCFIInstruction();
1451 Register &SrcReg2, int64_t &CmpMask,
1452 int64_t &CmpValue)
const {
1456 assert(
MI.getNumOperands() >= 2 &&
"All AArch64 cmps should have 2 operands");
1457 if (!
MI.getOperand(1).isReg() ||
MI.getOperand(1).getSubReg())
1460 switch (
MI.getOpcode()) {
1463 case AArch64::PTEST_PP:
1464 case AArch64::PTEST_PP_ANY:
1465 case AArch64::PTEST_PP_FIRST:
1466 SrcReg =
MI.getOperand(0).getReg();
1467 SrcReg2 =
MI.getOperand(1).getReg();
1468 if (
MI.getOperand(2).getSubReg())
1475 case AArch64::SUBSWrr:
1476 case AArch64::SUBSWrs:
1477 case AArch64::SUBSWrx:
1478 case AArch64::SUBSXrr:
1479 case AArch64::SUBSXrs:
1480 case AArch64::SUBSXrx:
1481 case AArch64::ADDSWrr:
1482 case AArch64::ADDSWrs:
1483 case AArch64::ADDSWrx:
1484 case AArch64::ADDSXrr:
1485 case AArch64::ADDSXrs:
1486 case AArch64::ADDSXrx:
1488 SrcReg =
MI.getOperand(1).getReg();
1489 SrcReg2 =
MI.getOperand(2).getReg();
1492 if (
MI.getOperand(2).getSubReg())
1498 case AArch64::SUBSWri:
1499 case AArch64::ADDSWri:
1500 case AArch64::SUBSXri:
1501 case AArch64::ADDSXri:
1502 SrcReg =
MI.getOperand(1).getReg();
1505 CmpValue =
MI.getOperand(2).getImm();
1507 case AArch64::ANDSWri:
1508 case AArch64::ANDSXri:
1511 SrcReg =
MI.getOperand(1).getReg();
1515 MI.getOperand(2).getImm(),
1516 MI.getOpcode() == AArch64::ANDSWri ? 32 : 64);
1525 assert(
MBB &&
"Can't get MachineBasicBlock here");
1527 assert(MF &&
"Can't get MachineFunction here");
1532 for (
unsigned OpIdx = 0, EndIdx = Instr.getNumOperands();
OpIdx < EndIdx;
1539 if (!OpRegCstraints)
1547 "Operand has register constraints without being a register!");
1550 if (
Reg.isPhysical()) {
1554 !
MRI->constrainRegClass(
Reg, OpRegCstraints))
1567 bool MIDefinesZeroReg =
false;
1568 if (
MI.definesRegister(AArch64::WZR,
nullptr) ||
1569 MI.definesRegister(AArch64::XZR,
nullptr))
1570 MIDefinesZeroReg =
true;
1572 switch (
MI.getOpcode()) {
1574 return MI.getOpcode();
1575 case AArch64::ADDSWrr:
1576 return AArch64::ADDWrr;
1577 case AArch64::ADDSWri:
1578 return MIDefinesZeroReg ? AArch64::ADDSWri : AArch64::ADDWri;
1579 case AArch64::ADDSWrs:
1580 return MIDefinesZeroReg ? AArch64::ADDSWrs : AArch64::ADDWrs;
1581 case AArch64::ADDSWrx:
1582 return AArch64::ADDWrx;
1583 case AArch64::ADDSXrr:
1584 return AArch64::ADDXrr;
1585 case AArch64::ADDSXri:
1586 return MIDefinesZeroReg ? AArch64::ADDSXri : AArch64::ADDXri;
1587 case AArch64::ADDSXrs:
1588 return MIDefinesZeroReg ? AArch64::ADDSXrs : AArch64::ADDXrs;
1589 case AArch64::ADDSXrx:
1590 return AArch64::ADDXrx;
1591 case AArch64::SUBSWrr:
1592 return AArch64::SUBWrr;
1593 case AArch64::SUBSWri:
1594 return MIDefinesZeroReg ? AArch64::SUBSWri : AArch64::SUBWri;
1595 case AArch64::SUBSWrs:
1596 return MIDefinesZeroReg ? AArch64::SUBSWrs : AArch64::SUBWrs;
1597 case AArch64::SUBSWrx:
1598 return AArch64::SUBWrx;
1599 case AArch64::SUBSXrr:
1600 return AArch64::SUBXrr;
1601 case AArch64::SUBSXri:
1602 return MIDefinesZeroReg ? AArch64::SUBSXri : AArch64::SUBXri;
1603 case AArch64::SUBSXrs:
1604 return MIDefinesZeroReg ? AArch64::SUBSXrs : AArch64::SUBXrs;
1605 case AArch64::SUBSXrx:
1606 return AArch64::SUBXrx;
1621 if (To == To->getParent()->begin())
1626 if (To->getParent() != From->getParent())
1638 Instr.modifiesRegister(AArch64::NZCV,
TRI)) ||
1639 ((AccessToCheck &
AK_Read) && Instr.readsRegister(AArch64::NZCV,
TRI)))
1645std::optional<unsigned>
1649 unsigned MaskOpcode =
Mask->getOpcode();
1650 unsigned PredOpcode = Pred->
getOpcode();
1651 bool PredIsPTestLike = isPTestLikeOpcode(PredOpcode);
1652 bool PredIsWhileLike = isWhileOpcode(PredOpcode);
1654 if (PredIsWhileLike) {
1658 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1665 getElementSizeForOpcode(MaskOpcode) ==
1666 getElementSizeForOpcode(PredOpcode))
1672 if (PTest->
getOpcode() == AArch64::PTEST_PP_FIRST &&
1679 if (PredIsPTestLike) {
1684 if ((Mask == Pred) && PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1692 if (Mask != PTestLikeMask && PTestLikeMask->isFullCopy() &&
1693 PTestLikeMask->getOperand(1).getReg().isVirtual())
1695 MRI->getUniqueVRegDef(PTestLikeMask->getOperand(1).getReg());
1701 getElementSizeForOpcode(MaskOpcode) ==
1702 getElementSizeForOpcode(PredOpcode)) {
1703 if (Mask == PTestLikeMask || PTest->
getOpcode() == AArch64::PTEST_PP_ANY)
1729 uint64_t PredElementSize = getElementSizeForOpcode(PredOpcode);
1731 PTest->
getOpcode() == AArch64::PTEST_PP_ANY))
1739 switch (PredOpcode) {
1740 case AArch64::AND_PPzPP:
1741 case AArch64::BIC_PPzPP:
1742 case AArch64::EOR_PPzPP:
1743 case AArch64::NAND_PPzPP:
1744 case AArch64::NOR_PPzPP:
1745 case AArch64::ORN_PPzPP:
1746 case AArch64::ORR_PPzPP:
1747 case AArch64::BRKA_PPzP:
1748 case AArch64::BRKPA_PPzPP:
1749 case AArch64::BRKB_PPzP:
1750 case AArch64::BRKPB_PPzPP:
1751 case AArch64::RDFFR_PPz: {
1755 if (Mask != PredMask)
1759 case AArch64::BRKN_PPzP: {
1763 if ((MaskOpcode != AArch64::PTRUE_B) ||
1764 (
Mask->getOperand(1).getImm() != 31))
1768 case AArch64::PTRUE_B:
1781bool AArch64InstrInfo::optimizePTestInstr(
1782 MachineInstr *PTest,
unsigned MaskReg,
unsigned PredReg,
1784 auto *
Mask =
MRI->getUniqueVRegDef(MaskReg);
1785 auto *Pred =
MRI->getUniqueVRegDef(PredReg);
1787 if (Pred->
isCopy() && PTest->
getOpcode() == AArch64::PTEST_PP_FIRST) {
1791 if (
Op.isReg() &&
Op.getReg().isVirtual() &&
1792 Op.getSubReg() == AArch64::psub0)
1793 Pred =
MRI->getUniqueVRegDef(
Op.getReg());
1796 unsigned PredOpcode = Pred->
getOpcode();
1797 auto NewOp = canRemovePTestInstr(PTest, Mask, Pred,
MRI);
1813 if (*NewOp != PredOpcode) {
1824 for (; i !=
e; ++i) {
1855 if (DeadNZCVIdx != -1) {
1874 if (CmpInstr.
getOpcode() == AArch64::PTEST_PP ||
1875 CmpInstr.
getOpcode() == AArch64::PTEST_PP_ANY ||
1876 CmpInstr.
getOpcode() == AArch64::PTEST_PP_FIRST)
1877 return optimizePTestInstr(&CmpInstr, SrcReg, SrcReg2,
MRI);
1886 if (CmpValue == 0 && substituteCmpToZero(CmpInstr, SrcReg, *
MRI))
1888 return (CmpValue == 0 || CmpValue == 1) &&
1889 removeCmpToZeroOrOne(CmpInstr, SrcReg, CmpValue, *
MRI);
1897 switch (Instr.getOpcode()) {
1899 return AArch64::INSTRUCTION_LIST_END;
1901 case AArch64::ADDSWrr:
1902 case AArch64::ADDSWri:
1903 case AArch64::ADDSXrr:
1904 case AArch64::ADDSXri:
1905 case AArch64::ADDSWrx:
1906 case AArch64::ADDSXrx:
1907 case AArch64::SUBSWrr:
1908 case AArch64::SUBSWri:
1909 case AArch64::SUBSWrx:
1910 case AArch64::SUBSXrr:
1911 case AArch64::SUBSXri:
1912 case AArch64::SUBSXrx:
1913 case AArch64::ANDSWri:
1914 case AArch64::ANDSWrr:
1915 case AArch64::ANDSWrs:
1916 case AArch64::ANDSXri:
1917 case AArch64::ANDSXrr:
1918 case AArch64::ANDSXrs:
1919 case AArch64::BICSWrr:
1920 case AArch64::BICSXrr:
1921 case AArch64::BICSWrs:
1922 case AArch64::BICSXrs:
1923 return Instr.getOpcode();
1925 case AArch64::ADDWrr:
1926 return AArch64::ADDSWrr;
1927 case AArch64::ADDWri:
1928 return AArch64::ADDSWri;
1929 case AArch64::ADDXrr:
1930 return AArch64::ADDSXrr;
1931 case AArch64::ADDXri:
1932 return AArch64::ADDSXri;
1933 case AArch64::ADDWrx:
1934 return AArch64::ADDSWrx;
1935 case AArch64::ADDXrx:
1936 return AArch64::ADDSXrx;
1937 case AArch64::ADCWr:
1938 return AArch64::ADCSWr;
1939 case AArch64::ADCXr:
1940 return AArch64::ADCSXr;
1941 case AArch64::SUBWrr:
1942 return AArch64::SUBSWrr;
1943 case AArch64::SUBWri:
1944 return AArch64::SUBSWri;
1945 case AArch64::SUBXrr:
1946 return AArch64::SUBSXrr;
1947 case AArch64::SUBXri:
1948 return AArch64::SUBSXri;
1949 case AArch64::SUBWrx:
1950 return AArch64::SUBSWrx;
1951 case AArch64::SUBXrx:
1952 return AArch64::SUBSXrx;
1953 case AArch64::SBCWr:
1954 return AArch64::SBCSWr;
1955 case AArch64::SBCXr:
1956 return AArch64::SBCSXr;
1957 case AArch64::ANDWri:
1958 return AArch64::ANDSWri;
1959 case AArch64::ANDXri:
1960 return AArch64::ANDSXri;
1961 case AArch64::ANDWrr:
1962 return AArch64::ANDSWrr;
1963 case AArch64::ANDWrs:
1964 return AArch64::ANDSWrs;
1965 case AArch64::ANDXrr:
1966 return AArch64::ANDSXrr;
1967 case AArch64::ANDXrs:
1968 return AArch64::ANDSXrs;
1969 case AArch64::BICWrr:
1970 return AArch64::BICSWrr;
1971 case AArch64::BICXrr:
1972 return AArch64::BICSXrr;
1973 case AArch64::BICWrs:
1974 return AArch64::BICSWrs;
1975 case AArch64::BICXrs:
1976 return AArch64::BICSXrs;
1982 for (
auto *BB :
MBB->successors())
1983 if (BB->isLiveIn(AArch64::NZCV))
1992 switch (Instr.getOpcode()) {
1996 case AArch64::Bcc: {
1997 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2002 case AArch64::CSINVWr:
2003 case AArch64::CSINVXr:
2004 case AArch64::CSINCWr:
2005 case AArch64::CSINCXr:
2006 case AArch64::CSELWr:
2007 case AArch64::CSELXr:
2008 case AArch64::CSNEGWr:
2009 case AArch64::CSNEGXr:
2010 case AArch64::FCSELSrrr:
2011 case AArch64::FCSELDrrr: {
2012 int Idx = Instr.findRegisterUseOperandIdx(AArch64::NZCV,
nullptr);
2025 Instr.getOperand(CCIdx).
getImm())
2078std::optional<UsedNZCV>
2083 if (
MI.getParent() != CmpParent)
2084 return std::nullopt;
2087 return std::nullopt;
2092 if (Instr.readsRegister(AArch64::NZCV, &
TRI)) {
2095 return std::nullopt;
2100 if (Instr.modifiesRegister(AArch64::NZCV, &
TRI))
2103 return NZCVUsedAfterCmp;
2107 return Opcode == AArch64::ADDSWri || Opcode == AArch64::ADDSXri;
2111 return Opcode == AArch64::SUBSWri || Opcode == AArch64::SUBSXri;
2117 case AArch64::ANDSWri:
2118 case AArch64::ANDSWrr:
2119 case AArch64::ANDSWrs:
2120 case AArch64::ANDSXri:
2121 case AArch64::ANDSXrr:
2122 case AArch64::ANDSXrs:
2123 case AArch64::BICSWrr:
2124 case AArch64::BICSXrr:
2125 case AArch64::BICSWrs:
2126 case AArch64::BICSXrs:
2152 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2158 "Caller guarantees that CmpInstr compares with constant 0");
2161 if (!NZVCUsed || NZVCUsed->C)
2184bool AArch64InstrInfo::substituteCmpToZero(
2188 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2195 if (NewOpc == AArch64::INSTRUCTION_LIST_END)
2202 MI->setDesc(
get(NewOpc));
2207 MI->addRegisterDefined(AArch64::NZCV, &
TRI);
2219 assert((CmpValue == 0 || CmpValue == 1) &&
2220 "Only comparisons to 0 or 1 considered for removal!");
2223 unsigned MIOpc =
MI.getOpcode();
2224 if (MIOpc == AArch64::CSINCWr) {
2225 if (
MI.getOperand(1).getReg() != AArch64::WZR ||
2226 MI.getOperand(2).getReg() != AArch64::WZR)
2228 }
else if (MIOpc == AArch64::CSINCXr) {
2229 if (
MI.getOperand(1).getReg() != AArch64::XZR ||
2230 MI.getOperand(2).getReg() != AArch64::XZR)
2240 if (
MI.findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) != -1)
2244 const unsigned CmpOpcode = CmpInstr.
getOpcode();
2246 if (CmpValue && !IsSubsRegImm)
2248 if (!CmpValue && !IsSubsRegImm && !
isADDSRegImm(CmpOpcode))
2253 if (MIUsedNZCV.
C || MIUsedNZCV.
V)
2256 std::optional<UsedNZCV> NZCVUsedAfterCmp =
2260 if (!NZCVUsedAfterCmp || NZCVUsedAfterCmp->C || NZCVUsedAfterCmp->V)
2263 if ((MIUsedNZCV.
Z && NZCVUsedAfterCmp->N) ||
2264 (MIUsedNZCV.
N && NZCVUsedAfterCmp->Z))
2267 if (MIUsedNZCV.
N && !CmpValue)
2309bool AArch64InstrInfo::removeCmpToZeroOrOne(
2312 MachineInstr *
MI =
MRI.getUniqueVRegDef(SrcReg);
2316 SmallVector<MachineInstr *, 4> CCUseInstrs;
2317 bool IsInvertCC =
false;
2325 for (MachineInstr *CCUseInstr : CCUseInstrs) {
2327 assert(Idx >= 0 &&
"Unexpected instruction using CC.");
2328 MachineOperand &CCOperand = CCUseInstr->getOperand(Idx);
2337bool AArch64InstrInfo::expandPostRAPseudo(
MachineInstr &
MI)
const {
2338 if (
MI.getOpcode() != TargetOpcode::LOAD_STACK_GUARD &&
2339 MI.getOpcode() != AArch64::CATCHRET)
2342 MachineBasicBlock &
MBB = *
MI.getParent();
2344 auto TRI = Subtarget.getRegisterInfo();
2347 if (
MI.getOpcode() == AArch64::CATCHRET) {
2349 const TargetInstrInfo *
TII =
2351 MachineBasicBlock *TargetMBB =
MI.getOperand(0).getMBB();
2356 FirstEpilogSEH = std::prev(FirstEpilogSEH);
2358 FirstEpilogSEH = std::next(FirstEpilogSEH);
2373 if (
M.getStackProtectorGuard() ==
"sysreg") {
2374 const AArch64SysReg::SysReg *SrcReg =
2375 AArch64SysReg::lookupSysRegByName(
M.getStackProtectorGuardReg());
2383 int Offset =
M.getStackProtectorGuardOffset();
2434 const GlobalValue *GV =
2437 unsigned OpFlags = Subtarget.ClassifyGlobalReference(GV, TM);
2443 if (Subtarget.isTargetILP32()) {
2444 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2458 assert(!Subtarget.isTargetILP32() &&
"how can large exist in ILP32?");
2485 if (Subtarget.isTargetILP32()) {
2486 unsigned Reg32 =
TRI->getSubReg(
Reg, AArch64::sub_32);
2509 switch (
MI.getOpcode()) {
2512 case AArch64::MOVZWi:
2513 case AArch64::MOVZXi:
2514 if (
MI.getOperand(1).isImm() &&
MI.getOperand(1).getImm() == 0) {
2515 assert(
MI.getDesc().getNumOperands() == 3 &&
2516 MI.getOperand(2).getImm() == 0 &&
"invalid MOVZi operands");
2520 case AArch64::ANDWri:
2521 return MI.getOperand(1).getReg() == AArch64::WZR;
2522 case AArch64::ANDXri:
2523 return MI.getOperand(1).getReg() == AArch64::XZR;
2524 case TargetOpcode::COPY:
2525 return MI.getOperand(1).getReg() == AArch64::WZR;
2533 switch (
MI.getOpcode()) {
2536 case TargetOpcode::COPY: {
2539 return (AArch64::GPR32RegClass.
contains(DstReg) ||
2540 AArch64::GPR64RegClass.
contains(DstReg));
2542 case AArch64::ORRXrs:
2543 if (
MI.getOperand(1).getReg() == AArch64::XZR) {
2544 assert(
MI.getDesc().getNumOperands() == 4 &&
2545 MI.getOperand(3).getImm() == 0 &&
"invalid ORRrs operands");
2549 case AArch64::ADDXri:
2550 if (
MI.getOperand(2).getImm() == 0) {
2551 assert(
MI.getDesc().getNumOperands() == 4 &&
2552 MI.getOperand(3).getImm() == 0 &&
"invalid ADDXri operands");
2563 switch (
MI.getOpcode()) {
2566 case TargetOpcode::COPY: {
2568 return AArch64::FPR128RegClass.contains(DstReg);
2570 case AArch64::ORRv16i8:
2571 if (
MI.getOperand(1).getReg() ==
MI.getOperand(2).getReg()) {
2572 assert(
MI.getDesc().getNumOperands() == 3 &&
MI.getOperand(0).isReg() &&
2573 "invalid ORRv16i8 operands");
2585 case AArch64::LDRWui:
2586 case AArch64::LDRXui:
2587 case AArch64::LDRBui:
2588 case AArch64::LDRHui:
2589 case AArch64::LDRSui:
2590 case AArch64::LDRDui:
2591 case AArch64::LDRQui:
2592 case AArch64::LDR_PXI:
2598 int &FrameIndex)
const {
2602 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2603 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2604 FrameIndex =
MI.getOperand(1).getIndex();
2605 return MI.getOperand(0).getReg();
2614 case AArch64::STRWui:
2615 case AArch64::STRXui:
2616 case AArch64::STRBui:
2617 case AArch64::STRHui:
2618 case AArch64::STRSui:
2619 case AArch64::STRDui:
2620 case AArch64::STRQui:
2621 case AArch64::STR_PXI:
2627 int &FrameIndex)
const {
2631 if (
MI.getOperand(0).getSubReg() == 0 &&
MI.getOperand(1).isFI() &&
2632 MI.getOperand(2).isImm() &&
MI.getOperand(2).getImm() == 0) {
2633 FrameIndex =
MI.getOperand(1).getIndex();
2634 return MI.getOperand(0).getReg();
2640 int &FrameIndex)
const {
2655 return MI.getOperand(0).getReg();
2661 int &FrameIndex)
const {
2676 return MI.getOperand(0).getReg();
2684 return MMO->getFlags() & MOSuppressPair;
2690 if (
MI.memoperands_empty())
2698 return MMO->getFlags() & MOStridedAccess;
2706 case AArch64::STURSi:
2707 case AArch64::STRSpre:
2708 case AArch64::STURDi:
2709 case AArch64::STRDpre:
2710 case AArch64::STURQi:
2711 case AArch64::STRQpre:
2712 case AArch64::STURBBi:
2713 case AArch64::STURHHi:
2714 case AArch64::STURWi:
2715 case AArch64::STRWpre:
2716 case AArch64::STURXi:
2717 case AArch64::STRXpre:
2718 case AArch64::LDURSi:
2719 case AArch64::LDRSpre:
2720 case AArch64::LDURDi:
2721 case AArch64::LDRDpre:
2722 case AArch64::LDURQi:
2723 case AArch64::LDRQpre:
2724 case AArch64::LDURWi:
2725 case AArch64::LDRWpre:
2726 case AArch64::LDURXi:
2727 case AArch64::LDRXpre:
2728 case AArch64::LDRSWpre:
2729 case AArch64::LDURSWi:
2730 case AArch64::LDURHHi:
2731 case AArch64::LDURBBi:
2732 case AArch64::LDURSBWi:
2733 case AArch64::LDURSHWi:
2741 case AArch64::PRFMui:
return AArch64::PRFUMi;
2742 case AArch64::LDRXui:
return AArch64::LDURXi;
2743 case AArch64::LDRWui:
return AArch64::LDURWi;
2744 case AArch64::LDRBui:
return AArch64::LDURBi;
2745 case AArch64::LDRHui:
return AArch64::LDURHi;
2746 case AArch64::LDRSui:
return AArch64::LDURSi;
2747 case AArch64::LDRDui:
return AArch64::LDURDi;
2748 case AArch64::LDRQui:
return AArch64::LDURQi;
2749 case AArch64::LDRBBui:
return AArch64::LDURBBi;
2750 case AArch64::LDRHHui:
return AArch64::LDURHHi;
2751 case AArch64::LDRSBXui:
return AArch64::LDURSBXi;
2752 case AArch64::LDRSBWui:
return AArch64::LDURSBWi;
2753 case AArch64::LDRSHXui:
return AArch64::LDURSHXi;
2754 case AArch64::LDRSHWui:
return AArch64::LDURSHWi;
2755 case AArch64::LDRSWui:
return AArch64::LDURSWi;
2756 case AArch64::STRXui:
return AArch64::STURXi;
2757 case AArch64::STRWui:
return AArch64::STURWi;
2758 case AArch64::STRBui:
return AArch64::STURBi;
2759 case AArch64::STRHui:
return AArch64::STURHi;
2760 case AArch64::STRSui:
return AArch64::STURSi;
2761 case AArch64::STRDui:
return AArch64::STURDi;
2762 case AArch64::STRQui:
return AArch64::STURQi;
2763 case AArch64::STRBBui:
return AArch64::STURBBi;
2764 case AArch64::STRHHui:
return AArch64::STURHHi;
2773 case AArch64::LDAPURBi:
2774 case AArch64::LDAPURHi:
2775 case AArch64::LDAPURi:
2776 case AArch64::LDAPURSBWi:
2777 case AArch64::LDAPURSBXi:
2778 case AArch64::LDAPURSHWi:
2779 case AArch64::LDAPURSHXi:
2780 case AArch64::LDAPURSWi:
2781 case AArch64::LDAPURXi:
2782 case AArch64::LDR_PPXI:
2783 case AArch64::LDR_PXI:
2784 case AArch64::LDR_ZXI:
2785 case AArch64::LDR_ZZXI:
2786 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
2787 case AArch64::LDR_ZZZXI:
2788 case AArch64::LDR_ZZZZXI:
2789 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
2790 case AArch64::LDRBBui:
2791 case AArch64::LDRBui:
2792 case AArch64::LDRDui:
2793 case AArch64::LDRHHui:
2794 case AArch64::LDRHui:
2795 case AArch64::LDRQui:
2796 case AArch64::LDRSBWui:
2797 case AArch64::LDRSBXui:
2798 case AArch64::LDRSHWui:
2799 case AArch64::LDRSHXui:
2800 case AArch64::LDRSui:
2801 case AArch64::LDRSWui:
2802 case AArch64::LDRWui:
2803 case AArch64::LDRXui:
2804 case AArch64::LDURBBi:
2805 case AArch64::LDURBi:
2806 case AArch64::LDURDi:
2807 case AArch64::LDURHHi:
2808 case AArch64::LDURHi:
2809 case AArch64::LDURQi:
2810 case AArch64::LDURSBWi:
2811 case AArch64::LDURSBXi:
2812 case AArch64::LDURSHWi:
2813 case AArch64::LDURSHXi:
2814 case AArch64::LDURSi:
2815 case AArch64::LDURSWi:
2816 case AArch64::LDURWi:
2817 case AArch64::LDURXi:
2818 case AArch64::PRFMui:
2819 case AArch64::PRFUMi:
2820 case AArch64::ST2Gi:
2822 case AArch64::STLURBi:
2823 case AArch64::STLURHi:
2824 case AArch64::STLURWi:
2825 case AArch64::STLURXi:
2826 case AArch64::StoreSwiftAsyncContext:
2827 case AArch64::STR_PPXI:
2828 case AArch64::STR_PXI:
2829 case AArch64::STR_ZXI:
2830 case AArch64::STR_ZZXI:
2831 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
2832 case AArch64::STR_ZZZXI:
2833 case AArch64::STR_ZZZZXI:
2834 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
2835 case AArch64::STRBBui:
2836 case AArch64::STRBui:
2837 case AArch64::STRDui:
2838 case AArch64::STRHHui:
2839 case AArch64::STRHui:
2840 case AArch64::STRQui:
2841 case AArch64::STRSui:
2842 case AArch64::STRWui:
2843 case AArch64::STRXui:
2844 case AArch64::STURBBi:
2845 case AArch64::STURBi:
2846 case AArch64::STURDi:
2847 case AArch64::STURHHi:
2848 case AArch64::STURHi:
2849 case AArch64::STURQi:
2850 case AArch64::STURSi:
2851 case AArch64::STURWi:
2852 case AArch64::STURXi:
2853 case AArch64::STZ2Gi:
2854 case AArch64::STZGi:
2855 case AArch64::TAGPstack:
2857 case AArch64::LD1B_D_IMM:
2858 case AArch64::LD1B_H_IMM:
2859 case AArch64::LD1B_IMM:
2860 case AArch64::LD1B_S_IMM:
2861 case AArch64::LD1D_IMM:
2862 case AArch64::LD1H_D_IMM:
2863 case AArch64::LD1H_IMM:
2864 case AArch64::LD1H_S_IMM:
2865 case AArch64::LD1RB_D_IMM:
2866 case AArch64::LD1RB_H_IMM:
2867 case AArch64::LD1RB_IMM:
2868 case AArch64::LD1RB_S_IMM:
2869 case AArch64::LD1RD_IMM:
2870 case AArch64::LD1RH_D_IMM:
2871 case AArch64::LD1RH_IMM:
2872 case AArch64::LD1RH_S_IMM:
2873 case AArch64::LD1RSB_D_IMM:
2874 case AArch64::LD1RSB_H_IMM:
2875 case AArch64::LD1RSB_S_IMM:
2876 case AArch64::LD1RSH_D_IMM:
2877 case AArch64::LD1RSH_S_IMM:
2878 case AArch64::LD1RSW_IMM:
2879 case AArch64::LD1RW_D_IMM:
2880 case AArch64::LD1RW_IMM:
2881 case AArch64::LD1SB_D_IMM:
2882 case AArch64::LD1SB_H_IMM:
2883 case AArch64::LD1SB_S_IMM:
2884 case AArch64::LD1SH_D_IMM:
2885 case AArch64::LD1SH_S_IMM:
2886 case AArch64::LD1SW_D_IMM:
2887 case AArch64::LD1W_D_IMM:
2888 case AArch64::LD1W_IMM:
2889 case AArch64::LD2B_IMM:
2890 case AArch64::LD2D_IMM:
2891 case AArch64::LD2H_IMM:
2892 case AArch64::LD2W_IMM:
2893 case AArch64::LD3B_IMM:
2894 case AArch64::LD3D_IMM:
2895 case AArch64::LD3H_IMM:
2896 case AArch64::LD3W_IMM:
2897 case AArch64::LD4B_IMM:
2898 case AArch64::LD4D_IMM:
2899 case AArch64::LD4H_IMM:
2900 case AArch64::LD4W_IMM:
2902 case AArch64::LDNF1B_D_IMM:
2903 case AArch64::LDNF1B_H_IMM:
2904 case AArch64::LDNF1B_IMM:
2905 case AArch64::LDNF1B_S_IMM:
2906 case AArch64::LDNF1D_IMM:
2907 case AArch64::LDNF1H_D_IMM:
2908 case AArch64::LDNF1H_IMM:
2909 case AArch64::LDNF1H_S_IMM:
2910 case AArch64::LDNF1SB_D_IMM:
2911 case AArch64::LDNF1SB_H_IMM:
2912 case AArch64::LDNF1SB_S_IMM:
2913 case AArch64::LDNF1SH_D_IMM:
2914 case AArch64::LDNF1SH_S_IMM:
2915 case AArch64::LDNF1SW_D_IMM:
2916 case AArch64::LDNF1W_D_IMM:
2917 case AArch64::LDNF1W_IMM:
2918 case AArch64::LDNPDi:
2919 case AArch64::LDNPQi:
2920 case AArch64::LDNPSi:
2921 case AArch64::LDNPWi:
2922 case AArch64::LDNPXi:
2923 case AArch64::LDNT1B_ZRI:
2924 case AArch64::LDNT1D_ZRI:
2925 case AArch64::LDNT1H_ZRI:
2926 case AArch64::LDNT1W_ZRI:
2927 case AArch64::LDPDi:
2928 case AArch64::LDPQi:
2929 case AArch64::LDPSi:
2930 case AArch64::LDPWi:
2931 case AArch64::LDPXi:
2932 case AArch64::LDRBBpost:
2933 case AArch64::LDRBBpre:
2934 case AArch64::LDRBpost:
2935 case AArch64::LDRBpre:
2936 case AArch64::LDRDpost:
2937 case AArch64::LDRDpre:
2938 case AArch64::LDRHHpost:
2939 case AArch64::LDRHHpre:
2940 case AArch64::LDRHpost:
2941 case AArch64::LDRHpre:
2942 case AArch64::LDRQpost:
2943 case AArch64::LDRQpre:
2944 case AArch64::LDRSpost:
2945 case AArch64::LDRSpre:
2946 case AArch64::LDRWpost:
2947 case AArch64::LDRWpre:
2948 case AArch64::LDRXpost:
2949 case AArch64::LDRXpre:
2950 case AArch64::ST1B_D_IMM:
2951 case AArch64::ST1B_H_IMM:
2952 case AArch64::ST1B_IMM:
2953 case AArch64::ST1B_S_IMM:
2954 case AArch64::ST1D_IMM:
2955 case AArch64::ST1H_D_IMM:
2956 case AArch64::ST1H_IMM:
2957 case AArch64::ST1H_S_IMM:
2958 case AArch64::ST1W_D_IMM:
2959 case AArch64::ST1W_IMM:
2960 case AArch64::ST2B_IMM:
2961 case AArch64::ST2D_IMM:
2962 case AArch64::ST2H_IMM:
2963 case AArch64::ST2W_IMM:
2964 case AArch64::ST3B_IMM:
2965 case AArch64::ST3D_IMM:
2966 case AArch64::ST3H_IMM:
2967 case AArch64::ST3W_IMM:
2968 case AArch64::ST4B_IMM:
2969 case AArch64::ST4D_IMM:
2970 case AArch64::ST4H_IMM:
2971 case AArch64::ST4W_IMM:
2972 case AArch64::STGPi:
2973 case AArch64::STGPreIndex:
2974 case AArch64::STZGPreIndex:
2975 case AArch64::ST2GPreIndex:
2976 case AArch64::STZ2GPreIndex:
2977 case AArch64::STGPostIndex:
2978 case AArch64::STZGPostIndex:
2979 case AArch64::ST2GPostIndex:
2980 case AArch64::STZ2GPostIndex:
2981 case AArch64::STNPDi:
2982 case AArch64::STNPQi:
2983 case AArch64::STNPSi:
2984 case AArch64::STNPWi:
2985 case AArch64::STNPXi:
2986 case AArch64::STNT1B_ZRI:
2987 case AArch64::STNT1D_ZRI:
2988 case AArch64::STNT1H_ZRI:
2989 case AArch64::STNT1W_ZRI:
2990 case AArch64::STPDi:
2991 case AArch64::STPQi:
2992 case AArch64::STPSi:
2993 case AArch64::STPWi:
2994 case AArch64::STPXi:
2995 case AArch64::STRBBpost:
2996 case AArch64::STRBBpre:
2997 case AArch64::STRBpost:
2998 case AArch64::STRBpre:
2999 case AArch64::STRDpost:
3000 case AArch64::STRDpre:
3001 case AArch64::STRHHpost:
3002 case AArch64::STRHHpre:
3003 case AArch64::STRHpost:
3004 case AArch64::STRHpre:
3005 case AArch64::STRQpost:
3006 case AArch64::STRQpre:
3007 case AArch64::STRSpost:
3008 case AArch64::STRSpre:
3009 case AArch64::STRWpost:
3010 case AArch64::STRWpre:
3011 case AArch64::STRXpost:
3012 case AArch64::STRXpre:
3014 case AArch64::LDPDpost:
3015 case AArch64::LDPDpre:
3016 case AArch64::LDPQpost:
3017 case AArch64::LDPQpre:
3018 case AArch64::LDPSpost:
3019 case AArch64::LDPSpre:
3020 case AArch64::LDPWpost:
3021 case AArch64::LDPWpre:
3022 case AArch64::LDPXpost:
3023 case AArch64::LDPXpre:
3024 case AArch64::STGPpre:
3025 case AArch64::STGPpost:
3026 case AArch64::STPDpost:
3027 case AArch64::STPDpre:
3028 case AArch64::STPQpost:
3029 case AArch64::STPQpre:
3030 case AArch64::STPSpost:
3031 case AArch64::STPSpre:
3032 case AArch64::STPWpost:
3033 case AArch64::STPWpre:
3034 case AArch64::STPXpost:
3035 case AArch64::STPXpre:
3041 switch (
MI.getOpcode()) {
3045 case AArch64::STRSui:
3046 case AArch64::STRDui:
3047 case AArch64::STRQui:
3048 case AArch64::STRXui:
3049 case AArch64::STRWui:
3050 case AArch64::LDRSui:
3051 case AArch64::LDRDui:
3052 case AArch64::LDRQui:
3053 case AArch64::LDRXui:
3054 case AArch64::LDRWui:
3055 case AArch64::LDRSWui:
3057 case AArch64::STURSi:
3058 case AArch64::STRSpre:
3059 case AArch64::STURDi:
3060 case AArch64::STRDpre:
3061 case AArch64::STURQi:
3062 case AArch64::STRQpre:
3063 case AArch64::STURWi:
3064 case AArch64::STRWpre:
3065 case AArch64::STURXi:
3066 case AArch64::STRXpre:
3067 case AArch64::LDURSi:
3068 case AArch64::LDRSpre:
3069 case AArch64::LDURDi:
3070 case AArch64::LDRDpre:
3071 case AArch64::LDURQi:
3072 case AArch64::LDRQpre:
3073 case AArch64::LDURWi:
3074 case AArch64::LDRWpre:
3075 case AArch64::LDURXi:
3076 case AArch64::LDRXpre:
3077 case AArch64::LDURSWi:
3078 case AArch64::LDRSWpre:
3080 case AArch64::LDR_ZXI:
3081 case AArch64::STR_ZXI:
3087 switch (
MI.getOpcode()) {
3090 "Unexpected instruction - was a new tail call opcode introduced?");
3092 case AArch64::TCRETURNdi:
3093 case AArch64::TCRETURNri:
3094 case AArch64::TCRETURNrix16x17:
3095 case AArch64::TCRETURNrix17:
3096 case AArch64::TCRETURNrinotx16:
3097 case AArch64::TCRETURNriALL:
3098 case AArch64::AUTH_TCRETURN:
3099 case AArch64::AUTH_TCRETURN_BTI:
3109 case AArch64::ADDWri:
3110 return AArch64::ADDSWri;
3111 case AArch64::ADDWrr:
3112 return AArch64::ADDSWrr;
3113 case AArch64::ADDWrs:
3114 return AArch64::ADDSWrs;
3115 case AArch64::ADDWrx:
3116 return AArch64::ADDSWrx;
3117 case AArch64::ANDWri:
3118 return AArch64::ANDSWri;
3119 case AArch64::ANDWrr:
3120 return AArch64::ANDSWrr;
3121 case AArch64::ANDWrs:
3122 return AArch64::ANDSWrs;
3123 case AArch64::BICWrr:
3124 return AArch64::BICSWrr;
3125 case AArch64::BICWrs:
3126 return AArch64::BICSWrs;
3127 case AArch64::SUBWri:
3128 return AArch64::SUBSWri;
3129 case AArch64::SUBWrr:
3130 return AArch64::SUBSWrr;
3131 case AArch64::SUBWrs:
3132 return AArch64::SUBSWrs;
3133 case AArch64::SUBWrx:
3134 return AArch64::SUBSWrx;
3136 case AArch64::ADDXri:
3137 return AArch64::ADDSXri;
3138 case AArch64::ADDXrr:
3139 return AArch64::ADDSXrr;
3140 case AArch64::ADDXrs:
3141 return AArch64::ADDSXrs;
3142 case AArch64::ADDXrx:
3143 return AArch64::ADDSXrx;
3144 case AArch64::ANDXri:
3145 return AArch64::ANDSXri;
3146 case AArch64::ANDXrr:
3147 return AArch64::ANDSXrr;
3148 case AArch64::ANDXrs:
3149 return AArch64::ANDSXrs;
3150 case AArch64::BICXrr:
3151 return AArch64::BICSXrr;
3152 case AArch64::BICXrs:
3153 return AArch64::BICSXrs;
3154 case AArch64::SUBXri:
3155 return AArch64::SUBSXri;
3156 case AArch64::SUBXrr:
3157 return AArch64::SUBSXrr;
3158 case AArch64::SUBXrs:
3159 return AArch64::SUBSXrs;
3160 case AArch64::SUBXrx:
3161 return AArch64::SUBSXrx;
3163 case AArch64::AND_PPzPP:
3164 return AArch64::ANDS_PPzPP;
3165 case AArch64::BIC_PPzPP:
3166 return AArch64::BICS_PPzPP;
3167 case AArch64::EOR_PPzPP:
3168 return AArch64::EORS_PPzPP;
3169 case AArch64::NAND_PPzPP:
3170 return AArch64::NANDS_PPzPP;
3171 case AArch64::NOR_PPzPP:
3172 return AArch64::NORS_PPzPP;
3173 case AArch64::ORN_PPzPP:
3174 return AArch64::ORNS_PPzPP;
3175 case AArch64::ORR_PPzPP:
3176 return AArch64::ORRS_PPzPP;
3177 case AArch64::BRKA_PPzP:
3178 return AArch64::BRKAS_PPzP;
3179 case AArch64::BRKPA_PPzPP:
3180 return AArch64::BRKPAS_PPzPP;
3181 case AArch64::BRKB_PPzP:
3182 return AArch64::BRKBS_PPzP;
3183 case AArch64::BRKPB_PPzPP:
3184 return AArch64::BRKPBS_PPzPP;
3185 case AArch64::BRKN_PPzP:
3186 return AArch64::BRKNS_PPzP;
3187 case AArch64::RDFFR_PPz:
3188 return AArch64::RDFFRS_PPz;
3189 case AArch64::PTRUE_B:
3190 return AArch64::PTRUES_B;
3201 if (
MI.hasOrderedMemoryRef())
3206 assert((
MI.getOperand(IsPreLdSt ? 2 : 1).isReg() ||
3207 MI.getOperand(IsPreLdSt ? 2 : 1).isFI()) &&
3208 "Expected a reg or frame index operand.");
3212 bool IsImmPreLdSt = IsPreLdSt &&
MI.getOperand(3).isImm();
3214 if (!
MI.getOperand(2).isImm() && !IsImmPreLdSt)
3227 if (
MI.getOperand(1).isReg() && !IsPreLdSt) {
3228 Register BaseReg =
MI.getOperand(1).getReg();
3230 if (
MI.modifiesRegister(BaseReg,
TRI))
3236 switch (
MI.getOpcode()) {
3239 case AArch64::LDR_ZXI:
3240 case AArch64::STR_ZXI:
3241 if (!Subtarget.isLittleEndian() ||
3242 Subtarget.getSVEVectorSizeInBits() != 128)
3255 const MCAsmInfo *MAI =
MI.getMF()->getTarget().getMCAsmInfo();
3257 MI.getMF()->getFunction().needsUnwindTableEntry();
3263 if (Subtarget.isPaired128Slow()) {
3264 switch (
MI.getOpcode()) {
3267 case AArch64::LDURQi:
3268 case AArch64::STURQi:
3269 case AArch64::LDRQui:
3270 case AArch64::STRQui:
3297std::optional<ExtAddrMode>
3302 bool OffsetIsScalable;
3303 if (!getMemOperandWithOffset(MemI,
Base,
Offset, OffsetIsScalable,
TRI))
3304 return std::nullopt;
3307 return std::nullopt;
3322 int64_t OffsetScale = 1;
3327 case AArch64::LDURQi:
3328 case AArch64::STURQi:
3332 case AArch64::LDURDi:
3333 case AArch64::STURDi:
3334 case AArch64::LDURXi:
3335 case AArch64::STURXi:
3339 case AArch64::LDURWi:
3340 case AArch64::LDURSWi:
3341 case AArch64::STURWi:
3345 case AArch64::LDURHi:
3346 case AArch64::STURHi:
3347 case AArch64::LDURHHi:
3348 case AArch64::STURHHi:
3349 case AArch64::LDURSHXi:
3350 case AArch64::LDURSHWi:
3354 case AArch64::LDRBroX:
3355 case AArch64::LDRBBroX:
3356 case AArch64::LDRSBXroX:
3357 case AArch64::LDRSBWroX:
3358 case AArch64::STRBroX:
3359 case AArch64::STRBBroX:
3360 case AArch64::LDURBi:
3361 case AArch64::LDURBBi:
3362 case AArch64::LDURSBXi:
3363 case AArch64::LDURSBWi:
3364 case AArch64::STURBi:
3365 case AArch64::STURBBi:
3366 case AArch64::LDRBui:
3367 case AArch64::LDRBBui:
3368 case AArch64::LDRSBXui:
3369 case AArch64::LDRSBWui:
3370 case AArch64::STRBui:
3371 case AArch64::STRBBui:
3375 case AArch64::LDRQroX:
3376 case AArch64::STRQroX:
3377 case AArch64::LDRQui:
3378 case AArch64::STRQui:
3383 case AArch64::LDRDroX:
3384 case AArch64::STRDroX:
3385 case AArch64::LDRXroX:
3386 case AArch64::STRXroX:
3387 case AArch64::LDRDui:
3388 case AArch64::STRDui:
3389 case AArch64::LDRXui:
3390 case AArch64::STRXui:
3395 case AArch64::LDRWroX:
3396 case AArch64::LDRSWroX:
3397 case AArch64::STRWroX:
3398 case AArch64::LDRWui:
3399 case AArch64::LDRSWui:
3400 case AArch64::STRWui:
3405 case AArch64::LDRHroX:
3406 case AArch64::STRHroX:
3407 case AArch64::LDRHHroX:
3408 case AArch64::STRHHroX:
3409 case AArch64::LDRSHXroX:
3410 case AArch64::LDRSHWroX:
3411 case AArch64::LDRHui:
3412 case AArch64::STRHui:
3413 case AArch64::LDRHHui:
3414 case AArch64::STRHHui:
3415 case AArch64::LDRSHXui:
3416 case AArch64::LDRSHWui:
3424 if (BaseRegOp.
isReg() && BaseRegOp.
getReg() == Reg)
3448 case AArch64::SBFMXri:
3461 AM.
Scale = OffsetScale;
3466 case TargetOpcode::SUBREG_TO_REG: {
3479 if (!OffsetReg.
isVirtual() || !
MRI.hasOneNonDBGUse(OffsetReg))
3483 if (
DefMI.getOpcode() != AArch64::ORRWrs ||
3484 DefMI.getOperand(1).getReg() != AArch64::WZR ||
3485 DefMI.getOperand(3).getImm() != 0)
3492 AM.
Scale = OffsetScale;
3503 auto validateOffsetForLDP = [](
unsigned NumBytes, int64_t OldOffset,
3504 int64_t NewOffset) ->
bool {
3505 int64_t MinOffset, MaxOffset;
3522 return OldOffset < MinOffset || OldOffset > MaxOffset ||
3523 (NewOffset >= MinOffset && NewOffset <= MaxOffset);
3525 auto canFoldAddSubImmIntoAddrMode = [&](int64_t Disp) ->
bool {
3527 int64_t NewOffset = OldOffset + Disp;
3528 if (!isLegalAddressingMode(NumBytes, NewOffset, 0))
3532 if (!validateOffsetForLDP(NumBytes, OldOffset, NewOffset))
3542 auto canFoldAddRegIntoAddrMode =
3547 if ((
unsigned)Scale != Scale)
3549 if (!isLegalAddressingMode(NumBytes, 0, Scale))
3561 return (Opcode == AArch64::STURQi || Opcode == AArch64::STRQui) &&
3562 Subtarget.isSTRQroSlow();
3571 case AArch64::ADDXri:
3577 return canFoldAddSubImmIntoAddrMode(Disp);
3579 case AArch64::SUBXri:
3585 return canFoldAddSubImmIntoAddrMode(-Disp);
3587 case AArch64::ADDXrs: {
3600 if (Shift != 2 && Shift != 3 && Subtarget.hasAddrLSLSlow14())
3602 if (avoidSlowSTRQ(MemI))
3605 return canFoldAddRegIntoAddrMode(1ULL << Shift);
3608 case AArch64::ADDXrr:
3616 if (!OptSize && avoidSlowSTRQ(MemI))
3618 return canFoldAddRegIntoAddrMode(1);
3620 case AArch64::ADDXrx:
3628 if (!OptSize && avoidSlowSTRQ(MemI))
3637 return canFoldAddRegIntoAddrMode(
3652 case AArch64::LDURQi:
3653 case AArch64::LDRQui:
3654 return AArch64::LDRQroX;
3655 case AArch64::STURQi:
3656 case AArch64::STRQui:
3657 return AArch64::STRQroX;
3658 case AArch64::LDURDi:
3659 case AArch64::LDRDui:
3660 return AArch64::LDRDroX;
3661 case AArch64::STURDi:
3662 case AArch64::STRDui:
3663 return AArch64::STRDroX;
3664 case AArch64::LDURXi:
3665 case AArch64::LDRXui:
3666 return AArch64::LDRXroX;
3667 case AArch64::STURXi:
3668 case AArch64::STRXui:
3669 return AArch64::STRXroX;
3670 case AArch64::LDURWi:
3671 case AArch64::LDRWui:
3672 return AArch64::LDRWroX;
3673 case AArch64::LDURSWi:
3674 case AArch64::LDRSWui:
3675 return AArch64::LDRSWroX;
3676 case AArch64::STURWi:
3677 case AArch64::STRWui:
3678 return AArch64::STRWroX;
3679 case AArch64::LDURHi:
3680 case AArch64::LDRHui:
3681 return AArch64::LDRHroX;
3682 case AArch64::STURHi:
3683 case AArch64::STRHui:
3684 return AArch64::STRHroX;
3685 case AArch64::LDURHHi:
3686 case AArch64::LDRHHui:
3687 return AArch64::LDRHHroX;
3688 case AArch64::STURHHi:
3689 case AArch64::STRHHui:
3690 return AArch64::STRHHroX;
3691 case AArch64::LDURSHXi:
3692 case AArch64::LDRSHXui:
3693 return AArch64::LDRSHXroX;
3694 case AArch64::LDURSHWi:
3695 case AArch64::LDRSHWui:
3696 return AArch64::LDRSHWroX;
3697 case AArch64::LDURBi:
3698 case AArch64::LDRBui:
3699 return AArch64::LDRBroX;
3700 case AArch64::LDURBBi:
3701 case AArch64::LDRBBui:
3702 return AArch64::LDRBBroX;
3703 case AArch64::LDURSBXi:
3704 case AArch64::LDRSBXui:
3705 return AArch64::LDRSBXroX;
3706 case AArch64::LDURSBWi:
3707 case AArch64::LDRSBWui:
3708 return AArch64::LDRSBWroX;
3709 case AArch64::STURBi:
3710 case AArch64::STRBui:
3711 return AArch64::STRBroX;
3712 case AArch64::STURBBi:
3713 case AArch64::STRBBui:
3714 return AArch64::STRBBroX;
3726 case AArch64::LDURQi:
3728 return AArch64::LDRQui;
3729 case AArch64::STURQi:
3731 return AArch64::STRQui;
3732 case AArch64::LDURDi:
3734 return AArch64::LDRDui;
3735 case AArch64::STURDi:
3737 return AArch64::STRDui;
3738 case AArch64::LDURXi:
3740 return AArch64::LDRXui;
3741 case AArch64::STURXi:
3743 return AArch64::STRXui;
3744 case AArch64::LDURWi:
3746 return AArch64::LDRWui;
3747 case AArch64::LDURSWi:
3749 return AArch64::LDRSWui;
3750 case AArch64::STURWi:
3752 return AArch64::STRWui;
3753 case AArch64::LDURHi:
3755 return AArch64::LDRHui;
3756 case AArch64::STURHi:
3758 return AArch64::STRHui;
3759 case AArch64::LDURHHi:
3761 return AArch64::LDRHHui;
3762 case AArch64::STURHHi:
3764 return AArch64::STRHHui;
3765 case AArch64::LDURSHXi:
3767 return AArch64::LDRSHXui;
3768 case AArch64::LDURSHWi:
3770 return AArch64::LDRSHWui;
3771 case AArch64::LDURBi:
3773 return AArch64::LDRBui;
3774 case AArch64::LDURBBi:
3776 return AArch64::LDRBBui;
3777 case AArch64::LDURSBXi:
3779 return AArch64::LDRSBXui;
3780 case AArch64::LDURSBWi:
3782 return AArch64::LDRSBWui;
3783 case AArch64::STURBi:
3785 return AArch64::STRBui;
3786 case AArch64::STURBBi:
3788 return AArch64::STRBBui;
3789 case AArch64::LDRQui:
3790 case AArch64::STRQui:
3793 case AArch64::LDRDui:
3794 case AArch64::STRDui:
3795 case AArch64::LDRXui:
3796 case AArch64::STRXui:
3799 case AArch64::LDRWui:
3800 case AArch64::LDRSWui:
3801 case AArch64::STRWui:
3804 case AArch64::LDRHui:
3805 case AArch64::STRHui:
3806 case AArch64::LDRHHui:
3807 case AArch64::STRHHui:
3808 case AArch64::LDRSHXui:
3809 case AArch64::LDRSHWui:
3812 case AArch64::LDRBui:
3813 case AArch64::LDRBBui:
3814 case AArch64::LDRSBXui:
3815 case AArch64::LDRSBWui:
3816 case AArch64::STRBui:
3817 case AArch64::STRBBui:
3831 case AArch64::LDURQi:
3832 case AArch64::STURQi:
3833 case AArch64::LDURDi:
3834 case AArch64::STURDi:
3835 case AArch64::LDURXi:
3836 case AArch64::STURXi:
3837 case AArch64::LDURWi:
3838 case AArch64::LDURSWi:
3839 case AArch64::STURWi:
3840 case AArch64::LDURHi:
3841 case AArch64::STURHi:
3842 case AArch64::LDURHHi:
3843 case AArch64::STURHHi:
3844 case AArch64::LDURSHXi:
3845 case AArch64::LDURSHWi:
3846 case AArch64::LDURBi:
3847 case AArch64::STURBi:
3848 case AArch64::LDURBBi:
3849 case AArch64::STURBBi:
3850 case AArch64::LDURSBWi:
3851 case AArch64::LDURSBXi:
3853 case AArch64::LDRQui:
3854 return AArch64::LDURQi;
3855 case AArch64::STRQui:
3856 return AArch64::STURQi;
3857 case AArch64::LDRDui:
3858 return AArch64::LDURDi;
3859 case AArch64::STRDui:
3860 return AArch64::STURDi;
3861 case AArch64::LDRXui:
3862 return AArch64::LDURXi;
3863 case AArch64::STRXui:
3864 return AArch64::STURXi;
3865 case AArch64::LDRWui:
3866 return AArch64::LDURWi;
3867 case AArch64::LDRSWui:
3868 return AArch64::LDURSWi;
3869 case AArch64::STRWui:
3870 return AArch64::STURWi;
3871 case AArch64::LDRHui:
3872 return AArch64::LDURHi;
3873 case AArch64::STRHui:
3874 return AArch64::STURHi;
3875 case AArch64::LDRHHui:
3876 return AArch64::LDURHHi;
3877 case AArch64::STRHHui:
3878 return AArch64::STURHHi;
3879 case AArch64::LDRSHXui:
3880 return AArch64::LDURSHXi;
3881 case AArch64::LDRSHWui:
3882 return AArch64::LDURSHWi;
3883 case AArch64::LDRBBui:
3884 return AArch64::LDURBBi;
3885 case AArch64::LDRBui:
3886 return AArch64::LDURBi;
3887 case AArch64::STRBBui:
3888 return AArch64::STURBBi;
3889 case AArch64::STRBui:
3890 return AArch64::STURBi;
3891 case AArch64::LDRSBWui:
3892 return AArch64::LDURSBWi;
3893 case AArch64::LDRSBXui:
3894 return AArch64::LDURSBXi;
3907 case AArch64::LDRQroX:
3908 case AArch64::LDURQi:
3909 case AArch64::LDRQui:
3910 return AArch64::LDRQroW;
3911 case AArch64::STRQroX:
3912 case AArch64::STURQi:
3913 case AArch64::STRQui:
3914 return AArch64::STRQroW;
3915 case AArch64::LDRDroX:
3916 case AArch64::LDURDi:
3917 case AArch64::LDRDui:
3918 return AArch64::LDRDroW;
3919 case AArch64::STRDroX:
3920 case AArch64::STURDi:
3921 case AArch64::STRDui:
3922 return AArch64::STRDroW;
3923 case AArch64::LDRXroX:
3924 case AArch64::LDURXi:
3925 case AArch64::LDRXui:
3926 return AArch64::LDRXroW;
3927 case AArch64::STRXroX:
3928 case AArch64::STURXi:
3929 case AArch64::STRXui:
3930 return AArch64::STRXroW;
3931 case AArch64::LDRWroX:
3932 case AArch64::LDURWi:
3933 case AArch64::LDRWui:
3934 return AArch64::LDRWroW;
3935 case AArch64::LDRSWroX:
3936 case AArch64::LDURSWi:
3937 case AArch64::LDRSWui:
3938 return AArch64::LDRSWroW;
3939 case AArch64::STRWroX:
3940 case AArch64::STURWi:
3941 case AArch64::STRWui:
3942 return AArch64::STRWroW;
3943 case AArch64::LDRHroX:
3944 case AArch64::LDURHi:
3945 case AArch64::LDRHui:
3946 return AArch64::LDRHroW;
3947 case AArch64::STRHroX:
3948 case AArch64::STURHi:
3949 case AArch64::STRHui:
3950 return AArch64::STRHroW;
3951 case AArch64::LDRHHroX:
3952 case AArch64::LDURHHi:
3953 case AArch64::LDRHHui:
3954 return AArch64::LDRHHroW;
3955 case AArch64::STRHHroX:
3956 case AArch64::STURHHi:
3957 case AArch64::STRHHui:
3958 return AArch64::STRHHroW;
3959 case AArch64::LDRSHXroX:
3960 case AArch64::LDURSHXi:
3961 case AArch64::LDRSHXui:
3962 return AArch64::LDRSHXroW;
3963 case AArch64::LDRSHWroX:
3964 case AArch64::LDURSHWi:
3965 case AArch64::LDRSHWui:
3966 return AArch64::LDRSHWroW;
3967 case AArch64::LDRBroX:
3968 case AArch64::LDURBi:
3969 case AArch64::LDRBui:
3970 return AArch64::LDRBroW;
3971 case AArch64::LDRBBroX:
3972 case AArch64::LDURBBi:
3973 case AArch64::LDRBBui:
3974 return AArch64::LDRBBroW;
3975 case AArch64::LDRSBXroX:
3976 case AArch64::LDURSBXi:
3977 case AArch64::LDRSBXui:
3978 return AArch64::LDRSBXroW;
3979 case AArch64::LDRSBWroX:
3980 case AArch64::LDURSBWi:
3981 case AArch64::LDRSBWui:
3982 return AArch64::LDRSBWroW;
3983 case AArch64::STRBroX:
3984 case AArch64::STURBi:
3985 case AArch64::STRBui:
3986 return AArch64::STRBroW;
3987 case AArch64::STRBBroX:
3988 case AArch64::STURBBi:
3989 case AArch64::STRBBui:
3990 return AArch64::STRBBroW;
4005 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4015 return B.getInstr();
4019 "Addressing mode not supported for folding");
4036 return B.getInstr();
4043 "Address offset can be a register or an immediate, but not both");
4045 MRI.constrainRegClass(AM.
BaseReg, &AArch64::GPR64spRegClass);
4050 OffsetReg =
MRI.createVirtualRegister(&AArch64::GPR32RegClass);
4064 return B.getInstr();
4068 "Function must not be called with an addressing mode it can't handle");
4077 case AArch64::LD1Fourv16b_POST:
4078 case AArch64::LD1Fourv1d_POST:
4079 case AArch64::LD1Fourv2d_POST:
4080 case AArch64::LD1Fourv2s_POST:
4081 case AArch64::LD1Fourv4h_POST:
4082 case AArch64::LD1Fourv4s_POST:
4083 case AArch64::LD1Fourv8b_POST:
4084 case AArch64::LD1Fourv8h_POST:
4085 case AArch64::LD1Onev16b_POST:
4086 case AArch64::LD1Onev1d_POST:
4087 case AArch64::LD1Onev2d_POST:
4088 case AArch64::LD1Onev2s_POST:
4089 case AArch64::LD1Onev4h_POST:
4090 case AArch64::LD1Onev4s_POST:
4091 case AArch64::LD1Onev8b_POST:
4092 case AArch64::LD1Onev8h_POST:
4093 case AArch64::LD1Rv16b_POST:
4094 case AArch64::LD1Rv1d_POST:
4095 case AArch64::LD1Rv2d_POST:
4096 case AArch64::LD1Rv2s_POST:
4097 case AArch64::LD1Rv4h_POST:
4098 case AArch64::LD1Rv4s_POST:
4099 case AArch64::LD1Rv8b_POST:
4100 case AArch64::LD1Rv8h_POST:
4101 case AArch64::LD1Threev16b_POST:
4102 case AArch64::LD1Threev1d_POST:
4103 case AArch64::LD1Threev2d_POST:
4104 case AArch64::LD1Threev2s_POST:
4105 case AArch64::LD1Threev4h_POST:
4106 case AArch64::LD1Threev4s_POST:
4107 case AArch64::LD1Threev8b_POST:
4108 case AArch64::LD1Threev8h_POST:
4109 case AArch64::LD1Twov16b_POST:
4110 case AArch64::LD1Twov1d_POST:
4111 case AArch64::LD1Twov2d_POST:
4112 case AArch64::LD1Twov2s_POST:
4113 case AArch64::LD1Twov4h_POST:
4114 case AArch64::LD1Twov4s_POST:
4115 case AArch64::LD1Twov8b_POST:
4116 case AArch64::LD1Twov8h_POST:
4117 case AArch64::LD1i16_POST:
4118 case AArch64::LD1i32_POST:
4119 case AArch64::LD1i64_POST:
4120 case AArch64::LD1i8_POST:
4121 case AArch64::LD2Rv16b_POST:
4122 case AArch64::LD2Rv1d_POST:
4123 case AArch64::LD2Rv2d_POST:
4124 case AArch64::LD2Rv2s_POST:
4125 case AArch64::LD2Rv4h_POST:
4126 case AArch64::LD2Rv4s_POST:
4127 case AArch64::LD2Rv8b_POST:
4128 case AArch64::LD2Rv8h_POST:
4129 case AArch64::LD2Twov16b_POST:
4130 case AArch64::LD2Twov2d_POST:
4131 case AArch64::LD2Twov2s_POST:
4132 case AArch64::LD2Twov4h_POST:
4133 case AArch64::LD2Twov4s_POST:
4134 case AArch64::LD2Twov8b_POST:
4135 case AArch64::LD2Twov8h_POST:
4136 case AArch64::LD2i16_POST:
4137 case AArch64::LD2i32_POST:
4138 case AArch64::LD2i64_POST:
4139 case AArch64::LD2i8_POST:
4140 case AArch64::LD3Rv16b_POST:
4141 case AArch64::LD3Rv1d_POST:
4142 case AArch64::LD3Rv2d_POST:
4143 case AArch64::LD3Rv2s_POST:
4144 case AArch64::LD3Rv4h_POST:
4145 case AArch64::LD3Rv4s_POST:
4146 case AArch64::LD3Rv8b_POST:
4147 case AArch64::LD3Rv8h_POST:
4148 case AArch64::LD3Threev16b_POST:
4149 case AArch64::LD3Threev2d_POST:
4150 case AArch64::LD3Threev2s_POST:
4151 case AArch64::LD3Threev4h_POST:
4152 case AArch64::LD3Threev4s_POST:
4153 case AArch64::LD3Threev8b_POST:
4154 case AArch64::LD3Threev8h_POST:
4155 case AArch64::LD3i16_POST:
4156 case AArch64::LD3i32_POST:
4157 case AArch64::LD3i64_POST:
4158 case AArch64::LD3i8_POST:
4159 case AArch64::LD4Fourv16b_POST:
4160 case AArch64::LD4Fourv2d_POST:
4161 case AArch64::LD4Fourv2s_POST:
4162 case AArch64::LD4Fourv4h_POST:
4163 case AArch64::LD4Fourv4s_POST:
4164 case AArch64::LD4Fourv8b_POST:
4165 case AArch64::LD4Fourv8h_POST:
4166 case AArch64::LD4Rv16b_POST:
4167 case AArch64::LD4Rv1d_POST:
4168 case AArch64::LD4Rv2d_POST:
4169 case AArch64::LD4Rv2s_POST:
4170 case AArch64::LD4Rv4h_POST:
4171 case AArch64::LD4Rv4s_POST:
4172 case AArch64::LD4Rv8b_POST:
4173 case AArch64::LD4Rv8h_POST:
4174 case AArch64::LD4i16_POST:
4175 case AArch64::LD4i32_POST:
4176 case AArch64::LD4i64_POST:
4177 case AArch64::LD4i8_POST:
4178 case AArch64::LDAPRWpost:
4179 case AArch64::LDAPRXpost:
4180 case AArch64::LDIAPPWpost:
4181 case AArch64::LDIAPPXpost:
4182 case AArch64::LDPDpost:
4183 case AArch64::LDPQpost:
4184 case AArch64::LDPSWpost:
4185 case AArch64::LDPSpost:
4186 case AArch64::LDPWpost:
4187 case AArch64::LDPXpost:
4188 case AArch64::LDRBBpost:
4189 case AArch64::LDRBpost:
4190 case AArch64::LDRDpost:
4191 case AArch64::LDRHHpost:
4192 case AArch64::LDRHpost:
4193 case AArch64::LDRQpost:
4194 case AArch64::LDRSBWpost:
4195 case AArch64::LDRSBXpost:
4196 case AArch64::LDRSHWpost:
4197 case AArch64::LDRSHXpost:
4198 case AArch64::LDRSWpost:
4199 case AArch64::LDRSpost:
4200 case AArch64::LDRWpost:
4201 case AArch64::LDRXpost:
4202 case AArch64::ST1Fourv16b_POST:
4203 case AArch64::ST1Fourv1d_POST:
4204 case AArch64::ST1Fourv2d_POST:
4205 case AArch64::ST1Fourv2s_POST:
4206 case AArch64::ST1Fourv4h_POST:
4207 case AArch64::ST1Fourv4s_POST:
4208 case AArch64::ST1Fourv8b_POST:
4209 case AArch64::ST1Fourv8h_POST:
4210 case AArch64::ST1Onev16b_POST:
4211 case AArch64::ST1Onev1d_POST:
4212 case AArch64::ST1Onev2d_POST:
4213 case AArch64::ST1Onev2s_POST:
4214 case AArch64::ST1Onev4h_POST:
4215 case AArch64::ST1Onev4s_POST:
4216 case AArch64::ST1Onev8b_POST:
4217 case AArch64::ST1Onev8h_POST:
4218 case AArch64::ST1Threev16b_POST:
4219 case AArch64::ST1Threev1d_POST:
4220 case AArch64::ST1Threev2d_POST:
4221 case AArch64::ST1Threev2s_POST:
4222 case AArch64::ST1Threev4h_POST:
4223 case AArch64::ST1Threev4s_POST:
4224 case AArch64::ST1Threev8b_POST:
4225 case AArch64::ST1Threev8h_POST:
4226 case AArch64::ST1Twov16b_POST:
4227 case AArch64::ST1Twov1d_POST:
4228 case AArch64::ST1Twov2d_POST:
4229 case AArch64::ST1Twov2s_POST:
4230 case AArch64::ST1Twov4h_POST:
4231 case AArch64::ST1Twov4s_POST:
4232 case AArch64::ST1Twov8b_POST:
4233 case AArch64::ST1Twov8h_POST:
4234 case AArch64::ST1i16_POST:
4235 case AArch64::ST1i32_POST:
4236 case AArch64::ST1i64_POST:
4237 case AArch64::ST1i8_POST:
4238 case AArch64::ST2GPostIndex:
4239 case AArch64::ST2Twov16b_POST:
4240 case AArch64::ST2Twov2d_POST:
4241 case AArch64::ST2Twov2s_POST:
4242 case AArch64::ST2Twov4h_POST:
4243 case AArch64::ST2Twov4s_POST:
4244 case AArch64::ST2Twov8b_POST:
4245 case AArch64::ST2Twov8h_POST:
4246 case AArch64::ST2i16_POST:
4247 case AArch64::ST2i32_POST:
4248 case AArch64::ST2i64_POST:
4249 case AArch64::ST2i8_POST:
4250 case AArch64::ST3Threev16b_POST:
4251 case AArch64::ST3Threev2d_POST:
4252 case AArch64::ST3Threev2s_POST:
4253 case AArch64::ST3Threev4h_POST:
4254 case AArch64::ST3Threev4s_POST:
4255 case AArch64::ST3Threev8b_POST:
4256 case AArch64::ST3Threev8h_POST:
4257 case AArch64::ST3i16_POST:
4258 case AArch64::ST3i32_POST:
4259 case AArch64::ST3i64_POST:
4260 case AArch64::ST3i8_POST:
4261 case AArch64::ST4Fourv16b_POST:
4262 case AArch64::ST4Fourv2d_POST:
4263 case AArch64::ST4Fourv2s_POST:
4264 case AArch64::ST4Fourv4h_POST:
4265 case AArch64::ST4Fourv4s_POST:
4266 case AArch64::ST4Fourv8b_POST:
4267 case AArch64::ST4Fourv8h_POST:
4268 case AArch64::ST4i16_POST:
4269 case AArch64::ST4i32_POST:
4270 case AArch64::ST4i64_POST:
4271 case AArch64::ST4i8_POST:
4272 case AArch64::STGPostIndex:
4273 case AArch64::STGPpost:
4274 case AArch64::STPDpost:
4275 case AArch64::STPQpost:
4276 case AArch64::STPSpost:
4277 case AArch64::STPWpost:
4278 case AArch64::STPXpost:
4279 case AArch64::STRBBpost:
4280 case AArch64::STRBpost:
4281 case AArch64::STRDpost:
4282 case AArch64::STRHHpost:
4283 case AArch64::STRHpost:
4284 case AArch64::STRQpost:
4285 case AArch64::STRSpost:
4286 case AArch64::STRWpost:
4287 case AArch64::STRXpost:
4288 case AArch64::STZ2GPostIndex:
4289 case AArch64::STZGPostIndex:
4296 bool &OffsetIsScalable,
TypeSize &Width,
4317 int64_t Dummy1, Dummy2;
4339 return BaseOp->
isReg() || BaseOp->
isFI();
4346 assert(OfsOp.
isImm() &&
"Offset operand wasn't immediate.");
4351 TypeSize &Width, int64_t &MinOffset,
4352 int64_t &MaxOffset) {
4358 MinOffset = MaxOffset = 0;
4361 case AArch64::LDRQui:
4362 case AArch64::STRQui:
4368 case AArch64::LDRXui:
4369 case AArch64::LDRDui:
4370 case AArch64::STRXui:
4371 case AArch64::STRDui:
4372 case AArch64::PRFMui:
4378 case AArch64::LDRWui:
4379 case AArch64::LDRSui:
4380 case AArch64::LDRSWui:
4381 case AArch64::STRWui:
4382 case AArch64::STRSui:
4388 case AArch64::LDRHui:
4389 case AArch64::LDRHHui:
4390 case AArch64::LDRSHWui:
4391 case AArch64::LDRSHXui:
4392 case AArch64::STRHui:
4393 case AArch64::STRHHui:
4399 case AArch64::LDRBui:
4400 case AArch64::LDRBBui:
4401 case AArch64::LDRSBWui:
4402 case AArch64::LDRSBXui:
4403 case AArch64::STRBui:
4404 case AArch64::STRBBui:
4411 case AArch64::STRQpre:
4412 case AArch64::LDRQpost:
4418 case AArch64::LDRDpost:
4419 case AArch64::LDRDpre:
4420 case AArch64::LDRXpost:
4421 case AArch64::LDRXpre:
4422 case AArch64::STRDpost:
4423 case AArch64::STRDpre:
4424 case AArch64::STRXpost:
4425 case AArch64::STRXpre:
4431 case AArch64::STRWpost:
4432 case AArch64::STRWpre:
4433 case AArch64::LDRWpost:
4434 case AArch64::LDRWpre:
4435 case AArch64::STRSpost:
4436 case AArch64::STRSpre:
4437 case AArch64::LDRSpost:
4438 case AArch64::LDRSpre:
4444 case AArch64::LDRHpost:
4445 case AArch64::LDRHpre:
4446 case AArch64::STRHpost:
4447 case AArch64::STRHpre:
4448 case AArch64::LDRHHpost:
4449 case AArch64::LDRHHpre:
4450 case AArch64::STRHHpost:
4451 case AArch64::STRHHpre:
4457 case AArch64::LDRBpost:
4458 case AArch64::LDRBpre:
4459 case AArch64::STRBpost:
4460 case AArch64::STRBpre:
4461 case AArch64::LDRBBpost:
4462 case AArch64::LDRBBpre:
4463 case AArch64::STRBBpost:
4464 case AArch64::STRBBpre:
4471 case AArch64::LDURQi:
4472 case AArch64::STURQi:
4478 case AArch64::LDURXi:
4479 case AArch64::LDURDi:
4480 case AArch64::LDAPURXi:
4481 case AArch64::STURXi:
4482 case AArch64::STURDi:
4483 case AArch64::STLURXi:
4484 case AArch64::PRFUMi:
4490 case AArch64::LDURWi:
4491 case AArch64::LDURSi:
4492 case AArch64::LDURSWi:
4493 case AArch64::LDAPURi:
4494 case AArch64::LDAPURSWi:
4495 case AArch64::STURWi:
4496 case AArch64::STURSi:
4497 case AArch64::STLURWi:
4503 case AArch64::LDURHi:
4504 case AArch64::LDURHHi:
4505 case AArch64::LDURSHXi:
4506 case AArch64::LDURSHWi:
4507 case AArch64::LDAPURHi:
4508 case AArch64::LDAPURSHWi:
4509 case AArch64::LDAPURSHXi:
4510 case AArch64::STURHi:
4511 case AArch64::STURHHi:
4512 case AArch64::STLURHi:
4518 case AArch64::LDURBi:
4519 case AArch64::LDURBBi:
4520 case AArch64::LDURSBXi:
4521 case AArch64::LDURSBWi:
4522 case AArch64::LDAPURBi:
4523 case AArch64::LDAPURSBWi:
4524 case AArch64::LDAPURSBXi:
4525 case AArch64::STURBi:
4526 case AArch64::STURBBi:
4527 case AArch64::STLURBi:
4534 case AArch64::LDPQi:
4535 case AArch64::LDNPQi:
4536 case AArch64::STPQi:
4537 case AArch64::STNPQi:
4538 case AArch64::LDPQpost:
4539 case AArch64::LDPQpre:
4540 case AArch64::STPQpost:
4541 case AArch64::STPQpre:
4547 case AArch64::LDPXi:
4548 case AArch64::LDPDi:
4549 case AArch64::LDNPXi:
4550 case AArch64::LDNPDi:
4551 case AArch64::STPXi:
4552 case AArch64::STPDi:
4553 case AArch64::STNPXi:
4554 case AArch64::STNPDi:
4555 case AArch64::LDPDpost:
4556 case AArch64::LDPDpre:
4557 case AArch64::LDPXpost:
4558 case AArch64::LDPXpre:
4559 case AArch64::STPDpost:
4560 case AArch64::STPDpre:
4561 case AArch64::STPXpost:
4562 case AArch64::STPXpre:
4568 case AArch64::LDPWi:
4569 case AArch64::LDPSi:
4570 case AArch64::LDNPWi:
4571 case AArch64::LDNPSi:
4572 case AArch64::STPWi:
4573 case AArch64::STPSi:
4574 case AArch64::STNPWi:
4575 case AArch64::STNPSi:
4576 case AArch64::LDPSpost:
4577 case AArch64::LDPSpre:
4578 case AArch64::LDPWpost:
4579 case AArch64::LDPWpre:
4580 case AArch64::STPSpost:
4581 case AArch64::STPSpre:
4582 case AArch64::STPWpost:
4583 case AArch64::STPWpre:
4589 case AArch64::StoreSwiftAsyncContext:
4602 case AArch64::TAGPstack:
4612 case AArch64::STGPreIndex:
4613 case AArch64::STGPostIndex:
4614 case AArch64::STZGi:
4615 case AArch64::STZGPreIndex:
4616 case AArch64::STZGPostIndex:
4623 case AArch64::STR_ZZZZXI:
4624 case AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS:
4625 case AArch64::LDR_ZZZZXI:
4626 case AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS:
4632 case AArch64::STR_ZZZXI:
4633 case AArch64::LDR_ZZZXI:
4639 case AArch64::STR_ZZXI:
4640 case AArch64::STR_ZZXI_STRIDED_CONTIGUOUS:
4641 case AArch64::LDR_ZZXI:
4642 case AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS:
4648 case AArch64::LDR_PXI:
4649 case AArch64::STR_PXI:
4655 case AArch64::LDR_PPXI:
4656 case AArch64::STR_PPXI:
4662 case AArch64::LDR_ZXI:
4663 case AArch64::STR_ZXI:
4669 case AArch64::LD1B_IMM:
4670 case AArch64::LD1H_IMM:
4671 case AArch64::LD1W_IMM:
4672 case AArch64::LD1D_IMM:
4673 case AArch64::LDNT1B_ZRI:
4674 case AArch64::LDNT1H_ZRI:
4675 case AArch64::LDNT1W_ZRI:
4676 case AArch64::LDNT1D_ZRI:
4677 case AArch64::ST1B_IMM:
4678 case AArch64::ST1H_IMM:
4679 case AArch64::ST1W_IMM:
4680 case AArch64::ST1D_IMM:
4681 case AArch64::STNT1B_ZRI:
4682 case AArch64::STNT1H_ZRI:
4683 case AArch64::STNT1W_ZRI:
4684 case AArch64::STNT1D_ZRI:
4685 case AArch64::LDNF1B_IMM:
4686 case AArch64::LDNF1H_IMM:
4687 case AArch64::LDNF1W_IMM:
4688 case AArch64::LDNF1D_IMM:
4696 case AArch64::LD2B_IMM:
4697 case AArch64::LD2H_IMM:
4698 case AArch64::LD2W_IMM:
4699 case AArch64::LD2D_IMM:
4700 case AArch64::ST2B_IMM:
4701 case AArch64::ST2H_IMM:
4702 case AArch64::ST2W_IMM:
4703 case AArch64::ST2D_IMM:
4709 case AArch64::LD3B_IMM:
4710 case AArch64::LD3H_IMM:
4711 case AArch64::LD3W_IMM:
4712 case AArch64::LD3D_IMM:
4713 case AArch64::ST3B_IMM:
4714 case AArch64::ST3H_IMM:
4715 case AArch64::ST3W_IMM:
4716 case AArch64::ST3D_IMM:
4722 case AArch64::LD4B_IMM:
4723 case AArch64::LD4H_IMM:
4724 case AArch64::LD4W_IMM:
4725 case AArch64::LD4D_IMM:
4726 case AArch64::ST4B_IMM:
4727 case AArch64::ST4H_IMM:
4728 case AArch64::ST4W_IMM:
4729 case AArch64::ST4D_IMM:
4735 case AArch64::LD1B_H_IMM:
4736 case AArch64::LD1SB_H_IMM:
4737 case AArch64::LD1H_S_IMM:
4738 case AArch64::LD1SH_S_IMM:
4739 case AArch64::LD1W_D_IMM:
4740 case AArch64::LD1SW_D_IMM:
4741 case AArch64::ST1B_H_IMM:
4742 case AArch64::ST1H_S_IMM:
4743 case AArch64::ST1W_D_IMM:
4744 case AArch64::LDNF1B_H_IMM:
4745 case AArch64::LDNF1SB_H_IMM:
4746 case AArch64::LDNF1H_S_IMM:
4747 case AArch64::LDNF1SH_S_IMM:
4748 case AArch64::LDNF1W_D_IMM:
4749 case AArch64::LDNF1SW_D_IMM:
4757 case AArch64::LD1B_S_IMM:
4758 case AArch64::LD1SB_S_IMM:
4759 case AArch64::LD1H_D_IMM:
4760 case AArch64::LD1SH_D_IMM:
4761 case AArch64::ST1B_S_IMM:
4762 case AArch64::ST1H_D_IMM:
4763 case AArch64::LDNF1B_S_IMM:
4764 case AArch64::LDNF1SB_S_IMM:
4765 case AArch64::LDNF1H_D_IMM:
4766 case AArch64::LDNF1SH_D_IMM:
4774 case AArch64::LD1B_D_IMM:
4775 case AArch64::LD1SB_D_IMM:
4776 case AArch64::ST1B_D_IMM:
4777 case AArch64::LDNF1B_D_IMM:
4778 case AArch64::LDNF1SB_D_IMM:
4786 case AArch64::ST2Gi:
4787 case AArch64::ST2GPreIndex:
4788 case AArch64::ST2GPostIndex:
4789 case AArch64::STZ2Gi:
4790 case AArch64::STZ2GPreIndex:
4791 case AArch64::STZ2GPostIndex:
4797 case AArch64::STGPi:
4798 case AArch64::STGPpost:
4799 case AArch64::STGPpre:
4805 case AArch64::LD1RB_IMM:
4806 case AArch64::LD1RB_H_IMM:
4807 case AArch64::LD1RB_S_IMM:
4808 case AArch64::LD1RB_D_IMM:
4809 case AArch64::LD1RSB_H_IMM:
4810 case AArch64::LD1RSB_S_IMM:
4811 case AArch64::LD1RSB_D_IMM:
4817 case AArch64::LD1RH_IMM:
4818 case AArch64::LD1RH_S_IMM:
4819 case AArch64::LD1RH_D_IMM:
4820 case AArch64::LD1RSH_S_IMM:
4821 case AArch64::LD1RSH_D_IMM:
4827 case AArch64::LD1RW_IMM:
4828 case AArch64::LD1RW_D_IMM:
4829 case AArch64::LD1RSW_IMM:
4835 case AArch64::LD1RD_IMM:
4851 case AArch64::LDRBBui:
4852 case AArch64::LDURBBi:
4853 case AArch64::LDRSBWui:
4854 case AArch64::LDURSBWi:
4855 case AArch64::STRBBui:
4856 case AArch64::STURBBi:
4858 case AArch64::LDRHHui:
4859 case AArch64::LDURHHi:
4860 case AArch64::LDRSHWui:
4861 case AArch64::LDURSHWi:
4862 case AArch64::STRHHui:
4863 case AArch64::STURHHi:
4865 case AArch64::LDRSui:
4866 case AArch64::LDURSi:
4867 case AArch64::LDRSpre:
4868 case AArch64::LDRSWui:
4869 case AArch64::LDURSWi:
4870 case AArch64::LDRSWpre:
4871 case AArch64::LDRWpre:
4872 case AArch64::LDRWui:
4873 case AArch64::LDURWi:
4874 case AArch64::STRSui:
4875 case AArch64::STURSi:
4876 case AArch64::STRSpre:
4877 case AArch64::STRWui:
4878 case AArch64::STURWi:
4879 case AArch64::STRWpre:
4880 case AArch64::LDPSi:
4881 case AArch64::LDPSWi:
4882 case AArch64::LDPWi:
4883 case AArch64::STPSi:
4884 case AArch64::STPWi:
4886 case AArch64::LDRDui:
4887 case AArch64::LDURDi:
4888 case AArch64::LDRDpre:
4889 case AArch64::LDRXui:
4890 case AArch64::LDURXi:
4891 case AArch64::LDRXpre:
4892 case AArch64::STRDui:
4893 case AArch64::STURDi:
4894 case AArch64::STRDpre:
4895 case AArch64::STRXui:
4896 case AArch64::STURXi:
4897 case AArch64::STRXpre:
4898 case AArch64::LDPDi:
4899 case AArch64::LDPXi:
4900 case AArch64::STPDi:
4901 case AArch64::STPXi:
4903 case AArch64::LDRQui:
4904 case AArch64::LDURQi:
4905 case AArch64::STRQui:
4906 case AArch64::STURQi:
4907 case AArch64::STRQpre:
4908 case AArch64::LDPQi:
4909 case AArch64::LDRQpre:
4910 case AArch64::STPQi:
4912 case AArch64::STZGi:
4913 case AArch64::ST2Gi:
4914 case AArch64::STZ2Gi:
4915 case AArch64::STGPi:
4921 switch (
MI.getOpcode()) {
4924 case AArch64::LDRWpre:
4925 case AArch64::LDRXpre:
4926 case AArch64::LDRSWpre:
4927 case AArch64::LDRSpre:
4928 case AArch64::LDRDpre:
4929 case AArch64::LDRQpre:
4935 switch (
MI.getOpcode()) {
4938 case AArch64::STRWpre:
4939 case AArch64::STRXpre:
4940 case AArch64::STRSpre:
4941 case AArch64::STRDpre:
4942 case AArch64::STRQpre:
4952 switch (
MI.getOpcode()) {
4955 case AArch64::LDPSi:
4956 case AArch64::LDPSWi:
4957 case AArch64::LDPDi:
4958 case AArch64::LDPQi:
4959 case AArch64::LDPWi:
4960 case AArch64::LDPXi:
4961 case AArch64::STPSi:
4962 case AArch64::STPDi:
4963 case AArch64::STPQi:
4964 case AArch64::STPWi:
4965 case AArch64::STPXi:
4966 case AArch64::STGPi:
4972 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
4976 return MI.getOperand(Idx);
4981 assert(
MI.mayLoadOrStore() &&
"Load or store instruction expected");
4985 return MI.getOperand(Idx);
4990 switch (
MI.getOpcode()) {
4993 case AArch64::LDRBroX:
4994 case AArch64::LDRBBroX:
4995 case AArch64::LDRSBXroX:
4996 case AArch64::LDRSBWroX:
4997 case AArch64::LDRHroX:
4998 case AArch64::LDRHHroX:
4999 case AArch64::LDRSHXroX:
5000 case AArch64::LDRSHWroX:
5001 case AArch64::LDRWroX:
5002 case AArch64::LDRSroX:
5003 case AArch64::LDRSWroX:
5004 case AArch64::LDRDroX:
5005 case AArch64::LDRXroX:
5006 case AArch64::LDRQroX:
5007 return MI.getOperand(4);
5013 if (
MI.getParent() ==
nullptr)
5023 auto Reg =
Op.getReg();
5024 if (Reg.isPhysical())
5025 return AArch64::FPR16RegClass.contains(Reg);
5027 return TRC == &AArch64::FPR16RegClass ||
5028 TRC == &AArch64::FPR16_loRegClass;
5037 auto Reg =
Op.getReg();
5038 if (Reg.isPhysical())
5039 return AArch64::FPR128RegClass.contains(Reg);
5041 return TRC == &AArch64::FPR128RegClass ||
5042 TRC == &AArch64::FPR128_loRegClass;
5048 switch (
MI.getOpcode()) {
5051 case AArch64::PACIASP:
5052 case AArch64::PACIBSP:
5055 case AArch64::PAUTH_PROLOGUE:
5058 case AArch64::HINT: {
5059 unsigned Imm =
MI.getOperand(0).getImm();
5061 if (Imm == 32 || Imm == 34 || Imm == 36 || Imm == 38)
5064 if (Imm == 25 || Imm == 27)
5076 assert(Reg.isPhysical() &&
"Expected physical register in isFpOrNEON");
5077 return AArch64::FPR128RegClass.contains(Reg) ||
5078 AArch64::FPR64RegClass.contains(Reg) ||
5079 AArch64::FPR32RegClass.contains(Reg) ||
5080 AArch64::FPR16RegClass.contains(Reg) ||
5081 AArch64::FPR8RegClass.contains(Reg);
5088 auto Reg =
Op.getReg();
5089 if (Reg.isPhysical())
5093 return TRC == &AArch64::FPR128RegClass ||
5094 TRC == &AArch64::FPR128_loRegClass ||
5095 TRC == &AArch64::FPR64RegClass ||
5096 TRC == &AArch64::FPR64_loRegClass ||
5097 TRC == &AArch64::FPR32RegClass || TRC == &AArch64::FPR16RegClass ||
5098 TRC == &AArch64::FPR8RegClass;
5120 if (FirstOpc == SecondOpc)
5126 case AArch64::STRSui:
5127 case AArch64::STURSi:
5128 return SecondOpc == AArch64::STRSui || SecondOpc == AArch64::STURSi;
5129 case AArch64::STRDui:
5130 case AArch64::STURDi:
5131 return SecondOpc == AArch64::STRDui || SecondOpc == AArch64::STURDi;
5132 case AArch64::STRQui:
5133 case AArch64::STURQi:
5134 return SecondOpc == AArch64::STRQui || SecondOpc == AArch64::STURQi;
5135 case AArch64::STRWui:
5136 case AArch64::STURWi:
5137 return SecondOpc == AArch64::STRWui || SecondOpc == AArch64::STURWi;
5138 case AArch64::STRXui:
5139 case AArch64::STURXi:
5140 return SecondOpc == AArch64::STRXui || SecondOpc == AArch64::STURXi;
5141 case AArch64::LDRSui:
5142 case AArch64::LDURSi:
5143 return SecondOpc == AArch64::LDRSui || SecondOpc == AArch64::LDURSi;
5144 case AArch64::LDRDui:
5145 case AArch64::LDURDi:
5146 return SecondOpc == AArch64::LDRDui || SecondOpc == AArch64::LDURDi;
5147 case AArch64::LDRQui:
5148 case AArch64::LDURQi:
5149 return SecondOpc == AArch64::LDRQui || SecondOpc == AArch64::LDURQi;
5150 case AArch64::LDRWui:
5151 case AArch64::LDURWi:
5152 return SecondOpc == AArch64::LDRSWui || SecondOpc == AArch64::LDURSWi;
5153 case AArch64::LDRSWui:
5154 case AArch64::LDURSWi:
5155 return SecondOpc == AArch64::LDRWui || SecondOpc == AArch64::LDURWi;
5156 case AArch64::LDRXui:
5157 case AArch64::LDURXi:
5158 return SecondOpc == AArch64::LDRXui || SecondOpc == AArch64::LDURXi;
5165 int64_t Offset1,
unsigned Opcode1,
int FI2,
5166 int64_t Offset2,
unsigned Opcode2) {
5172 assert(ObjectOffset1 <= ObjectOffset2 &&
"Object offsets are not ordered.");
5175 if (ObjectOffset1 % Scale1 != 0)
5177 ObjectOffset1 /= Scale1;
5179 if (ObjectOffset2 % Scale2 != 0)
5181 ObjectOffset2 /= Scale2;
5182 ObjectOffset1 += Offset1;
5183 ObjectOffset2 += Offset2;
5184 return ObjectOffset1 + 1 == ObjectOffset2;
5196 int64_t OpOffset2,
bool OffsetIsScalable2,
unsigned ClusterSize,
5197 unsigned NumBytes)
const {
5207 "Only base registers and frame indices are supported.");
5214 if (ClusterSize > 2)
5221 unsigned FirstOpc = FirstLdSt.
getOpcode();
5222 unsigned SecondOpc = SecondLdSt.
getOpcode();
5242 if (Offset1 > 63 || Offset1 < -64)
5247 if (BaseOp1.
isFI()) {
5249 "Caller should have ordered offsets.");
5254 BaseOp2.
getIndex(), Offset2, SecondOpc);
5257 assert(Offset1 <= Offset2 &&
"Caller should have ordered offsets.");
5259 return Offset1 + 1 == Offset2;
5269 if (
Reg.isPhysical())
5278 return ((DestReg - SrcReg) & 0x1f) < NumRegs;
5287 assert(Subtarget.hasNEON() &&
"Unexpected register copy without NEON");
5289 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5290 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5291 unsigned NumRegs = Indices.
size();
5293 int SubReg = 0, End = NumRegs, Incr = 1;
5312 unsigned Opcode,
unsigned ZeroReg,
5315 unsigned NumRegs = Indices.
size();
5318 uint16_t DestEncoding =
TRI->getEncodingValue(DestReg);
5319 uint16_t SrcEncoding =
TRI->getEncodingValue(SrcReg);
5320 assert(DestEncoding % NumRegs == 0 && SrcEncoding % NumRegs == 0 &&
5321 "GPR reg sequences should not be able to overlap");
5338 bool RenamableSrc)
const {
5339 if (AArch64::GPR32spRegClass.
contains(DestReg) &&
5340 AArch64::GPR32spRegClass.
contains(SrcReg)) {
5341 if (DestReg == AArch64::WSP || SrcReg == AArch64::WSP) {
5343 if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5344 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5346 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5347 &AArch64::GPR64spRegClass);
5348 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5349 &AArch64::GPR64spRegClass);
5365 }
else if (Subtarget.hasZeroCycleRegMoveGPR64() &&
5366 !Subtarget.hasZeroCycleRegMoveGPR32()) {
5368 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5369 &AArch64::GPR64spRegClass);
5370 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5371 MCRegister SrcRegX = RI.getMatchingSuperReg(SrcReg, AArch64::sub_32,
5372 &AArch64::GPR64spRegClass);
5392 if (AArch64::GPR32spRegClass.
contains(DestReg) && SrcReg == AArch64::WZR) {
5393 if (Subtarget.hasZeroCycleZeroingGPR64() &&
5394 !Subtarget.hasZeroCycleZeroingGPR32()) {
5395 MCRegister DestRegX = RI.getMatchingSuperReg(DestReg, AArch64::sub_32,
5396 &AArch64::GPR64spRegClass);
5397 assert(DestRegX.
isValid() &&
"Destination super-reg not valid");
5401 }
else if (Subtarget.hasZeroCycleZeroingGPR32()) {
5413 if (AArch64::GPR64spRegClass.
contains(DestReg) &&
5414 AArch64::GPR64spRegClass.
contains(SrcReg)) {
5415 if (DestReg == AArch64::SP || SrcReg == AArch64::SP) {
5431 if (AArch64::GPR64spRegClass.
contains(DestReg) && SrcReg == AArch64::XZR) {
5432 if (Subtarget.hasZeroCycleZeroingGPR64()) {
5445 if (AArch64::PPRRegClass.
contains(DestReg) &&
5446 AArch64::PPRRegClass.
contains(SrcReg)) {
5447 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5448 "Unexpected SVE register.");
5458 bool DestIsPNR = AArch64::PNRRegClass.contains(DestReg);
5459 bool SrcIsPNR = AArch64::PNRRegClass.contains(SrcReg);
5460 if (DestIsPNR || SrcIsPNR) {
5462 return (R - AArch64::PN0) + AArch64::P0;
5467 if (PPRSrcReg != PPRDestReg) {
5479 if (AArch64::ZPRRegClass.
contains(DestReg) &&
5480 AArch64::ZPRRegClass.
contains(SrcReg)) {
5481 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5482 "Unexpected SVE register.");
5490 if ((AArch64::ZPR2RegClass.
contains(DestReg) ||
5491 AArch64::ZPR2StridedOrContiguousRegClass.
contains(DestReg)) &&
5492 (AArch64::ZPR2RegClass.
contains(SrcReg) ||
5493 AArch64::ZPR2StridedOrContiguousRegClass.
contains(SrcReg))) {
5494 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5495 "Unexpected SVE register.");
5496 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1};
5503 if (AArch64::ZPR3RegClass.
contains(DestReg) &&
5504 AArch64::ZPR3RegClass.
contains(SrcReg)) {
5505 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5506 "Unexpected SVE register.");
5507 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5515 if ((AArch64::ZPR4RegClass.
contains(DestReg) ||
5516 AArch64::ZPR4StridedOrContiguousRegClass.
contains(DestReg)) &&
5517 (AArch64::ZPR4RegClass.
contains(SrcReg) ||
5518 AArch64::ZPR4StridedOrContiguousRegClass.
contains(SrcReg))) {
5519 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5520 "Unexpected SVE register.");
5521 static const unsigned Indices[] = {AArch64::zsub0, AArch64::zsub1,
5522 AArch64::zsub2, AArch64::zsub3};
5529 if (AArch64::DDDDRegClass.
contains(DestReg) &&
5530 AArch64::DDDDRegClass.
contains(SrcReg)) {
5531 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5532 AArch64::dsub2, AArch64::dsub3};
5539 if (AArch64::DDDRegClass.
contains(DestReg) &&
5540 AArch64::DDDRegClass.
contains(SrcReg)) {
5541 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1,
5549 if (AArch64::DDRegClass.
contains(DestReg) &&
5550 AArch64::DDRegClass.
contains(SrcReg)) {
5551 static const unsigned Indices[] = {AArch64::dsub0, AArch64::dsub1};
5558 if (AArch64::QQQQRegClass.
contains(DestReg) &&
5559 AArch64::QQQQRegClass.
contains(SrcReg)) {
5560 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5561 AArch64::qsub2, AArch64::qsub3};
5568 if (AArch64::QQQRegClass.
contains(DestReg) &&
5569 AArch64::QQQRegClass.
contains(SrcReg)) {
5570 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1,
5578 if (AArch64::QQRegClass.
contains(DestReg) &&
5579 AArch64::QQRegClass.
contains(SrcReg)) {
5580 static const unsigned Indices[] = {AArch64::qsub0, AArch64::qsub1};
5586 if (AArch64::XSeqPairsClassRegClass.
contains(DestReg) &&
5587 AArch64::XSeqPairsClassRegClass.
contains(SrcReg)) {
5588 static const unsigned Indices[] = {AArch64::sube64, AArch64::subo64};
5590 AArch64::XZR, Indices);
5594 if (AArch64::WSeqPairsClassRegClass.
contains(DestReg) &&
5595 AArch64::WSeqPairsClassRegClass.
contains(SrcReg)) {
5596 static const unsigned Indices[] = {AArch64::sube32, AArch64::subo32};
5598 AArch64::WZR, Indices);
5602 if (AArch64::FPR128RegClass.
contains(DestReg) &&
5603 AArch64::FPR128RegClass.
contains(SrcReg)) {
5604 if (Subtarget.isSVEorStreamingSVEAvailable() &&
5605 !Subtarget.isNeonAvailable())
5608 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0))
5609 .
addReg(AArch64::Z0 + (SrcReg - AArch64::Q0));
5610 else if (Subtarget.isNeonAvailable())
5629 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5630 AArch64::FPR64RegClass.
contains(SrcReg)) {
5631 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5632 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5633 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5634 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::dsub,
5635 &AArch64::FPR128RegClass);
5636 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::dsub,
5637 &AArch64::FPR128RegClass);
5653 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5654 AArch64::FPR32RegClass.
contains(SrcReg)) {
5655 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5656 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5657 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5658 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5659 &AArch64::FPR128RegClass);
5660 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5661 &AArch64::FPR128RegClass);
5670 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5671 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5672 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::ssub,
5673 &AArch64::FPR64RegClass);
5674 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::ssub,
5675 &AArch64::FPR64RegClass);
5690 if (AArch64::FPR16RegClass.
contains(DestReg) &&
5691 AArch64::FPR16RegClass.
contains(SrcReg)) {
5692 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5693 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5694 !Subtarget.hasZeroCycleRegMoveFPR32() && Subtarget.isNeonAvailable()) {
5695 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5696 &AArch64::FPR128RegClass);
5697 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5698 &AArch64::FPR128RegClass);
5707 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5708 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5709 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5710 &AArch64::FPR64RegClass);
5711 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5712 &AArch64::FPR64RegClass);
5721 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::hsub,
5722 &AArch64::FPR32RegClass);
5723 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::hsub,
5724 &AArch64::FPR32RegClass);
5731 if (AArch64::FPR8RegClass.
contains(DestReg) &&
5732 AArch64::FPR8RegClass.
contains(SrcReg)) {
5733 if (Subtarget.hasZeroCycleRegMoveFPR128() &&
5734 !Subtarget.hasZeroCycleRegMoveFPR64() &&
5735 !Subtarget.hasZeroCycleRegMoveFPR64() && Subtarget.isNeonAvailable()) {
5736 MCRegister DestRegQ = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5737 &AArch64::FPR128RegClass);
5738 MCRegister SrcRegQ = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5739 &AArch64::FPR128RegClass);
5748 }
else if (Subtarget.hasZeroCycleRegMoveFPR64() &&
5749 !Subtarget.hasZeroCycleRegMoveFPR32()) {
5750 MCRegister DestRegD = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5751 &AArch64::FPR64RegClass);
5752 MCRegister SrcRegD = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5753 &AArch64::FPR64RegClass);
5762 DestReg = RI.getMatchingSuperReg(DestReg, AArch64::bsub,
5763 &AArch64::FPR32RegClass);
5764 SrcReg = RI.getMatchingSuperReg(SrcReg, AArch64::bsub,
5765 &AArch64::FPR32RegClass);
5773 if (AArch64::FPR64RegClass.
contains(DestReg) &&
5774 AArch64::GPR64RegClass.
contains(SrcReg)) {
5775 if (AArch64::XZR == SrcReg) {
5783 if (AArch64::GPR64RegClass.
contains(DestReg) &&
5784 AArch64::FPR64RegClass.
contains(SrcReg)) {
5790 if (AArch64::FPR32RegClass.
contains(DestReg) &&
5791 AArch64::GPR32RegClass.
contains(SrcReg)) {
5792 if (AArch64::WZR == SrcReg) {
5800 if (AArch64::GPR32RegClass.
contains(DestReg) &&
5801 AArch64::FPR32RegClass.
contains(SrcReg)) {
5807 if (DestReg == AArch64::NZCV) {
5808 assert(AArch64::GPR64RegClass.
contains(SrcReg) &&
"Invalid NZCV copy");
5810 .
addImm(AArch64SysReg::NZCV)
5816 if (SrcReg == AArch64::NZCV) {
5817 assert(AArch64::GPR64RegClass.
contains(DestReg) &&
"Invalid NZCV copy");
5819 .
addImm(AArch64SysReg::NZCV)
5825 errs() << RI.getRegAsmName(DestReg) <<
" = COPY " << RI.getRegAsmName(SrcReg)
5836 unsigned SubIdx0,
unsigned SubIdx1,
int FI,
5841 SrcReg0 =
TRI.getSubReg(SrcReg, SubIdx0);
5843 SrcReg1 =
TRI.getSubReg(SrcReg, SubIdx1);
5856 Register SrcReg,
bool isKill,
int FI,
5871 switch (RI.getSpillSize(*RC)) {
5873 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
5874 Opc = AArch64::STRBui;
5877 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
5878 Opc = AArch64::STRHui;
5879 else if (AArch64::PNRRegClass.hasSubClassEq(RC) ||
5880 AArch64::PPRRegClass.hasSubClassEq(RC)) {
5881 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5882 "Unexpected register store without SVE store instructions");
5883 Opc = AArch64::STR_PXI;
5889 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
5890 Opc = AArch64::STRWui;
5894 assert(SrcReg != AArch64::WSP);
5895 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
5896 Opc = AArch64::STRSui;
5897 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
5898 Opc = AArch64::STR_PPXI;
5903 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
5904 Opc = AArch64::STRXui;
5908 assert(SrcReg != AArch64::SP);
5909 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
5910 Opc = AArch64::STRDui;
5911 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
5913 get(AArch64::STPWi), SrcReg, isKill,
5914 AArch64::sube32, AArch64::subo32, FI, MMO);
5919 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
5920 Opc = AArch64::STRQui;
5921 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
5922 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5923 Opc = AArch64::ST1Twov1d;
5925 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
5927 get(AArch64::STPXi), SrcReg, isKill,
5928 AArch64::sube64, AArch64::subo64, FI, MMO);
5930 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
5931 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5932 "Unexpected register store without SVE store instructions");
5933 Opc = AArch64::STR_ZXI;
5938 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
5939 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5940 Opc = AArch64::ST1Threev1d;
5945 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
5946 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5947 Opc = AArch64::ST1Fourv1d;
5949 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
5950 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5951 Opc = AArch64::ST1Twov2d;
5953 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
5954 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5955 "Unexpected register store without SVE store instructions");
5956 Opc = AArch64::STR_ZZXI_STRIDED_CONTIGUOUS;
5958 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
5959 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5960 "Unexpected register store without SVE store instructions");
5961 Opc = AArch64::STR_ZZXI;
5966 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
5967 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5968 Opc = AArch64::ST1Threev2d;
5970 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
5971 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5972 "Unexpected register store without SVE store instructions");
5973 Opc = AArch64::STR_ZZZXI;
5978 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
5979 assert(Subtarget.hasNEON() &&
"Unexpected register store without NEON");
5980 Opc = AArch64::ST1Fourv2d;
5982 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
5983 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5984 "Unexpected register store without SVE store instructions");
5985 Opc = AArch64::STR_ZZZZXI_STRIDED_CONTIGUOUS;
5987 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
5988 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
5989 "Unexpected register store without SVE store instructions");
5990 Opc = AArch64::STR_ZZZZXI;
5995 assert(
Opc &&
"Unknown register class");
6006 MI.addMemOperand(MMO);
6013 Register DestReg,
unsigned SubIdx0,
6014 unsigned SubIdx1,
int FI,
6018 bool IsUndef =
true;
6020 DestReg0 =
TRI.getSubReg(DestReg, SubIdx0);
6022 DestReg1 =
TRI.getSubReg(DestReg, SubIdx1);
6051 switch (
TRI.getSpillSize(*RC)) {
6053 if (AArch64::FPR8RegClass.hasSubClassEq(RC))
6054 Opc = AArch64::LDRBui;
6057 bool IsPNR = AArch64::PNRRegClass.hasSubClassEq(RC);
6058 if (AArch64::FPR16RegClass.hasSubClassEq(RC))
6059 Opc = AArch64::LDRHui;
6060 else if (IsPNR || AArch64::PPRRegClass.hasSubClassEq(RC)) {
6061 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6062 "Unexpected register load without SVE load instructions");
6065 Opc = AArch64::LDR_PXI;
6071 if (AArch64::GPR32allRegClass.hasSubClassEq(RC)) {
6072 Opc = AArch64::LDRWui;
6076 assert(DestReg != AArch64::WSP);
6077 }
else if (AArch64::FPR32RegClass.hasSubClassEq(RC))
6078 Opc = AArch64::LDRSui;
6079 else if (AArch64::PPR2RegClass.hasSubClassEq(RC)) {
6080 Opc = AArch64::LDR_PPXI;
6085 if (AArch64::GPR64allRegClass.hasSubClassEq(RC)) {
6086 Opc = AArch64::LDRXui;
6090 assert(DestReg != AArch64::SP);
6091 }
else if (AArch64::FPR64RegClass.hasSubClassEq(RC)) {
6092 Opc = AArch64::LDRDui;
6093 }
else if (AArch64::WSeqPairsClassRegClass.hasSubClassEq(RC)) {
6095 get(AArch64::LDPWi), DestReg, AArch64::sube32,
6096 AArch64::subo32, FI, MMO);
6101 if (AArch64::FPR128RegClass.hasSubClassEq(RC))
6102 Opc = AArch64::LDRQui;
6103 else if (AArch64::DDRegClass.hasSubClassEq(RC)) {
6104 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6105 Opc = AArch64::LD1Twov1d;
6107 }
else if (AArch64::XSeqPairsClassRegClass.hasSubClassEq(RC)) {
6109 get(AArch64::LDPXi), DestReg, AArch64::sube64,
6110 AArch64::subo64, FI, MMO);
6112 }
else if (AArch64::ZPRRegClass.hasSubClassEq(RC)) {
6113 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6114 "Unexpected register load without SVE load instructions");
6115 Opc = AArch64::LDR_ZXI;
6120 if (AArch64::DDDRegClass.hasSubClassEq(RC)) {
6121 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6122 Opc = AArch64::LD1Threev1d;
6127 if (AArch64::DDDDRegClass.hasSubClassEq(RC)) {
6128 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6129 Opc = AArch64::LD1Fourv1d;
6131 }
else if (AArch64::QQRegClass.hasSubClassEq(RC)) {
6132 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6133 Opc = AArch64::LD1Twov2d;
6135 }
else if (AArch64::ZPR2StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6136 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6137 "Unexpected register load without SVE load instructions");
6138 Opc = AArch64::LDR_ZZXI_STRIDED_CONTIGUOUS;
6140 }
else if (AArch64::ZPR2RegClass.hasSubClassEq(RC)) {
6141 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6142 "Unexpected register load without SVE load instructions");
6143 Opc = AArch64::LDR_ZZXI;
6148 if (AArch64::QQQRegClass.hasSubClassEq(RC)) {
6149 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6150 Opc = AArch64::LD1Threev2d;
6152 }
else if (AArch64::ZPR3RegClass.hasSubClassEq(RC)) {
6153 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6154 "Unexpected register load without SVE load instructions");
6155 Opc = AArch64::LDR_ZZZXI;
6160 if (AArch64::QQQQRegClass.hasSubClassEq(RC)) {
6161 assert(Subtarget.hasNEON() &&
"Unexpected register load without NEON");
6162 Opc = AArch64::LD1Fourv2d;
6164 }
else if (AArch64::ZPR4StridedOrContiguousRegClass.hasSubClassEq(RC)) {
6165 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6166 "Unexpected register load without SVE load instructions");
6167 Opc = AArch64::LDR_ZZZZXI_STRIDED_CONTIGUOUS;
6169 }
else if (AArch64::ZPR4RegClass.hasSubClassEq(RC)) {
6170 assert(Subtarget.isSVEorStreamingSVEAvailable() &&
6171 "Unexpected register load without SVE load instructions");
6172 Opc = AArch64::LDR_ZZZZXI;
6178 assert(
Opc &&
"Unknown register class");
6188 MI.addMemOperand(MMO);
6195 UseMI.getIterator()),
6197 return I.modifiesRegister(AArch64::NZCV, TRI) ||
6198 I.readsRegister(AArch64::NZCV, TRI);
6202void AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6207 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6214 ByteSized =
Offset.getFixed();
6215 VGSized =
Offset.getScalable() / 2;
6221void AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6223 int64_t &NumDataVectors) {
6227 assert(
Offset.getScalable() % 2 == 0 &&
"Invalid frame offset");
6229 NumBytes =
Offset.getFixed();
6231 NumPredicateVectors =
Offset.getScalable() / 2;
6236 if (NumPredicateVectors % 8 == 0 || NumPredicateVectors < -64 ||
6237 NumPredicateVectors > 62) {
6238 NumDataVectors = NumPredicateVectors / 8;
6239 NumPredicateVectors -= NumDataVectors * 8;
6265 Expr.
push_back((
char)dwarf::DW_OP_bregx);
6273 int64_t OffsetFromDefCFA) {
6287 Comment << (NumBytes < 0 ?
" - " :
" + ") << std::abs(NumBytes);
6288 if (!RegScale.empty())
6298 int64_t NumBytes, NumVGScaledBytes;
6299 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
Offset, NumBytes,
6301 std::string CommentBuffer;
6304 if (
Reg == AArch64::SP)
6306 else if (
Reg == AArch64::FP)
6313 unsigned DwarfReg =
TRI.getDwarfRegNum(
Reg,
true);
6314 assert(DwarfReg <= 31 &&
"DwarfReg out of bounds (0..31)");
6316 Expr.
push_back(dwarf::DW_OP_breg0 + DwarfReg);
6319 if (NumVGScaledBytes) {
6329 DefCfaExpr.
push_back(dwarf::DW_CFA_def_cfa_expression);
6337 unsigned FrameReg,
unsigned Reg,
6339 bool LastAdjustmentWasScalable) {
6340 if (
Offset.getScalable())
6343 if (FrameReg == Reg && !LastAdjustmentWasScalable)
6346 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6353 std::optional<int64_t> IncomingVGOffsetFromDefCFA) {
6354 int64_t NumBytes, NumVGScaledBytes;
6355 AArch64InstrInfo::decomposeStackOffsetForDwarfOffsets(
6356 OffsetFromDefCFA, NumBytes, NumVGScaledBytes);
6358 unsigned DwarfReg =
TRI.getDwarfRegNum(Reg,
true);
6361 if (!NumVGScaledBytes)
6364 std::string CommentBuffer;
6369 assert(NumVGScaledBytes &&
"Expected scalable offset");
6373 if (IncomingVGOffsetFromDefCFA) {
6375 VGRegScale =
"* IncomingVG";
6378 VGRegScale =
"* VG";
6382 OffsetExpr.
push_back(dwarf::DW_OP_plus);
6391 CfaExpr.
push_back(dwarf::DW_CFA_expression);
6406 unsigned SrcReg, int64_t
Offset,
unsigned Opc,
6409 bool *HasWinCFI,
bool EmitCFAOffset,
6412 unsigned MaxEncoding, ShiftSize;
6414 case AArch64::ADDXri:
6415 case AArch64::ADDSXri:
6416 case AArch64::SUBXri:
6417 case AArch64::SUBSXri:
6418 MaxEncoding = 0xfff;
6421 case AArch64::ADDVL_XXI:
6422 case AArch64::ADDPL_XXI:
6423 case AArch64::ADDSVL_XXI:
6424 case AArch64::ADDSPL_XXI:
6439 if (
Opc == AArch64::ADDVL_XXI ||
Opc == AArch64::ADDSVL_XXI)
6441 else if (
Opc == AArch64::ADDPL_XXI ||
Opc == AArch64::ADDSPL_XXI)
6455 const unsigned MaxEncodableValue = MaxEncoding << ShiftSize;
6457 if (TmpReg == AArch64::XZR)
6458 TmpReg =
MBB.getParent()->getRegInfo().createVirtualRegister(
6459 &AArch64::GPR64RegClass);
6461 uint64_t ThisVal = std::min<uint64_t>(
Offset, MaxEncodableValue);
6462 unsigned LocalShiftSize = 0;
6463 if (ThisVal > MaxEncoding) {
6464 ThisVal = ThisVal >> ShiftSize;
6465 LocalShiftSize = ShiftSize;
6467 assert((ThisVal >> ShiftSize) <= MaxEncoding &&
6468 "Encoding cannot handle value that big");
6470 Offset -= ThisVal << LocalShiftSize;
6475 .
addImm(Sign * (
int)ThisVal);
6485 if (Sign == -1 ||
Opc == AArch64::SUBXri ||
Opc == AArch64::SUBSXri)
6486 CFAOffset += Change;
6488 CFAOffset -= Change;
6489 if (EmitCFAOffset && DestReg == TmpReg) {
6502 int Imm = (int)(ThisVal << LocalShiftSize);
6503 if (VScale != 1 && DestReg == AArch64::SP) {
6509 }
else if ((DestReg == AArch64::FP && SrcReg == AArch64::SP) ||
6510 (SrcReg == AArch64::FP && DestReg == AArch64::SP)) {
6511 assert(VScale == 1 &&
"Expected non-scalable operation");
6520 assert(
Offset == 0 &&
"Expected remaining offset to be zero to "
6521 "emit a single SEH directive");
6522 }
else if (DestReg == AArch64::SP) {
6523 assert(VScale == 1 &&
"Expected non-scalable operation");
6526 assert(SrcReg == AArch64::SP &&
"Unexpected SrcReg for SEH_StackAlloc");
6539 unsigned DestReg,
unsigned SrcReg,
6542 bool NeedsWinCFI,
bool *HasWinCFI,
6544 unsigned FrameReg) {
6551 bool UseSVL =
F.hasFnAttribute(
"aarch64_pstate_sm_body");
6553 int64_t Bytes, NumPredicateVectors, NumDataVectors;
6554 AArch64InstrInfo::decomposeStackOffsetForFrameOffsets(
6555 Offset, Bytes, NumPredicateVectors, NumDataVectors);
6558 bool NeedsFinalDefNZCV = SetNZCV && (NumPredicateVectors || NumDataVectors);
6559 if (NeedsFinalDefNZCV)
6563 if (Bytes || (!
Offset && SrcReg != DestReg)) {
6564 assert((DestReg != AArch64::SP || Bytes % 8 == 0) &&
6565 "SP increment/decrement not 8-byte aligned");
6566 unsigned Opc = SetNZCV ? AArch64::ADDSXri : AArch64::ADDXri;
6569 Opc = SetNZCV ? AArch64::SUBSXri : AArch64::SUBXri;
6572 NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6574 CFAOffset += (
Opc == AArch64::ADDXri ||
Opc == AArch64::ADDSXri)
6581 assert(!(NeedsWinCFI && NumPredicateVectors) &&
6582 "WinCFI can't allocate fractions of an SVE data vector");
6584 if (NumDataVectors) {
6586 UseSVL ? AArch64::ADDSVL_XXI : AArch64::ADDVL_XXI,
TII,
6587 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6593 if (NumPredicateVectors) {
6594 assert(DestReg != AArch64::SP &&
"Unaligned access to SP");
6596 UseSVL ? AArch64::ADDSPL_XXI : AArch64::ADDPL_XXI,
TII,
6597 Flag, NeedsWinCFI, HasWinCFI, EmitCFAOffset, CFAOffset,
6601 if (NeedsFinalDefNZCV)
6622 if (
MI.isFullCopy()) {
6625 if (SrcReg == AArch64::SP && DstReg.
isVirtual()) {
6629 if (DstReg == AArch64::SP && SrcReg.
isVirtual()) {
6634 if (SrcReg == AArch64::NZCV || DstReg == AArch64::NZCV)
6662 if (
MI.isCopy() &&
Ops.size() == 1 &&
6664 (
Ops[0] == 0 ||
Ops[0] == 1)) {
6665 bool IsSpill =
Ops[0] == 0;
6666 bool IsFill = !IsSpill;
6678 :
TRI.getMinimalPhysRegClass(Reg);
6684 "Mismatched register size in non subreg COPY");
6691 return &*--InsertPt;
6703 if (IsSpill && DstMO.
isUndef() && SrcReg == AArch64::WZR &&
6706 "Unexpected subreg on physical register");
6708 FrameIndex, &AArch64::GPR64RegClass,
Register());
6709 return &*--InsertPt;
6726 case AArch64::sub_32:
6727 if (AArch64::GPR64RegClass.hasSubClassEq(
getRegClass(DstReg)))
6728 FillRC = &AArch64::GPR32RegClass;
6731 FillRC = &AArch64::FPR32RegClass;
6734 FillRC = &AArch64::FPR64RegClass;
6740 TRI.getRegSizeInBits(*FillRC) &&
6741 "Mismatched regclass size on folded subreg COPY");
6760 bool *OutUseUnscaledOp,
6761 unsigned *OutUnscaledOp,
6762 int64_t *EmittableOffset) {
6764 if (EmittableOffset)
6765 *EmittableOffset = 0;
6766 if (OutUseUnscaledOp)
6767 *OutUseUnscaledOp =
false;
6773 switch (
MI.getOpcode()) {
6776 case AArch64::LD1Rv1d:
6777 case AArch64::LD1Rv2s:
6778 case AArch64::LD1Rv2d:
6779 case AArch64::LD1Rv4h:
6780 case AArch64::LD1Rv4s:
6781 case AArch64::LD1Rv8b:
6782 case AArch64::LD1Rv8h:
6783 case AArch64::LD1Rv16b:
6784 case AArch64::LD1Twov2d:
6785 case AArch64::LD1Threev2d:
6786 case AArch64::LD1Fourv2d:
6787 case AArch64::LD1Twov1d:
6788 case AArch64::LD1Threev1d:
6789 case AArch64::LD1Fourv1d:
6790 case AArch64::ST1Twov2d:
6791 case AArch64::ST1Threev2d:
6792 case AArch64::ST1Fourv2d:
6793 case AArch64::ST1Twov1d:
6794 case AArch64::ST1Threev1d:
6795 case AArch64::ST1Fourv1d:
6796 case AArch64::ST1i8:
6797 case AArch64::ST1i16:
6798 case AArch64::ST1i32:
6799 case AArch64::ST1i64:
6801 case AArch64::IRGstack:
6802 case AArch64::STGloop:
6803 case AArch64::STZGloop:
6808 TypeSize ScaleValue(0U,
false), Width(0U,
false);
6809 int64_t MinOff, MaxOff;
6815 bool IsMulVL = ScaleValue.isScalable();
6816 unsigned Scale = ScaleValue.getKnownMinValue();
6826 std::optional<unsigned> UnscaledOp =
6828 bool useUnscaledOp = UnscaledOp && (
Offset % Scale ||
Offset < 0);
6829 if (useUnscaledOp &&
6834 Scale = ScaleValue.getKnownMinValue();
6835 assert(IsMulVL == ScaleValue.isScalable() &&
6836 "Unscaled opcode has different value for scalable");
6838 int64_t Remainder =
Offset % Scale;
6839 assert(!(Remainder && useUnscaledOp) &&
6840 "Cannot have remainder when using unscaled op");
6842 assert(MinOff < MaxOff &&
"Unexpected Min/Max offsets");
6843 int64_t NewOffset =
Offset / Scale;
6844 if (MinOff <= NewOffset && NewOffset <= MaxOff)
6847 NewOffset = NewOffset < 0 ? MinOff : MaxOff;
6851 if (EmittableOffset)
6852 *EmittableOffset = NewOffset;
6853 if (OutUseUnscaledOp)
6854 *OutUseUnscaledOp = useUnscaledOp;
6855 if (OutUnscaledOp && UnscaledOp)
6856 *OutUnscaledOp = *UnscaledOp;
6869 unsigned Opcode =
MI.getOpcode();
6870 unsigned ImmIdx = FrameRegIdx + 1;
6872 if (Opcode == AArch64::ADDSXri || Opcode == AArch64::ADDXri) {
6877 MI.eraseFromParent();
6883 unsigned UnscaledOp;
6886 &UnscaledOp, &NewOffset);
6890 MI.getOperand(FrameRegIdx).ChangeToRegister(FrameReg,
false);
6892 MI.setDesc(
TII->get(UnscaledOp));
6894 MI.getOperand(ImmIdx).ChangeToImmediate(NewOffset);
6912bool AArch64InstrInfo::useMachineCombiner()
const {
return true; }
6917 case AArch64::ADDSWrr:
6918 case AArch64::ADDSWri:
6919 case AArch64::ADDSXrr:
6920 case AArch64::ADDSXri:
6921 case AArch64::SUBSWrr:
6922 case AArch64::SUBSXrr:
6924 case AArch64::SUBSWri:
6925 case AArch64::SUBSXri:
6936 case AArch64::ADDWrr:
6937 case AArch64::ADDWri:
6938 case AArch64::SUBWrr:
6939 case AArch64::ADDSWrr:
6940 case AArch64::ADDSWri:
6941 case AArch64::SUBSWrr:
6943 case AArch64::SUBWri:
6944 case AArch64::SUBSWri:
6955 case AArch64::ADDXrr:
6956 case AArch64::ADDXri:
6957 case AArch64::SUBXrr:
6958 case AArch64::ADDSXrr:
6959 case AArch64::ADDSXri:
6960 case AArch64::SUBSXrr:
6962 case AArch64::SUBXri:
6963 case AArch64::SUBSXri:
6964 case AArch64::ADDv8i8:
6965 case AArch64::ADDv16i8:
6966 case AArch64::ADDv4i16:
6967 case AArch64::ADDv8i16:
6968 case AArch64::ADDv2i32:
6969 case AArch64::ADDv4i32:
6970 case AArch64::SUBv8i8:
6971 case AArch64::SUBv16i8:
6972 case AArch64::SUBv4i16:
6973 case AArch64::SUBv8i16:
6974 case AArch64::SUBv2i32:
6975 case AArch64::SUBv4i32:
6988 case AArch64::FADDHrr:
6989 case AArch64::FADDSrr:
6990 case AArch64::FADDDrr:
6991 case AArch64::FADDv4f16:
6992 case AArch64::FADDv8f16:
6993 case AArch64::FADDv2f32:
6994 case AArch64::FADDv2f64:
6995 case AArch64::FADDv4f32:
6996 case AArch64::FSUBHrr:
6997 case AArch64::FSUBSrr:
6998 case AArch64::FSUBDrr:
6999 case AArch64::FSUBv4f16:
7000 case AArch64::FSUBv8f16:
7001 case AArch64::FSUBv2f32:
7002 case AArch64::FSUBv2f64:
7003 case AArch64::FSUBv4f32:
7022 unsigned CombineOpc,
unsigned ZeroReg = 0,
7023 bool CheckZeroReg =
false) {
7030 if (!
MI ||
MI->getParent() != &
MBB ||
MI->getOpcode() != CombineOpc)
7033 if (!
MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()))
7037 assert(
MI->getNumOperands() >= 4 &&
MI->getOperand(0).isReg() &&
7038 MI->getOperand(1).isReg() &&
MI->getOperand(2).isReg() &&
7039 MI->getOperand(3).isReg() &&
"MAdd/MSub must have a least 4 regs");
7041 if (
MI->getOperand(3).getReg() != ZeroReg)
7046 MI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
true) == -1)
7055 unsigned MulOpc,
unsigned ZeroReg) {
7070bool AArch64InstrInfo::isAssociativeAndCommutative(
const MachineInstr &Inst,
7071 bool Invert)
const {
7077 case AArch64::FADDHrr:
7078 case AArch64::FADDSrr:
7079 case AArch64::FADDDrr:
7080 case AArch64::FMULHrr:
7081 case AArch64::FMULSrr:
7082 case AArch64::FMULDrr:
7083 case AArch64::FMULX16:
7084 case AArch64::FMULX32:
7085 case AArch64::FMULX64:
7087 case AArch64::FADDv4f16:
7088 case AArch64::FADDv8f16:
7089 case AArch64::FADDv2f32:
7090 case AArch64::FADDv4f32:
7091 case AArch64::FADDv2f64:
7092 case AArch64::FMULv4f16:
7093 case AArch64::FMULv8f16:
7094 case AArch64::FMULv2f32:
7095 case AArch64::FMULv4f32:
7096 case AArch64::FMULv2f64:
7097 case AArch64::FMULXv4f16:
7098 case AArch64::FMULXv8f16:
7099 case AArch64::FMULXv2f32:
7100 case AArch64::FMULXv4f32:
7101 case AArch64::FMULXv2f64:
7105 case AArch64::FADD_ZZZ_H:
7106 case AArch64::FADD_ZZZ_S:
7107 case AArch64::FADD_ZZZ_D:
7108 case AArch64::FMUL_ZZZ_H:
7109 case AArch64::FMUL_ZZZ_S:
7110 case AArch64::FMUL_ZZZ_D:
7121 case AArch64::ADDWrr:
7122 case AArch64::ADDXrr:
7123 case AArch64::ANDWrr:
7124 case AArch64::ANDXrr:
7125 case AArch64::ORRWrr:
7126 case AArch64::ORRXrr:
7127 case AArch64::EORWrr:
7128 case AArch64::EORXrr:
7129 case AArch64::EONWrr:
7130 case AArch64::EONXrr:
7134 case AArch64::ADDv8i8:
7135 case AArch64::ADDv16i8:
7136 case AArch64::ADDv4i16:
7137 case AArch64::ADDv8i16:
7138 case AArch64::ADDv2i32:
7139 case AArch64::ADDv4i32:
7140 case AArch64::ADDv1i64:
7141 case AArch64::ADDv2i64:
7142 case AArch64::MULv8i8:
7143 case AArch64::MULv16i8:
7144 case AArch64::MULv4i16:
7145 case AArch64::MULv8i16:
7146 case AArch64::MULv2i32:
7147 case AArch64::MULv4i32:
7148 case AArch64::ANDv8i8:
7149 case AArch64::ANDv16i8:
7150 case AArch64::ORRv8i8:
7151 case AArch64::ORRv16i8:
7152 case AArch64::EORv8i8:
7153 case AArch64::EORv16i8:
7155 case AArch64::ADD_ZZZ_B:
7156 case AArch64::ADD_ZZZ_H:
7157 case AArch64::ADD_ZZZ_S:
7158 case AArch64::ADD_ZZZ_D:
7159 case AArch64::MUL_ZZZ_B:
7160 case AArch64::MUL_ZZZ_H:
7161 case AArch64::MUL_ZZZ_S:
7162 case AArch64::MUL_ZZZ_D:
7163 case AArch64::AND_ZZZ:
7164 case AArch64::ORR_ZZZ:
7165 case AArch64::EOR_ZZZ:
7196 auto setFound = [&](
int Opcode,
int Operand,
unsigned ZeroReg,
7204 auto setVFound = [&](
int Opcode,
int Operand,
unsigned Pattern) {
7216 case AArch64::ADDWrr:
7218 "ADDWrr does not have register operands");
7219 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDW_OP1);
7220 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULADDW_OP2);
7222 case AArch64::ADDXrr:
7223 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDX_OP1);
7224 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULADDX_OP2);
7226 case AArch64::SUBWrr:
7227 setFound(AArch64::MADDWrrr, 2, AArch64::WZR, MCP::MULSUBW_OP2);
7228 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBW_OP1);
7230 case AArch64::SUBXrr:
7231 setFound(AArch64::MADDXrrr, 2, AArch64::XZR, MCP::MULSUBX_OP2);
7232 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBX_OP1);
7234 case AArch64::ADDWri:
7235 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULADDWI_OP1);
7237 case AArch64::ADDXri:
7238 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULADDXI_OP1);
7240 case AArch64::SUBWri:
7241 setFound(AArch64::MADDWrrr, 1, AArch64::WZR, MCP::MULSUBWI_OP1);
7243 case AArch64::SUBXri:
7244 setFound(AArch64::MADDXrrr, 1, AArch64::XZR, MCP::MULSUBXI_OP1);
7246 case AArch64::ADDv8i8:
7247 setVFound(AArch64::MULv8i8, 1, MCP::MULADDv8i8_OP1);
7248 setVFound(AArch64::MULv8i8, 2, MCP::MULADDv8i8_OP2);
7250 case AArch64::ADDv16i8:
7251 setVFound(AArch64::MULv16i8, 1, MCP::MULADDv16i8_OP1);
7252 setVFound(AArch64::MULv16i8, 2, MCP::MULADDv16i8_OP2);
7254 case AArch64::ADDv4i16:
7255 setVFound(AArch64::MULv4i16, 1, MCP::MULADDv4i16_OP1);
7256 setVFound(AArch64::MULv4i16, 2, MCP::MULADDv4i16_OP2);
7257 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULADDv4i16_indexed_OP1);
7258 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULADDv4i16_indexed_OP2);
7260 case AArch64::ADDv8i16:
7261 setVFound(AArch64::MULv8i16, 1, MCP::MULADDv8i16_OP1);
7262 setVFound(AArch64::MULv8i16, 2, MCP::MULADDv8i16_OP2);
7263 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULADDv8i16_indexed_OP1);
7264 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULADDv8i16_indexed_OP2);
7266 case AArch64::ADDv2i32:
7267 setVFound(AArch64::MULv2i32, 1, MCP::MULADDv2i32_OP1);
7268 setVFound(AArch64::MULv2i32, 2, MCP::MULADDv2i32_OP2);
7269 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULADDv2i32_indexed_OP1);
7270 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULADDv2i32_indexed_OP2);
7272 case AArch64::ADDv4i32:
7273 setVFound(AArch64::MULv4i32, 1, MCP::MULADDv4i32_OP1);
7274 setVFound(AArch64::MULv4i32, 2, MCP::MULADDv4i32_OP2);
7275 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULADDv4i32_indexed_OP1);
7276 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULADDv4i32_indexed_OP2);
7278 case AArch64::SUBv8i8:
7279 setVFound(AArch64::MULv8i8, 1, MCP::MULSUBv8i8_OP1);
7280 setVFound(AArch64::MULv8i8, 2, MCP::MULSUBv8i8_OP2);
7282 case AArch64::SUBv16i8:
7283 setVFound(AArch64::MULv16i8, 1, MCP::MULSUBv16i8_OP1);
7284 setVFound(AArch64::MULv16i8, 2, MCP::MULSUBv16i8_OP2);
7286 case AArch64::SUBv4i16:
7287 setVFound(AArch64::MULv4i16, 1, MCP::MULSUBv4i16_OP1);
7288 setVFound(AArch64::MULv4i16, 2, MCP::MULSUBv4i16_OP2);
7289 setVFound(AArch64::MULv4i16_indexed, 1, MCP::MULSUBv4i16_indexed_OP1);
7290 setVFound(AArch64::MULv4i16_indexed, 2, MCP::MULSUBv4i16_indexed_OP2);
7292 case AArch64::SUBv8i16:
7293 setVFound(AArch64::MULv8i16, 1, MCP::MULSUBv8i16_OP1);
7294 setVFound(AArch64::MULv8i16, 2, MCP::MULSUBv8i16_OP2);
7295 setVFound(AArch64::MULv8i16_indexed, 1, MCP::MULSUBv8i16_indexed_OP1);
7296 setVFound(AArch64::MULv8i16_indexed, 2, MCP::MULSUBv8i16_indexed_OP2);
7298 case AArch64::SUBv2i32:
7299 setVFound(AArch64::MULv2i32, 1, MCP::MULSUBv2i32_OP1);
7300 setVFound(AArch64::MULv2i32, 2, MCP::MULSUBv2i32_OP2);
7301 setVFound(AArch64::MULv2i32_indexed, 1, MCP::MULSUBv2i32_indexed_OP1);
7302 setVFound(AArch64::MULv2i32_indexed, 2, MCP::MULSUBv2i32_indexed_OP2);
7304 case AArch64::SUBv4i32:
7305 setVFound(AArch64::MULv4i32, 1, MCP::MULSUBv4i32_OP1);
7306 setVFound(AArch64::MULv4i32, 2, MCP::MULSUBv4i32_OP2);
7307 setVFound(AArch64::MULv4i32_indexed, 1, MCP::MULSUBv4i32_indexed_OP1);
7308 setVFound(AArch64::MULv4i32_indexed, 2, MCP::MULSUBv4i32_indexed_OP2);
7314bool AArch64InstrInfo::isAccumulationOpcode(
unsigned Opcode)
const {
7318 case AArch64::UABALB_ZZZ_D:
7319 case AArch64::UABALB_ZZZ_H:
7320 case AArch64::UABALB_ZZZ_S:
7321 case AArch64::UABALT_ZZZ_D:
7322 case AArch64::UABALT_ZZZ_H:
7323 case AArch64::UABALT_ZZZ_S:
7324 case AArch64::SABALB_ZZZ_D:
7325 case AArch64::SABALB_ZZZ_S:
7326 case AArch64::SABALB_ZZZ_H:
7327 case AArch64::SABALT_ZZZ_D:
7328 case AArch64::SABALT_ZZZ_S:
7329 case AArch64::SABALT_ZZZ_H:
7330 case AArch64::UABALv16i8_v8i16:
7331 case AArch64::UABALv2i32_v2i64:
7332 case AArch64::UABALv4i16_v4i32:
7333 case AArch64::UABALv4i32_v2i64:
7334 case AArch64::UABALv8i16_v4i32:
7335 case AArch64::UABALv8i8_v8i16:
7336 case AArch64::UABAv16i8:
7337 case AArch64::UABAv2i32:
7338 case AArch64::UABAv4i16:
7339 case AArch64::UABAv4i32:
7340 case AArch64::UABAv8i16:
7341 case AArch64::UABAv8i8:
7342 case AArch64::SABALv16i8_v8i16:
7343 case AArch64::SABALv2i32_v2i64:
7344 case AArch64::SABALv4i16_v4i32:
7345 case AArch64::SABALv4i32_v2i64:
7346 case AArch64::SABALv8i16_v4i32:
7347 case AArch64::SABALv8i8_v8i16:
7348 case AArch64::SABAv16i8:
7349 case AArch64::SABAv2i32:
7350 case AArch64::SABAv4i16:
7351 case AArch64::SABAv4i32:
7352 case AArch64::SABAv8i16:
7353 case AArch64::SABAv8i8:
7360unsigned AArch64InstrInfo::getAccumulationStartOpcode(
7361 unsigned AccumulationOpcode)
const {
7362 switch (AccumulationOpcode) {
7365 case AArch64::UABALB_ZZZ_D:
7366 return AArch64::UABDLB_ZZZ_D;
7367 case AArch64::UABALB_ZZZ_H:
7368 return AArch64::UABDLB_ZZZ_H;
7369 case AArch64::UABALB_ZZZ_S:
7370 return AArch64::UABDLB_ZZZ_S;
7371 case AArch64::UABALT_ZZZ_D:
7372 return AArch64::UABDLT_ZZZ_D;
7373 case AArch64::UABALT_ZZZ_H:
7374 return AArch64::UABDLT_ZZZ_H;
7375 case AArch64::UABALT_ZZZ_S:
7376 return AArch64::UABDLT_ZZZ_S;
7377 case AArch64::UABALv16i8_v8i16:
7378 return AArch64::UABDLv16i8_v8i16;
7379 case AArch64::UABALv2i32_v2i64:
7380 return AArch64::UABDLv2i32_v2i64;
7381 case AArch64::UABALv4i16_v4i32:
7382 return AArch64::UABDLv4i16_v4i32;
7383 case AArch64::UABALv4i32_v2i64:
7384 return AArch64::UABDLv4i32_v2i64;
7385 case AArch64::UABALv8i16_v4i32:
7386 return AArch64::UABDLv8i16_v4i32;
7387 case AArch64::UABALv8i8_v8i16:
7388 return AArch64::UABDLv8i8_v8i16;
7389 case AArch64::UABAv16i8:
7390 return AArch64::UABDv16i8;
7391 case AArch64::UABAv2i32:
7392 return AArch64::UABDv2i32;
7393 case AArch64::UABAv4i16:
7394 return AArch64::UABDv4i16;
7395 case AArch64::UABAv4i32:
7396 return AArch64::UABDv4i32;
7397 case AArch64::UABAv8i16:
7398 return AArch64::UABDv8i16;
7399 case AArch64::UABAv8i8:
7400 return AArch64::UABDv8i8;
7401 case AArch64::SABALB_ZZZ_D:
7402 return AArch64::SABDLB_ZZZ_D;
7403 case AArch64::SABALB_ZZZ_S:
7404 return AArch64::SABDLB_ZZZ_S;
7405 case AArch64::SABALB_ZZZ_H:
7406 return AArch64::SABDLB_ZZZ_H;
7407 case AArch64::SABALT_ZZZ_D:
7408 return AArch64::SABDLT_ZZZ_D;
7409 case AArch64::SABALT_ZZZ_S:
7410 return AArch64::SABDLT_ZZZ_S;
7411 case AArch64::SABALT_ZZZ_H:
7412 return AArch64::SABDLT_ZZZ_H;
7413 case AArch64::SABALv16i8_v8i16:
7414 return AArch64::SABDLv16i8_v8i16;
7415 case AArch64::SABALv2i32_v2i64:
7416 return AArch64::SABDLv2i32_v2i64;
7417 case AArch64::SABALv4i16_v4i32:
7418 return AArch64::SABDLv4i16_v4i32;
7419 case AArch64::SABALv4i32_v2i64:
7420 return AArch64::SABDLv4i32_v2i64;
7421 case AArch64::SABALv8i16_v4i32:
7422 return AArch64::SABDLv8i16_v4i32;
7423 case AArch64::SABALv8i8_v8i16:
7424 return AArch64::SABDLv8i8_v8i16;
7425 case AArch64::SABAv16i8:
7426 return AArch64::SABDv16i8;
7427 case AArch64::SABAv2i32:
7428 return AArch64::SABAv2i32;
7429 case AArch64::SABAv4i16:
7430 return AArch64::SABDv4i16;
7431 case AArch64::SABAv4i32:
7432 return AArch64::SABDv4i32;
7433 case AArch64::SABAv8i16:
7434 return AArch64::SABDv8i16;
7435 case AArch64::SABAv8i8:
7436 return AArch64::SABDv8i8;
7452 auto Match = [&](
int Opcode,
int Operand,
unsigned Pattern) ->
bool {
7464 assert(
false &&
"Unsupported FP instruction in combiner\n");
7466 case AArch64::FADDHrr:
7468 "FADDHrr does not have register operands");
7470 Found = Match(AArch64::FMULHrr, 1, MCP::FMULADDH_OP1);
7471 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULADDH_OP2);
7473 case AArch64::FADDSrr:
7475 "FADDSrr does not have register operands");
7477 Found |= Match(AArch64::FMULSrr, 1, MCP::FMULADDS_OP1) ||
7478 Match(AArch64::FMULv1i32_indexed, 1, MCP::FMLAv1i32_indexed_OP1);
7480 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULADDS_OP2) ||
7481 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLAv1i32_indexed_OP2);
7483 case AArch64::FADDDrr:
7484 Found |= Match(AArch64::FMULDrr, 1, MCP::FMULADDD_OP1) ||
7485 Match(AArch64::FMULv1i64_indexed, 1, MCP::FMLAv1i64_indexed_OP1);
7487 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULADDD_OP2) ||
7488 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLAv1i64_indexed_OP2);
7490 case AArch64::FADDv4f16:
7491 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLAv4i16_indexed_OP1) ||
7492 Match(AArch64::FMULv4f16, 1, MCP::FMLAv4f16_OP1);
7494 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLAv4i16_indexed_OP2) ||
7495 Match(AArch64::FMULv4f16, 2, MCP::FMLAv4f16_OP2);
7497 case AArch64::FADDv8f16:
7498 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLAv8i16_indexed_OP1) ||
7499 Match(AArch64::FMULv8f16, 1, MCP::FMLAv8f16_OP1);
7501 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLAv8i16_indexed_OP2) ||
7502 Match(AArch64::FMULv8f16, 2, MCP::FMLAv8f16_OP2);
7504 case AArch64::FADDv2f32:
7505 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLAv2i32_indexed_OP1) ||
7506 Match(AArch64::FMULv2f32, 1, MCP::FMLAv2f32_OP1);
7508 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLAv2i32_indexed_OP2) ||
7509 Match(AArch64::FMULv2f32, 2, MCP::FMLAv2f32_OP2);
7511 case AArch64::FADDv2f64:
7512 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLAv2i64_indexed_OP1) ||
7513 Match(AArch64::FMULv2f64, 1, MCP::FMLAv2f64_OP1);
7515 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLAv2i64_indexed_OP2) ||
7516 Match(AArch64::FMULv2f64, 2, MCP::FMLAv2f64_OP2);
7518 case AArch64::FADDv4f32:
7519 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLAv4i32_indexed_OP1) ||
7520 Match(AArch64::FMULv4f32, 1, MCP::FMLAv4f32_OP1);
7522 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLAv4i32_indexed_OP2) ||
7523 Match(AArch64::FMULv4f32, 2, MCP::FMLAv4f32_OP2);
7525 case AArch64::FSUBHrr:
7526 Found = Match(AArch64::FMULHrr, 1, MCP::FMULSUBH_OP1);
7527 Found |= Match(AArch64::FMULHrr, 2, MCP::FMULSUBH_OP2);
7528 Found |= Match(AArch64::FNMULHrr, 1, MCP::FNMULSUBH_OP1);
7530 case AArch64::FSUBSrr:
7531 Found = Match(AArch64::FMULSrr, 1, MCP::FMULSUBS_OP1);
7533 Found |= Match(AArch64::FMULSrr, 2, MCP::FMULSUBS_OP2) ||
7534 Match(AArch64::FMULv1i32_indexed, 2, MCP::FMLSv1i32_indexed_OP2);
7536 Found |= Match(AArch64::FNMULSrr, 1, MCP::FNMULSUBS_OP1);
7538 case AArch64::FSUBDrr:
7539 Found = Match(AArch64::FMULDrr, 1, MCP::FMULSUBD_OP1);
7541 Found |= Match(AArch64::FMULDrr, 2, MCP::FMULSUBD_OP2) ||
7542 Match(AArch64::FMULv1i64_indexed, 2, MCP::FMLSv1i64_indexed_OP2);
7544 Found |= Match(AArch64::FNMULDrr, 1, MCP::FNMULSUBD_OP1);
7546 case AArch64::FSUBv4f16:
7547 Found |= Match(AArch64::FMULv4i16_indexed, 2, MCP::FMLSv4i16_indexed_OP2) ||
7548 Match(AArch64::FMULv4f16, 2, MCP::FMLSv4f16_OP2);
7550 Found |= Match(AArch64::FMULv4i16_indexed, 1, MCP::FMLSv4i16_indexed_OP1) ||
7551 Match(AArch64::FMULv4f16, 1, MCP::FMLSv4f16_OP1);
7553 case AArch64::FSUBv8f16:
7554 Found |= Match(AArch64::FMULv8i16_indexed, 2, MCP::FMLSv8i16_indexed_OP2) ||
7555 Match(AArch64::FMULv8f16, 2, MCP::FMLSv8f16_OP2);
7557 Found |= Match(AArch64::FMULv8i16_indexed, 1, MCP::FMLSv8i16_indexed_OP1) ||
7558 Match(AArch64::FMULv8f16, 1, MCP::FMLSv8f16_OP1);
7560 case AArch64::FSUBv2f32:
7561 Found |= Match(AArch64::FMULv2i32_indexed, 2, MCP::FMLSv2i32_indexed_OP2) ||
7562 Match(AArch64::FMULv2f32, 2, MCP::FMLSv2f32_OP2);
7564 Found |= Match(AArch64::FMULv2i32_indexed, 1, MCP::FMLSv2i32_indexed_OP1) ||
7565 Match(AArch64::FMULv2f32, 1, MCP::FMLSv2f32_OP1);
7567 case AArch64::FSUBv2f64:
7568 Found |= Match(AArch64::FMULv2i64_indexed, 2, MCP::FMLSv2i64_indexed_OP2) ||
7569 Match(AArch64::FMULv2f64, 2, MCP::FMLSv2f64_OP2);
7571 Found |= Match(AArch64::FMULv2i64_indexed, 1, MCP::FMLSv2i64_indexed_OP1) ||
7572 Match(AArch64::FMULv2f64, 1, MCP::FMLSv2f64_OP1);
7574 case AArch64::FSUBv4f32:
7575 Found |= Match(AArch64::FMULv4i32_indexed, 2, MCP::FMLSv4i32_indexed_OP2) ||
7576 Match(AArch64::FMULv4f32, 2, MCP::FMLSv4f32_OP2);
7578 Found |= Match(AArch64::FMULv4i32_indexed, 1, MCP::FMLSv4i32_indexed_OP1) ||
7579 Match(AArch64::FMULv4f32, 1, MCP::FMLSv4f32_OP1);
7590 auto Match = [&](
unsigned Opcode,
int Operand,
unsigned Pattern) ->
bool {
7597 if (
MI &&
MI->getOpcode() == TargetOpcode::COPY &&
7598 MI->getOperand(1).getReg().isVirtual())
7599 MI =
MRI.getUniqueVRegDef(
MI->getOperand(1).getReg());
7600 if (
MI &&
MI->getOpcode() == Opcode) {
7612 case AArch64::FMULv2f32:
7613 Found = Match(AArch64::DUPv2i32lane, 1, MCP::FMULv2i32_indexed_OP1);
7614 Found |= Match(AArch64::DUPv2i32lane, 2, MCP::FMULv2i32_indexed_OP2);
7616 case AArch64::FMULv2f64:
7617 Found = Match(AArch64::DUPv2i64lane, 1, MCP::FMULv2i64_indexed_OP1);
7618 Found |= Match(AArch64::DUPv2i64lane, 2, MCP::FMULv2i64_indexed_OP2);
7620 case AArch64::FMULv4f16:
7621 Found = Match(AArch64::DUPv4i16lane, 1, MCP::FMULv4i16_indexed_OP1);
7622 Found |= Match(AArch64::DUPv4i16lane, 2, MCP::FMULv4i16_indexed_OP2);
7624 case AArch64::FMULv4f32:
7625 Found = Match(AArch64::DUPv4i32lane, 1, MCP::FMULv4i32_indexed_OP1);
7626 Found |= Match(AArch64::DUPv4i32lane, 2, MCP::FMULv4i32_indexed_OP2);
7628 case AArch64::FMULv8f16:
7629 Found = Match(AArch64::DUPv8i16lane, 1, MCP::FMULv8i16_indexed_OP1);
7630 Found |= Match(AArch64::DUPv8i16lane, 2, MCP::FMULv8i16_indexed_OP2);
7643 auto Match = [&](
unsigned Opcode,
unsigned Pattern) ->
bool {
7646 if (
MI !=
nullptr && (
MI->getOpcode() == Opcode) &&
7647 MRI.hasOneNonDBGUse(
MI->getOperand(0).getReg()) &&
7661 case AArch64::FNEGDr:
7663 case AArch64::FNEGSr:
7795 case AArch64::SUBWrr:
7796 case AArch64::SUBSWrr:
7797 case AArch64::SUBXrr:
7798 case AArch64::SUBSXrr:
7843 unsigned LoadLaneOpCode,
unsigned NumLanes) {
7866 while (!RemainingLanes.
empty() && CurrInstr &&
7867 CurrInstr->getOpcode() == LoadLaneOpCode &&
7868 MRI.hasOneNonDBGUse(CurrInstr->getOperand(0).getReg()) &&
7869 CurrInstr->getNumOperands() == 4) {
7870 RemainingLanes.
erase(CurrInstr->getOperand(2).getImm());
7872 CurrInstr =
MRI.getUniqueVRegDef(CurrInstr->getOperand(1).getReg());
7876 if (!RemainingLanes.
empty())
7880 if (CurrInstr->getOpcode() != TargetOpcode::SUBREG_TO_REG)
7884 auto Lane0LoadReg = CurrInstr->getOperand(2).getReg();
7885 unsigned SingleLaneSizeInBits = 128 / NumLanes;
7886 if (
TRI->getRegSizeInBits(Lane0LoadReg,
MRI) != SingleLaneSizeInBits)
7890 if (!
MRI.hasOneNonDBGUse(Lane0LoadReg))
7893 LoadInstrs.
push_back(
MRI.getUniqueVRegDef(Lane0LoadReg));
7902 RemainingLoadInstrs.
insert(LoadInstrs.
begin(), LoadInstrs.
end());
7905 for (; MBBItr !=
MBB->begin() && RemainingSteps > 0 &&
7906 !RemainingLoadInstrs.
empty();
7907 --MBBItr, --RemainingSteps) {
7911 RemainingLoadInstrs.
erase(&CurrInstr);
7921 if (RemainingSteps == 0 && !RemainingLoadInstrs.
empty())
7947 case AArch64::LD1i32:
7949 case AArch64::LD1i16:
7951 case AArch64::LD1i8:
7967 unsigned Pattern,
unsigned NumLanes) {
7975 for (
unsigned i = 0; i < NumLanes - 1; ++i) {
7983 return A->getOperand(2).getImm() >
B->getOperand(2).getImm();
7988 MRI.getUniqueVRegDef(SubregToReg->getOperand(2).getReg()));
7989 auto LoadToLaneInstrsAscending =
llvm::reverse(LoadToLaneInstrs);
7995 auto CreateLD1Instruction = [&](
MachineInstr *OriginalInstr,
7996 Register SrcRegister,
unsigned Lane,
7998 bool OffsetRegisterKillState) {
7999 auto NewRegister =
MRI.createVirtualRegister(FPR128RegClass);
8006 InstrIdxForVirtReg.
insert(std::make_pair(NewRegister, InsInstrs.
size()));
8007 InsInstrs.
push_back(LoadIndexIntoRegister);
8013 auto CreateLDRInstruction = [&](
unsigned NumLanes,
Register DestReg,
8019 Opcode = AArch64::LDRSui;
8022 Opcode = AArch64::LDRHui;
8025 Opcode = AArch64::LDRBui;
8029 "Got unsupported number of lanes in machine-combiner gather pattern");
8038 auto LanesToLoadToReg0 =
8040 LoadToLaneInstrsAscending.begin() + NumLanes / 2);
8041 Register PrevReg = SubregToReg->getOperand(0).getReg();
8043 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8044 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8045 OffsetRegOperand.
getReg(),
8046 OffsetRegOperand.
isKill());
8053 MachineInstr *Lane0Load = *LoadToLaneInstrsAscending.begin();
8055 *std::next(LoadToLaneInstrsAscending.begin(), NumLanes / 2);
8056 Register DestRegForMiddleIndex =
MRI.createVirtualRegister(
8062 CreateLDRInstruction(NumLanes, DestRegForMiddleIndex,
8063 OriginalSplitToLoadOffsetOperand.
getReg(),
8064 OriginalSplitToLoadOffsetOperand.
isKill());
8066 InstrIdxForVirtReg.
insert(
8067 std::make_pair(DestRegForMiddleIndex, InsInstrs.
size()));
8068 InsInstrs.
push_back(MiddleIndexLoadInstr);
8072 Register DestRegForSubregToReg =
MRI.createVirtualRegister(FPR128RegClass);
8073 unsigned SubregType;
8076 SubregType = AArch64::ssub;
8079 SubregType = AArch64::hsub;
8082 SubregType = AArch64::bsub;
8086 "Got invalid NumLanes for machine-combiner gather pattern");
8089 auto SubRegToRegInstr =
8091 DestRegForSubregToReg)
8095 InstrIdxForVirtReg.
insert(
8096 std::make_pair(DestRegForSubregToReg, InsInstrs.
size()));
8100 auto LanesToLoadToReg1 =
8102 LoadToLaneInstrsAscending.end());
8103 PrevReg = SubRegToRegInstr->getOperand(0).getReg();
8105 const MachineOperand &OffsetRegOperand = LoadInstr->getOperand(3);
8106 PrevReg = CreateLD1Instruction(LoadInstr, PrevReg, Index + 1,
8107 OffsetRegOperand.
getReg(),
8108 OffsetRegOperand.
isKill());
8111 if (Index == NumLanes / 2 - 2) {
8146bool AArch64InstrInfo::getMachineCombinerPatterns(
8148 bool DoRegPressureReduce)
const {
8169 DoRegPressureReduce);
8198 const Register *ReplacedAddend =
nullptr) {
8199 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8201 unsigned IdxOtherOpd = IdxMulOpd == 1 ? 2 : 1;
8204 Register SrcReg0 = MUL->getOperand(1).getReg();
8205 bool Src0IsKill = MUL->getOperand(1).isKill();
8206 Register SrcReg1 = MUL->getOperand(2).getReg();
8207 bool Src1IsKill = MUL->getOperand(2).isKill();
8211 if (ReplacedAddend) {
8213 SrcReg2 = *ReplacedAddend;
8221 MRI.constrainRegClass(ResultReg, RC);
8223 MRI.constrainRegClass(SrcReg0, RC);
8225 MRI.constrainRegClass(SrcReg1, RC);
8227 MRI.constrainRegClass(SrcReg2, RC);
8240 .
addImm(MUL->getOperand(3).getImm());
8247 assert(
false &&
"Invalid FMA instruction kind \n");
8261 if (AArch64::FPR32RegClass.hasSubClassEq(RC))
8262 Opc = AArch64::FNMADDSrrr;
8263 else if (AArch64::FPR64RegClass.hasSubClassEq(RC))
8264 Opc = AArch64::FNMADDDrrr;
8276 MRI.constrainRegClass(ResultReg, RC);
8278 MRI.constrainRegClass(SrcReg0, RC);
8280 MRI.constrainRegClass(SrcReg1, RC);
8282 MRI.constrainRegClass(SrcReg2, RC);
8298 unsigned IdxDupOp,
unsigned MulOpc,
8300 assert(((IdxDupOp == 1) || (IdxDupOp == 2)) &&
8301 "Invalid index of FMUL operand");
8309 if (Dup->
getOpcode() == TargetOpcode::COPY)
8313 MRI.clearKillFlags(DupSrcReg);
8314 MRI.constrainRegClass(DupSrcReg, RC);
8318 unsigned IdxMulOp = IdxDupOp == 1 ? 2 : 1;
8359 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8374 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8401 genNeg(MF,
MRI,
TII, Root, InsInstrs, InstrIdxForVirtReg, MnegOpc, RC);
8429 unsigned IdxMulOpd,
unsigned MaddOpc,
unsigned VR,
8431 assert(IdxMulOpd == 1 || IdxMulOpd == 2);
8435 Register SrcReg0 = MUL->getOperand(1).getReg();
8436 bool Src0IsKill = MUL->getOperand(1).isKill();
8437 Register SrcReg1 = MUL->getOperand(2).getReg();
8438 bool Src1IsKill = MUL->getOperand(2).isKill();
8441 MRI.constrainRegClass(ResultReg, RC);
8443 MRI.constrainRegClass(SrcReg0, RC);
8445 MRI.constrainRegClass(SrcReg1, RC);
8447 MRI.constrainRegClass(VR, RC);
8468 assert(IdxOpd1 == 1 || IdxOpd1 == 2);
8469 unsigned IdxOtherOpd = IdxOpd1 == 1 ? 2 : 1;
8483 if (Opcode == AArch64::SUBSWrr)
8484 Opcode = AArch64::SUBWrr;
8485 else if (Opcode == AArch64::SUBSXrr)
8486 Opcode = AArch64::SUBXrr;
8488 assert((Opcode == AArch64::SUBWrr || Opcode == AArch64::SUBXrr) &&
8489 "Unexpected instruction opcode.");
8506 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8513unsigned AArch64InstrInfo::getReduceOpcodeForAccumulator(
8514 unsigned int AccumulatorOpCode)
const {
8515 switch (AccumulatorOpCode) {
8516 case AArch64::UABALB_ZZZ_D:
8517 case AArch64::SABALB_ZZZ_D:
8518 case AArch64::UABALT_ZZZ_D:
8519 case AArch64::SABALT_ZZZ_D:
8520 return AArch64::ADD_ZZZ_D;
8521 case AArch64::UABALB_ZZZ_H:
8522 case AArch64::SABALB_ZZZ_H:
8523 case AArch64::UABALT_ZZZ_H:
8524 case AArch64::SABALT_ZZZ_H:
8525 return AArch64::ADD_ZZZ_H;
8526 case AArch64::UABALB_ZZZ_S:
8527 case AArch64::SABALB_ZZZ_S:
8528 case AArch64::UABALT_ZZZ_S:
8529 case AArch64::SABALT_ZZZ_S:
8530 return AArch64::ADD_ZZZ_S;
8531 case AArch64::UABALv16i8_v8i16:
8532 case AArch64::SABALv8i8_v8i16:
8533 case AArch64::SABAv8i16:
8534 case AArch64::UABAv8i16:
8535 return AArch64::ADDv8i16;
8536 case AArch64::SABALv2i32_v2i64:
8537 case AArch64::UABALv2i32_v2i64:
8538 case AArch64::SABALv4i32_v2i64:
8539 return AArch64::ADDv2i64;
8540 case AArch64::UABALv4i16_v4i32:
8541 case AArch64::SABALv4i16_v4i32:
8542 case AArch64::SABALv8i16_v4i32:
8543 case AArch64::SABAv4i32:
8544 case AArch64::UABAv4i32:
8545 return AArch64::ADDv4i32;
8546 case AArch64::UABALv4i32_v2i64:
8547 return AArch64::ADDv2i64;
8548 case AArch64::UABALv8i16_v4i32:
8549 return AArch64::ADDv4i32;
8550 case AArch64::UABALv8i8_v8i16:
8551 case AArch64::SABALv16i8_v8i16:
8552 return AArch64::ADDv8i16;
8553 case AArch64::UABAv16i8:
8554 case AArch64::SABAv16i8:
8555 return AArch64::ADDv16i8;
8556 case AArch64::UABAv4i16:
8557 case AArch64::SABAv4i16:
8558 return AArch64::ADDv4i16;
8559 case AArch64::UABAv2i32:
8560 case AArch64::SABAv2i32:
8561 return AArch64::ADDv2i32;
8562 case AArch64::UABAv8i8:
8563 case AArch64::SABAv8i8:
8564 return AArch64::ADDv8i8;
8573void AArch64InstrInfo::genAlternativeCodeSequence(
8583 MachineInstr *
MUL =
nullptr;
8584 const TargetRegisterClass *RC;
8590 DelInstrs, InstrIdxForVirtReg);
8596 InstrIdxForVirtReg);
8602 InstrIdxForVirtReg);
8611 Opc = AArch64::MADDWrrr;
8612 RC = &AArch64::GPR32RegClass;
8614 Opc = AArch64::MADDXrrr;
8615 RC = &AArch64::GPR64RegClass;
8626 Opc = AArch64::MADDWrrr;
8627 RC = &AArch64::GPR32RegClass;
8629 Opc = AArch64::MADDXrrr;
8630 RC = &AArch64::GPR64RegClass;
8643 const TargetRegisterClass *RC;
8644 unsigned BitSize, MovImm;
8647 MovImm = AArch64::MOVi32imm;
8648 RC = &AArch64::GPR32spRegClass;
8650 Opc = AArch64::MADDWrrr;
8651 RC = &AArch64::GPR32RegClass;
8653 MovImm = AArch64::MOVi64imm;
8654 RC = &AArch64::GPR64spRegClass;
8656 Opc = AArch64::MADDXrrr;
8657 RC = &AArch64::GPR64RegClass;
8668 uint64_t UImm =
SignExtend64(IsSub ? -Imm : Imm, BitSize);
8672 if (Insn.
size() != 1)
8674 MachineInstrBuilder MIB1 =
8675 BuildMI(MF, MIMetadata(Root),
TII->get(MovImm), NewVR)
8676 .
addImm(IsSub ? -Imm : Imm);
8678 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8689 const TargetRegisterClass *SubRC;
8690 unsigned SubOpc, ZeroReg;
8692 SubOpc = AArch64::SUBWrr;
8693 SubRC = &AArch64::GPR32spRegClass;
8694 ZeroReg = AArch64::WZR;
8695 Opc = AArch64::MADDWrrr;
8696 RC = &AArch64::GPR32RegClass;
8698 SubOpc = AArch64::SUBXrr;
8699 SubRC = &AArch64::GPR64spRegClass;
8700 ZeroReg = AArch64::XZR;
8701 Opc = AArch64::MADDXrrr;
8702 RC = &AArch64::GPR64RegClass;
8704 Register NewVR =
MRI.createVirtualRegister(SubRC);
8706 MachineInstrBuilder MIB1 =
8707 BuildMI(MF, MIMetadata(Root),
TII->get(SubOpc), NewVR)
8711 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
8722 Opc = AArch64::MSUBWrrr;
8723 RC = &AArch64::GPR32RegClass;
8725 Opc = AArch64::MSUBXrrr;
8726 RC = &AArch64::GPR64RegClass;
8731 Opc = AArch64::MLAv8i8;
8732 RC = &AArch64::FPR64RegClass;
8736 Opc = AArch64::MLAv8i8;
8737 RC = &AArch64::FPR64RegClass;
8741 Opc = AArch64::MLAv16i8;
8742 RC = &AArch64::FPR128RegClass;
8746 Opc = AArch64::MLAv16i8;
8747 RC = &AArch64::FPR128RegClass;
8751 Opc = AArch64::MLAv4i16;
8752 RC = &AArch64::FPR64RegClass;
8756 Opc = AArch64::MLAv4i16;
8757 RC = &AArch64::FPR64RegClass;
8761 Opc = AArch64::MLAv8i16;
8762 RC = &AArch64::FPR128RegClass;
8766 Opc = AArch64::MLAv8i16;
8767 RC = &AArch64::FPR128RegClass;
8771 Opc = AArch64::MLAv2i32;
8772 RC = &AArch64::FPR64RegClass;
8776 Opc = AArch64::MLAv2i32;
8777 RC = &AArch64::FPR64RegClass;
8781 Opc = AArch64::MLAv4i32;
8782 RC = &AArch64::FPR128RegClass;
8786 Opc = AArch64::MLAv4i32;
8787 RC = &AArch64::FPR128RegClass;
8792 Opc = AArch64::MLAv8i8;
8793 RC = &AArch64::FPR64RegClass;
8795 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i8,
8799 Opc = AArch64::MLSv8i8;
8800 RC = &AArch64::FPR64RegClass;
8804 Opc = AArch64::MLAv16i8;
8805 RC = &AArch64::FPR128RegClass;
8807 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv16i8,
8811 Opc = AArch64::MLSv16i8;
8812 RC = &AArch64::FPR128RegClass;
8816 Opc = AArch64::MLAv4i16;
8817 RC = &AArch64::FPR64RegClass;
8819 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8823 Opc = AArch64::MLSv4i16;
8824 RC = &AArch64::FPR64RegClass;
8828 Opc = AArch64::MLAv8i16;
8829 RC = &AArch64::FPR128RegClass;
8831 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8835 Opc = AArch64::MLSv8i16;
8836 RC = &AArch64::FPR128RegClass;
8840 Opc = AArch64::MLAv2i32;
8841 RC = &AArch64::FPR64RegClass;
8843 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8847 Opc = AArch64::MLSv2i32;
8848 RC = &AArch64::FPR64RegClass;
8852 Opc = AArch64::MLAv4i32;
8853 RC = &AArch64::FPR128RegClass;
8855 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
8859 Opc = AArch64::MLSv4i32;
8860 RC = &AArch64::FPR128RegClass;
8865 Opc = AArch64::MLAv4i16_indexed;
8866 RC = &AArch64::FPR64RegClass;
8870 Opc = AArch64::MLAv4i16_indexed;
8871 RC = &AArch64::FPR64RegClass;
8875 Opc = AArch64::MLAv8i16_indexed;
8876 RC = &AArch64::FPR128RegClass;
8880 Opc = AArch64::MLAv8i16_indexed;
8881 RC = &AArch64::FPR128RegClass;
8885 Opc = AArch64::MLAv2i32_indexed;
8886 RC = &AArch64::FPR64RegClass;
8890 Opc = AArch64::MLAv2i32_indexed;
8891 RC = &AArch64::FPR64RegClass;
8895 Opc = AArch64::MLAv4i32_indexed;
8896 RC = &AArch64::FPR128RegClass;
8900 Opc = AArch64::MLAv4i32_indexed;
8901 RC = &AArch64::FPR128RegClass;
8906 Opc = AArch64::MLAv4i16_indexed;
8907 RC = &AArch64::FPR64RegClass;
8909 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i16,
8913 Opc = AArch64::MLSv4i16_indexed;
8914 RC = &AArch64::FPR64RegClass;
8918 Opc = AArch64::MLAv8i16_indexed;
8919 RC = &AArch64::FPR128RegClass;
8921 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv8i16,
8925 Opc = AArch64::MLSv8i16_indexed;
8926 RC = &AArch64::FPR128RegClass;
8930 Opc = AArch64::MLAv2i32_indexed;
8931 RC = &AArch64::FPR64RegClass;
8933 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv2i32,
8937 Opc = AArch64::MLSv2i32_indexed;
8938 RC = &AArch64::FPR64RegClass;
8942 Opc = AArch64::MLAv4i32_indexed;
8943 RC = &AArch64::FPR128RegClass;
8945 InstrIdxForVirtReg, 1,
Opc, AArch64::NEGv4i32,
8949 Opc = AArch64::MLSv4i32_indexed;
8950 RC = &AArch64::FPR128RegClass;
8956 Opc = AArch64::FMADDHrrr;
8957 RC = &AArch64::FPR16RegClass;
8961 Opc = AArch64::FMADDSrrr;
8962 RC = &AArch64::FPR32RegClass;
8966 Opc = AArch64::FMADDDrrr;
8967 RC = &AArch64::FPR64RegClass;
8972 Opc = AArch64::FMADDHrrr;
8973 RC = &AArch64::FPR16RegClass;
8977 Opc = AArch64::FMADDSrrr;
8978 RC = &AArch64::FPR32RegClass;
8982 Opc = AArch64::FMADDDrrr;
8983 RC = &AArch64::FPR64RegClass;
8988 Opc = AArch64::FMLAv1i32_indexed;
8989 RC = &AArch64::FPR32RegClass;
8994 Opc = AArch64::FMLAv1i32_indexed;
8995 RC = &AArch64::FPR32RegClass;
9001 Opc = AArch64::FMLAv1i64_indexed;
9002 RC = &AArch64::FPR64RegClass;
9007 Opc = AArch64::FMLAv1i64_indexed;
9008 RC = &AArch64::FPR64RegClass;
9014 RC = &AArch64::FPR64RegClass;
9015 Opc = AArch64::FMLAv4i16_indexed;
9020 RC = &AArch64::FPR64RegClass;
9021 Opc = AArch64::FMLAv4f16;
9026 RC = &AArch64::FPR64RegClass;
9027 Opc = AArch64::FMLAv4i16_indexed;
9032 RC = &AArch64::FPR64RegClass;
9033 Opc = AArch64::FMLAv4f16;
9040 RC = &AArch64::FPR64RegClass;
9042 Opc = AArch64::FMLAv2i32_indexed;
9046 Opc = AArch64::FMLAv2f32;
9053 RC = &AArch64::FPR64RegClass;
9055 Opc = AArch64::FMLAv2i32_indexed;
9059 Opc = AArch64::FMLAv2f32;
9066 RC = &AArch64::FPR128RegClass;
9067 Opc = AArch64::FMLAv8i16_indexed;
9072 RC = &AArch64::FPR128RegClass;
9073 Opc = AArch64::FMLAv8f16;
9078 RC = &AArch64::FPR128RegClass;
9079 Opc = AArch64::FMLAv8i16_indexed;
9084 RC = &AArch64::FPR128RegClass;
9085 Opc = AArch64::FMLAv8f16;
9092 RC = &AArch64::FPR128RegClass;
9094 Opc = AArch64::FMLAv2i64_indexed;
9098 Opc = AArch64::FMLAv2f64;
9105 RC = &AArch64::FPR128RegClass;
9107 Opc = AArch64::FMLAv2i64_indexed;
9111 Opc = AArch64::FMLAv2f64;
9119 RC = &AArch64::FPR128RegClass;
9121 Opc = AArch64::FMLAv4i32_indexed;
9125 Opc = AArch64::FMLAv4f32;
9133 RC = &AArch64::FPR128RegClass;
9135 Opc = AArch64::FMLAv4i32_indexed;
9139 Opc = AArch64::FMLAv4f32;
9146 Opc = AArch64::FNMSUBHrrr;
9147 RC = &AArch64::FPR16RegClass;
9151 Opc = AArch64::FNMSUBSrrr;
9152 RC = &AArch64::FPR32RegClass;
9156 Opc = AArch64::FNMSUBDrrr;
9157 RC = &AArch64::FPR64RegClass;
9162 Opc = AArch64::FNMADDHrrr;
9163 RC = &AArch64::FPR16RegClass;
9167 Opc = AArch64::FNMADDSrrr;
9168 RC = &AArch64::FPR32RegClass;
9172 Opc = AArch64::FNMADDDrrr;
9173 RC = &AArch64::FPR64RegClass;
9178 Opc = AArch64::FMSUBHrrr;
9179 RC = &AArch64::FPR16RegClass;
9183 Opc = AArch64::FMSUBSrrr;
9184 RC = &AArch64::FPR32RegClass;
9188 Opc = AArch64::FMSUBDrrr;
9189 RC = &AArch64::FPR64RegClass;
9194 Opc = AArch64::FMLSv1i32_indexed;
9195 RC = &AArch64::FPR32RegClass;
9201 Opc = AArch64::FMLSv1i64_indexed;
9202 RC = &AArch64::FPR64RegClass;
9209 RC = &AArch64::FPR64RegClass;
9211 MachineInstrBuilder MIB1 =
9212 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f16), NewVR)
9215 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9217 Opc = AArch64::FMLAv4f16;
9221 Opc = AArch64::FMLAv4i16_indexed;
9228 RC = &AArch64::FPR64RegClass;
9229 Opc = AArch64::FMLSv4f16;
9234 RC = &AArch64::FPR64RegClass;
9235 Opc = AArch64::FMLSv4i16_indexed;
9242 RC = &AArch64::FPR64RegClass;
9244 Opc = AArch64::FMLSv2i32_indexed;
9248 Opc = AArch64::FMLSv2f32;
9256 RC = &AArch64::FPR128RegClass;
9258 MachineInstrBuilder MIB1 =
9259 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv8f16), NewVR)
9262 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9264 Opc = AArch64::FMLAv8f16;
9268 Opc = AArch64::FMLAv8i16_indexed;
9275 RC = &AArch64::FPR128RegClass;
9276 Opc = AArch64::FMLSv8f16;
9281 RC = &AArch64::FPR128RegClass;
9282 Opc = AArch64::FMLSv8i16_indexed;
9289 RC = &AArch64::FPR128RegClass;
9291 Opc = AArch64::FMLSv2i64_indexed;
9295 Opc = AArch64::FMLSv2f64;
9303 RC = &AArch64::FPR128RegClass;
9305 Opc = AArch64::FMLSv4i32_indexed;
9309 Opc = AArch64::FMLSv4f32;
9316 RC = &AArch64::FPR64RegClass;
9318 MachineInstrBuilder MIB1 =
9319 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f32), NewVR)
9322 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9324 Opc = AArch64::FMLAv2i32_indexed;
9328 Opc = AArch64::FMLAv2f32;
9336 RC = &AArch64::FPR128RegClass;
9338 MachineInstrBuilder MIB1 =
9339 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv4f32), NewVR)
9342 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9344 Opc = AArch64::FMLAv4i32_indexed;
9348 Opc = AArch64::FMLAv4f32;
9356 RC = &AArch64::FPR128RegClass;
9358 MachineInstrBuilder MIB1 =
9359 BuildMI(MF, MIMetadata(Root),
TII->get(AArch64::FNEGv2f64), NewVR)
9362 InstrIdxForVirtReg.
insert(std::make_pair(NewVR, 0));
9364 Opc = AArch64::FMLAv2i64_indexed;
9368 Opc = AArch64::FMLAv2f64;
9380 &AArch64::FPR128RegClass,
MRI);
9389 &AArch64::FPR128RegClass,
MRI);
9398 &AArch64::FPR128_loRegClass,
MRI);
9407 &AArch64::FPR128RegClass,
MRI);
9416 &AArch64::FPR128_loRegClass,
MRI);
9450 for (
auto *
MI : InsInstrs)
9451 MI->setFlags(Flags);
9492 bool IsNegativeBranch =
false;
9493 bool IsTestAndBranch =
false;
9494 unsigned TargetBBInMI = 0;
9495 switch (
MI.getOpcode()) {
9499 case AArch64::CBWPri:
9500 case AArch64::CBXPri:
9501 case AArch64::CBBAssertExt:
9502 case AArch64::CBHAssertExt:
9503 case AArch64::CBWPrr:
9504 case AArch64::CBXPrr:
9510 case AArch64::CBNZW:
9511 case AArch64::CBNZX:
9513 IsNegativeBranch =
true;
9518 IsTestAndBranch =
true;
9520 case AArch64::TBNZW:
9521 case AArch64::TBNZX:
9523 IsNegativeBranch =
true;
9524 IsTestAndBranch =
true;
9530 if (IsTestAndBranch &&
MI.getOperand(1).getImm())
9534 assert(
MI.getParent() &&
"Incomplete machine instruction\n");
9545 while (
DefMI->isCopy()) {
9547 if (!
MRI->hasOneNonDBGUse(CopyVReg))
9549 if (!
MRI->hasOneDef(CopyVReg))
9554 switch (
DefMI->getOpcode()) {
9558 case AArch64::ANDWri:
9559 case AArch64::ANDXri: {
9560 if (IsTestAndBranch)
9564 if (!
MRI->hasOneNonDBGUse(VReg))
9567 bool Is32Bit = (
DefMI->getOpcode() == AArch64::ANDWri);
9569 DefMI->getOperand(2).getImm(), Is32Bit ? 32 : 64);
9578 assert(!
MRI->def_empty(NewReg) &&
"Register must be defined.");
9584 unsigned Opc = (Imm < 32)
9585 ? (IsNegativeBranch ? AArch64::TBNZW : AArch64::TBZW)
9586 : (IsNegativeBranch ? AArch64::TBNZX : AArch64::TBZX);
9599 if (!Is32Bit && Imm < 32)
9601 MI.eraseFromParent();
9605 case AArch64::CSINCWr:
9606 case AArch64::CSINCXr: {
9607 if (!(
DefMI->getOperand(1).getReg() == AArch64::WZR &&
9608 DefMI->getOperand(2).getReg() == AArch64::WZR) &&
9609 !(
DefMI->getOperand(1).getReg() == AArch64::XZR &&
9610 DefMI->getOperand(2).getReg() == AArch64::XZR))
9613 if (
DefMI->findRegisterDefOperandIdx(AArch64::NZCV,
nullptr,
9626 if (IsNegativeBranch)
9629 MI.eraseFromParent();
9635std::pair<unsigned, unsigned>
9636AArch64InstrInfo::decomposeMachineOperandsTargetFlags(
unsigned TF)
const {
9638 return std::make_pair(TF & Mask, TF & ~Mask);
9642AArch64InstrInfo::getSerializableDirectMachineOperandTargetFlags()
const {
9645 static const std::pair<unsigned, const char *> TargetFlags[] = {
9646 {MO_PAGE,
"aarch64-page"}, {
MO_PAGEOFF,
"aarch64-pageoff"},
9647 {
MO_G3,
"aarch64-g3"}, {
MO_G2,
"aarch64-g2"},
9648 {
MO_G1,
"aarch64-g1"}, {
MO_G0,
"aarch64-g0"},
9654AArch64InstrInfo::getSerializableBitmaskMachineOperandTargetFlags()
const {
9655 using namespace AArch64II;
9657 static const std::pair<unsigned, const char *> TargetFlags[] = {
9660 {
MO_NC,
"aarch64-nc"},
9661 {
MO_S,
"aarch64-s"},
9672AArch64InstrInfo::getSerializableMachineMemOperandTargetFlags()
const {
9673 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
9775 MachineFunction *MF =
C.getMF();
9777 const AArch64RegisterInfo *ARI =
9778 static_cast<const AArch64RegisterInfo *
>(&
TRI);
9781 for (
unsigned Reg : AArch64::GPR64RegClass) {
9783 Reg != AArch64::LR &&
9784 Reg != AArch64::X16 &&
9785 Reg != AArch64::X17 &&
9786 C.isAvailableAcrossAndOutOfSeq(
Reg,
TRI) &&
9787 C.isAvailableInsideSeq(
Reg,
TRI))
9818 return SubtargetA.hasV8_3aOps() == SubtargetB.hasV8_3aOps();
9821std::optional<std::unique_ptr<outliner::OutlinedFunction>>
9822AArch64InstrInfo::getOutliningCandidateInfo(
9824 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
9825 unsigned MinRepeats)
const {
9826 unsigned SequenceSize = 0;
9827 for (
auto &
MI : RepeatedSequenceLocs[0])
9830 unsigned NumBytesToCreateFrame = 0;
9836 MachineInstr &LastMI = RepeatedSequenceLocs[0].back();
9837 MachineInstr &FirstMI = RepeatedSequenceLocs[0].front();
9838 if (LastMI.
getOpcode() == AArch64::ADRP &&
9841 return std::nullopt;
9846 if ((FirstMI.
getOpcode() == AArch64::ADDXri ||
9847 FirstMI.
getOpcode() == AArch64::LDRXui) &&
9850 return std::nullopt;
9861 if (std::adjacent_find(
9862 RepeatedSequenceLocs.begin(), RepeatedSequenceLocs.end(),
9863 [](
const outliner::Candidate &a,
const outliner::Candidate &b) {
9866 if (outliningCandidatesSigningScopeConsensus(a, b) &&
9867 outliningCandidatesSigningKeyConsensus(a, b) &&
9868 outliningCandidatesV8_3OpsConsensus(a, b)) {
9872 }) != RepeatedSequenceLocs.end()) {
9873 return std::nullopt;
9890 unsigned NumBytesToCheckLRInTCEpilogue = 0;
9891 const auto RASignCondition = RepeatedSequenceLocs[0]
9894 ->getSignReturnAddressCondition();
9897 NumBytesToCreateFrame += 8;
9900 auto LRCheckMethod = Subtarget.getAuthenticatedLRCheckMethod(
9901 *RepeatedSequenceLocs[0].getMF());
9902 NumBytesToCheckLRInTCEpilogue =
9906 if (isTailCallReturnInst(RepeatedSequenceLocs[0].
back()))
9907 SequenceSize += NumBytesToCheckLRInTCEpilogue;
9915 for (
auto &
MI :
C) {
9916 if (
MI.modifiesRegister(AArch64::SP, &
TRI)) {
9917 switch (
MI.getOpcode()) {
9918 case AArch64::ADDXri:
9919 case AArch64::ADDWri:
9920 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
9922 "Expected operand to be immediate");
9924 "Expected operand to be a register");
9928 if (
MI.getOperand(1).getReg() == AArch64::SP)
9929 SPValue +=
MI.getOperand(2).getImm();
9933 case AArch64::SUBXri:
9934 case AArch64::SUBWri:
9935 assert(
MI.getNumOperands() == 4 &&
"Wrong number of operands");
9937 "Expected operand to be immediate");
9939 "Expected operand to be a register");
9943 if (
MI.getOperand(1).getReg() == AArch64::SP)
9944 SPValue -=
MI.getOperand(2).getImm();
9961 if (RepeatedSequenceLocs.size() < MinRepeats)
9962 return std::nullopt;
9966 unsigned FlagsSetInAll = 0xF;
9970 FlagsSetInAll &=
C.Flags;
9972 unsigned LastInstrOpcode = RepeatedSequenceLocs[0].back().getOpcode();
9975 auto SetCandidateCallInfo =
9976 [&RepeatedSequenceLocs](
unsigned CallID,
unsigned NumBytesForCall) {
9978 C.setCallInfo(CallID, NumBytesForCall);
9982 NumBytesToCreateFrame += 4;
9990 unsigned CFICount = 0;
9991 for (
auto &
I : RepeatedSequenceLocs[0]) {
9992 if (
I.isCFIInstruction())
10002 std::vector<MCCFIInstruction> CFIInstructions =
10003 C.getMF()->getFrameInstructions();
10005 if (CFICount > 0 && CFICount != CFIInstructions.size())
10006 return std::nullopt;
10014 if (!
MI.modifiesRegister(AArch64::SP, &
TRI) &&
10015 !
MI.readsRegister(AArch64::SP, &
TRI))
10021 if (
MI.modifiesRegister(AArch64::SP, &
TRI))
10026 if (
MI.mayLoadOrStore()) {
10029 bool OffsetIsScalable;
10033 if (!getMemOperandWithOffset(
MI,
Base,
Offset, OffsetIsScalable, &
TRI) ||
10034 !
Base->isReg() ||
Base->getReg() != AArch64::SP)
10038 if (OffsetIsScalable)
10046 TypeSize Scale(0U,
false), DummyWidth(0U,
false);
10047 getMemOpInfo(
MI.getOpcode(), Scale, DummyWidth, MinOffset, MaxOffset);
10050 if (
Offset < MinOffset * (int64_t)Scale.getFixedValue() ||
10051 Offset > MaxOffset * (int64_t)Scale.getFixedValue())
10066 bool AllStackInstrsSafe =
10071 if (RepeatedSequenceLocs[0].
back().isTerminator()) {
10073 NumBytesToCreateFrame = 0;
10074 unsigned NumBytesForCall = 4 + NumBytesToCheckLRInTCEpilogue;
10078 else if (LastInstrOpcode == AArch64::BL ||
10079 ((LastInstrOpcode == AArch64::BLR ||
10080 LastInstrOpcode == AArch64::BLRNoIP) &&
10084 NumBytesToCreateFrame = NumBytesToCheckLRInTCEpilogue;
10092 unsigned NumBytesNoStackCalls = 0;
10093 std::vector<outliner::Candidate> CandidatesWithoutStackFixups;
10099 ?
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI)
10108 C.getMF()->getFunction().hasFnAttribute(Attribute::NoReturn);
10111 if (LRAvailable && !IsNoReturn) {
10112 NumBytesNoStackCalls += 4;
10114 CandidatesWithoutStackFixups.push_back(
C);
10119 else if (findRegisterToSaveLRTo(
C)) {
10120 NumBytesNoStackCalls += 12;
10122 CandidatesWithoutStackFixups.push_back(
C);
10127 else if (
C.isAvailableInsideSeq(AArch64::SP,
TRI)) {
10128 NumBytesNoStackCalls += 12;
10130 CandidatesWithoutStackFixups.push_back(
C);
10136 NumBytesNoStackCalls += SequenceSize;
10143 if (!AllStackInstrsSafe ||
10144 NumBytesNoStackCalls <= RepeatedSequenceLocs.size() * 12) {
10145 RepeatedSequenceLocs = CandidatesWithoutStackFixups;
10147 if (RepeatedSequenceLocs.size() < MinRepeats)
10148 return std::nullopt;
10201 (!
C.isAvailableAcrossAndOutOfSeq(AArch64::LR,
TRI) ||
10202 !findRegisterToSaveLRTo(
C));
10208 if (RepeatedSequenceLocs.size() < MinRepeats)
10209 return std::nullopt;
10218 bool ModStackToSaveLR =
false;
10221 ModStackToSaveLR =
true;
10230 ModStackToSaveLR =
true;
10232 if (ModStackToSaveLR) {
10234 if (!AllStackInstrsSafe)
10235 return std::nullopt;
10238 NumBytesToCreateFrame += 8;
10245 return std::nullopt;
10247 return std::make_unique<outliner::OutlinedFunction>(
10248 RepeatedSequenceLocs, SequenceSize, NumBytesToCreateFrame, FrameID);
10251void AArch64InstrInfo::mergeOutliningCandidateAttributes(
10252 Function &
F, std::vector<outliner::Candidate> &Candidates)
const {
10256 const auto &CFn = Candidates.front().getMF()->getFunction();
10258 if (CFn.hasFnAttribute(
"ptrauth-returns"))
10259 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-returns"));
10260 if (CFn.hasFnAttribute(
"ptrauth-auth-traps"))
10261 F.addFnAttr(CFn.getFnAttribute(
"ptrauth-auth-traps"));
10264 if (CFn.hasFnAttribute(
"sign-return-address"))
10265 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address"));
10266 if (CFn.hasFnAttribute(
"sign-return-address-key"))
10267 F.addFnAttr(CFn.getFnAttribute(
"sign-return-address-key"));
10269 AArch64GenInstrInfo::mergeOutliningCandidateAttributes(
F, Candidates);
10272bool AArch64InstrInfo::isFunctionSafeToOutlineFrom(
10277 if (!OutlineFromLinkOnceODRs &&
F.hasLinkOnceODRLinkage())
10284 if (
F.hasSection())
10290 AArch64FunctionInfo *AFI = MF.
getInfo<AArch64FunctionInfo>();
10291 if (!AFI || AFI->
hasRedZone().value_or(
true))
10311 unsigned &Flags)
const {
10313 "Must track liveness!");
10315 std::pair<MachineBasicBlock::iterator, MachineBasicBlock::iterator>>
10330 auto AreAllUnsafeRegsDead = [&LRU]() {
10331 return LRU.available(AArch64::W16) && LRU.available(AArch64::W17) &&
10332 LRU.available(AArch64::NZCV);
10347 bool LRAvailableEverywhere =
true;
10349 LRU.addLiveOuts(
MBB);
10351 auto UpdateWholeMBBFlags = [&
Flags](
const MachineInstr &
MI) {
10352 if (
MI.isCall() && !
MI.isTerminator())
10358 auto CreateNewRangeStartingAt =
10359 [&RangeBegin, &RangeEnd,
10361 RangeBegin = NewBegin;
10362 RangeEnd = std::next(RangeBegin);
10365 auto SaveRangeIfNonEmpty = [&RangeLen, &
Ranges, &RangeBegin, &RangeEnd]() {
10371 if (!RangeBegin.isEnd() && RangeBegin->isBundledWithPred())
10373 if (!RangeEnd.isEnd() && RangeEnd->isBundledWithPred())
10375 Ranges.emplace_back(RangeBegin, RangeEnd);
10383 for (; FirstPossibleEndPt !=
MBB.
instr_rend(); ++FirstPossibleEndPt) {
10384 LRU.stepBackward(*FirstPossibleEndPt);
10387 UpdateWholeMBBFlags(*FirstPossibleEndPt);
10388 if (AreAllUnsafeRegsDead())
10395 CreateNewRangeStartingAt(FirstPossibleEndPt->getIterator());
10400 LRU.stepBackward(
MI);
10401 UpdateWholeMBBFlags(
MI);
10402 if (!AreAllUnsafeRegsDead()) {
10403 SaveRangeIfNonEmpty();
10404 CreateNewRangeStartingAt(
MI.getIterator());
10407 LRAvailableEverywhere &= LRU.available(AArch64::LR);
10408 RangeBegin =
MI.getIterator();
10413 if (AreAllUnsafeRegsDead())
10414 SaveRangeIfNonEmpty();
10422 if (!LRAvailableEverywhere)
10430 unsigned Flags)
const {
10431 MachineInstr &
MI = *MIT;
10435 switch (
MI.getOpcode()) {
10436 case AArch64::PACM:
10437 case AArch64::PACIASP:
10438 case AArch64::PACIBSP:
10439 case AArch64::PACIASPPC:
10440 case AArch64::PACIBSPPC:
10441 case AArch64::AUTIASP:
10442 case AArch64::AUTIBSP:
10443 case AArch64::AUTIASPPCi:
10444 case AArch64::AUTIASPPCr:
10445 case AArch64::AUTIBSPPCi:
10446 case AArch64::AUTIBSPPCr:
10447 case AArch64::RETAA:
10448 case AArch64::RETAB:
10449 case AArch64::RETAASPPCi:
10450 case AArch64::RETAASPPCr:
10451 case AArch64::RETABSPPCi:
10452 case AArch64::RETABSPPCr:
10453 case AArch64::EMITBKEY:
10454 case AArch64::PAUTH_PROLOGUE:
10455 case AArch64::PAUTH_EPILOGUE:
10465 if (
MI.isCFIInstruction())
10469 if (
MI.isTerminator())
10475 for (
const MachineOperand &MOP :
MI.operands()) {
10478 assert(!MOP.isCFIIndex());
10481 if (MOP.isReg() && !MOP.isImplicit() &&
10482 (MOP.getReg() == AArch64::LR || MOP.getReg() == AArch64::W30))
10489 if (
MI.getOpcode() == AArch64::ADRP)
10509 for (
const MachineOperand &MOP :
MI.operands()) {
10510 if (MOP.isGlobal()) {
10518 if (Callee &&
Callee->getName() ==
"\01_mcount")
10526 if (
MI.getOpcode() == AArch64::BLR ||
10527 MI.getOpcode() == AArch64::BLRNoIP ||
MI.getOpcode() == AArch64::BL)
10531 return UnknownCallOutlineType;
10539 return UnknownCallOutlineType;
10547 return UnknownCallOutlineType;
10568 for (MachineInstr &
MI :
MBB) {
10569 const MachineOperand *
Base;
10570 TypeSize Width(0,
false);
10572 bool OffsetIsScalable;
10575 if (!
MI.mayLoadOrStore() ||
10578 (
Base->isReg() &&
Base->getReg() != AArch64::SP))
10582 TypeSize Scale(0U,
false);
10583 int64_t Dummy1, Dummy2;
10586 assert(StackOffsetOperand.
isImm() &&
"Stack offset wasn't immediate!");
10588 assert(Scale != 0 &&
"Unexpected opcode!");
10589 assert(!OffsetIsScalable &&
"Expected offset to be a byte offset");
10594 int64_t NewImm = (
Offset + 16) / (int64_t)Scale.getFixedValue();
10595 StackOffsetOperand.
setImm(NewImm);
10601 bool ShouldSignReturnAddr) {
10602 if (!ShouldSignReturnAddr)
10608 TII->get(AArch64::PAUTH_EPILOGUE))
10612void AArch64InstrInfo::buildOutlinedFrame(
10616 AArch64FunctionInfo *FI = MF.
getInfo<AArch64FunctionInfo>();
10624 unsigned TailOpcode;
10626 TailOpcode = AArch64::TCRETURNdi;
10630 TailOpcode = AArch64::TCRETURNriALL;
10641 bool IsLeafFunction =
true;
10644 auto IsNonTailCall = [](
const MachineInstr &
MI) {
10645 return MI.isCall() && !
MI.isReturn();
10655 "Can only fix up stack references once");
10656 fixupPostOutline(
MBB);
10658 IsLeafFunction =
false;
10669 Et = std::prev(
MBB.
end());
10679 if (MF.
getInfo<AArch64FunctionInfo>()->needsDwarfUnwindInfo(MF)) {
10683 CFIBuilder.buildDefCFAOffset(16);
10687 CFIBuilder.buildOffset(AArch64::LR, -16);
10701 RASignCondition, !IsLeafFunction);
10730 fixupPostOutline(
MBB);
10741 .addGlobalAddress(
M.getNamedValue(MF.
getName()))
10751 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10760 MachineInstr *Save;
10761 MachineInstr *Restore;
10767 assert(
Reg &&
"No callee-saved register available?");
10801 .addGlobalAddress(
M.getNamedValue(MF.
getName())));
10809bool AArch64InstrInfo::shouldOutlineFromFunctionByDefault(
10817 bool AllowSideEffects)
const {
10819 const AArch64Subtarget &STI = MF.
getSubtarget<AArch64Subtarget>();
10822 if (
TRI.isGeneralPurposeRegister(MF,
Reg)) {
10835 assert(STI.hasNEON() &&
"Expected to have NEON.");
10841std::optional<DestSourcePair>
10846 if (((
MI.getOpcode() == AArch64::ORRWrs &&
10847 MI.getOperand(1).getReg() == AArch64::WZR &&
10848 MI.getOperand(3).getImm() == 0x0) ||
10849 (
MI.getOpcode() == AArch64::ORRWrr &&
10850 MI.getOperand(1).getReg() == AArch64::WZR)) &&
10852 (!
MI.getOperand(0).getReg().isVirtual() ||
10853 MI.getOperand(0).getSubReg() == 0) &&
10854 (!
MI.getOperand(0).getReg().isPhysical() ||
10859 if (
MI.getOpcode() == AArch64::ORRXrs &&
10860 MI.getOperand(1).getReg() == AArch64::XZR &&
10861 MI.getOperand(3).getImm() == 0x0)
10864 return std::nullopt;
10867std::optional<DestSourcePair>
10869 if ((
MI.getOpcode() == AArch64::ORRWrs &&
10870 MI.getOperand(1).getReg() == AArch64::WZR &&
10871 MI.getOperand(3).getImm() == 0x0) ||
10872 (
MI.getOpcode() == AArch64::ORRWrr &&
10873 MI.getOperand(1).getReg() == AArch64::WZR))
10875 return std::nullopt;
10878std::optional<RegImmPair>
10887 return std::nullopt;
10889 switch (
MI.getOpcode()) {
10891 return std::nullopt;
10892 case AArch64::SUBWri:
10893 case AArch64::SUBXri:
10894 case AArch64::SUBSWri:
10895 case AArch64::SUBSXri:
10898 case AArch64::ADDSWri:
10899 case AArch64::ADDSXri:
10900 case AArch64::ADDWri:
10901 case AArch64::ADDXri: {
10903 if (!
MI.getOperand(0).isReg() || !
MI.getOperand(1).isReg() ||
10904 !
MI.getOperand(2).isImm())
10905 return std::nullopt;
10906 int Shift =
MI.getOperand(3).getImm();
10907 assert((Shift == 0 || Shift == 12) &&
"Shift can be either 0 or 12");
10911 return RegImmPair{
MI.getOperand(1).getReg(),
Offset};
10917static std::optional<ParamLoadedValue>
10921 auto DestSrc =
TII->isCopyLikeInstr(
MI);
10923 return std::nullopt;
10925 Register DestReg = DestSrc->Destination->getReg();
10926 Register SrcReg = DestSrc->Source->getReg();
10929 return std::nullopt;
10934 if (DestReg == DescribedReg)
10938 if (
MI.getOpcode() == AArch64::ORRWrs &&
10939 TRI->isSuperRegister(DestReg, DescribedReg))
10943 if (
MI.getOpcode() == AArch64::ORRXrs &&
10944 TRI->isSubRegister(DestReg, DescribedReg)) {
10945 Register SrcSubReg =
TRI->getSubReg(SrcReg, AArch64::sub_32);
10949 assert(!
TRI->isSuperOrSubRegisterEq(DestReg, DescribedReg) &&
10950 "Unhandled ORR[XW]rs copy case");
10952 return std::nullopt;
10955bool AArch64InstrInfo::isFunctionSafeToSplit(
const MachineFunction &MF)
const {
10960 if (MF.
getInfo<AArch64FunctionInfo>()->hasRedZone().value_or(
true))
10966bool AArch64InstrInfo::isMBBSafeToSplitToCold(
10970 auto isAsmGoto = [](
const MachineInstr &
MI) {
10971 return MI.getOpcode() == AArch64::INLINEASM_BR;
10981 auto containsMBB = [&
MBB](
const MachineJumpTableEntry &JTE) {
10988 for (
const MachineInstr &
MI :
MBB) {
10989 switch (
MI.getOpcode()) {
10990 case TargetOpcode::G_BRJT:
10991 case AArch64::JumpTableDest32:
10992 case AArch64::JumpTableDest16:
10993 case AArch64::JumpTableDest8:
11004std::optional<ParamLoadedValue>
11007 const MachineFunction *MF =
MI.getMF();
11009 switch (
MI.getOpcode()) {
11010 case AArch64::MOVZWi:
11011 case AArch64::MOVZXi: {
11014 if (!
TRI->isSuperRegisterEq(
MI.getOperand(0).getReg(),
Reg))
11015 return std::nullopt;
11017 if (!
MI.getOperand(1).isImm())
11018 return std::nullopt;
11019 int64_t Immediate =
MI.getOperand(1).getImm();
11020 int Shift =
MI.getOperand(2).getImm();
11024 case AArch64::ORRWrs:
11025 case AArch64::ORRXrs:
11032bool AArch64InstrInfo::isExtendLikelyToBeFolded(
11035 ExtMI.
getOpcode() == TargetOpcode::G_ZEXT ||
11036 ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT);
11039 if (ExtMI.
getOpcode() == TargetOpcode::G_ANYEXT)
11043 if (!
MRI.hasOneNonDBGUse(DefReg))
11048 auto *UserMI = &*
MRI.use_instr_nodbg_begin(DefReg);
11049 return UserMI->getOpcode() == TargetOpcode::G_PTR_ADD;
11052uint64_t AArch64InstrInfo::getElementSizeForOpcode(
unsigned Opc)
const {
11056bool AArch64InstrInfo::isPTestLikeOpcode(
unsigned Opc)
const {
11060bool AArch64InstrInfo::isWhileOpcode(
unsigned Opc)
const {
11065AArch64InstrInfo::getTailDuplicateSize(
CodeGenOptLevel OptLevel)
const {
11069bool AArch64InstrInfo::isLegalAddressingMode(
unsigned NumBytes, int64_t
Offset,
11070 unsigned Scale)
const {
11081 unsigned Shift =
Log2_64(NumBytes);
11082 if (NumBytes &&
Offset > 0 && (
Offset / NumBytes) <= (1LL << 12) - 1 &&
11090 return Scale == 1 || (Scale > 0 && Scale == NumBytes);
11095 return AArch64::BLRNoIP;
11097 return AArch64::BLR;
11102 Register TargetReg,
bool FrameSetup)
const {
11103 assert(TargetReg != AArch64::SP &&
"New top of stack cannot already be in SP");
11115 MF.
insert(MBBInsertPoint, LoopTestMBB);
11118 MF.
insert(MBBInsertPoint, LoopBodyMBB);
11120 MF.
insert(MBBInsertPoint, ExitMBB);
11130 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::SUBSXrx64),
11138 BuildMI(*LoopTestMBB, LoopTestMBB->
end(),
DL,
TII->get(AArch64::Bcc))
11144 BuildMI(*LoopBodyMBB, LoopBodyMBB->
end(),
DL,
TII->get(AArch64::STRXui))
11157 BuildMI(*ExitMBB, ExitMBB->
end(),
DL,
TII->get(AArch64::ADDXri), AArch64::SP)
11176 MBB.addSuccessor(LoopTestMBB);
11182 return ExitMBB->
begin();
11199 unsigned CompCounterOprNum;
11203 unsigned UpdateCounterOprNum;
11207 bool IsUpdatePriorComp;
11219 TII(MF->getSubtarget().getInstrInfo()),
11220 TRI(MF->getSubtarget().getRegisterInfo()),
MRI(MF->getRegInfo()),
11221 LoopBB(LoopBB), CondBranch(CondBranch), Comp(Comp),
11222 CompCounterOprNum(CompCounterOprNum), Update(Update),
11223 UpdateCounterOprNum(UpdateCounterOprNum),
Init(
Init),
11224 IsUpdatePriorComp(IsUpdatePriorComp),
Cond(
Cond.begin(),
Cond.end()) {}
11226 bool shouldIgnoreForPipelining(
const MachineInstr *
MI)
const override {
11232 std::optional<bool> createTripCountGreaterCondition(
11233 int TC, MachineBasicBlock &
MBB,
11234 SmallVectorImpl<MachineOperand> &CondParam)
override {
11242 void createRemainingIterationsGreaterCondition(
11243 int TC, MachineBasicBlock &
MBB, SmallVectorImpl<MachineOperand> &
Cond,
11244 DenseMap<MachineInstr *, MachineInstr *> &LastStage0Insts)
override;
11246 void setPreheader(MachineBasicBlock *NewPreheader)
override {}
11248 void adjustTripCount(
int TripCountAdjust)
override {}
11250 bool isMVEExpanderSupported()
override {
return true; }
11266 Result =
MRI.createVirtualRegister(
11269 }
else if (
I == ReplaceOprNum) {
11270 MRI.constrainRegClass(ReplaceReg,
TII->getRegClass(NewMI->
getDesc(),
I));
11274 MBB.insert(InsertTo, NewMI);
11278void AArch64PipelinerLoopInfo::createRemainingIterationsGreaterCondition(
11294 assert(CondBranch->getOpcode() == AArch64::Bcc);
11298 if (CondBranch->getOperand(1).getMBB() == LoopBB)
11305 auto AccumulateCond = [&](
Register CurCond,
11307 Register NewCond =
MRI.createVirtualRegister(&AArch64::GPR64commonRegClass);
11316 if (!LastStage0Insts.
empty() && LastStage0Insts[Comp]->getParent() == &
MBB) {
11320 for (
int I = 0;
I <= TC; ++
I) {
11326 AccCond = AccumulateCond(AccCond, CC);
11330 if (Update != Comp && IsUpdatePriorComp) {
11332 LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11333 NextCounter =
cloneInstr(Update, UpdateCounterOprNum, Counter,
MBB,
11337 NextCounter = LastStage0Insts[Update]->getOperand(0).getReg();
11339 }
else if (Update != Comp) {
11344 Counter = NextCounter;
11348 if (LastStage0Insts.
empty()) {
11352 if (IsUpdatePriorComp)
11357 Counter = LastStage0Insts[Comp]->getOperand(CompCounterOprNum).getReg();
11360 for (
int I = 0;
I <= TC; ++
I) {
11364 AccCond = AccumulateCond(AccCond, CC);
11365 if (
I != TC && Update != Comp)
11368 Counter = NextCounter;
11384 assert(Phi.getNumOperands() == 5);
11385 if (Phi.getOperand(2).getMBB() ==
MBB) {
11386 RegMBB = Phi.getOperand(1).getReg();
11387 RegOther = Phi.getOperand(3).getReg();
11389 assert(Phi.getOperand(4).getMBB() ==
MBB);
11390 RegMBB = Phi.getOperand(3).getReg();
11391 RegOther = Phi.getOperand(1).getReg();
11396 if (!
Reg.isVirtual())
11399 return MRI.getVRegDef(
Reg)->getParent() != BB;
11405 unsigned &UpdateCounterOprNum,
Register &InitReg,
11406 bool &IsUpdatePriorComp) {
11420 if (!
Reg.isVirtual())
11423 UpdateInst =
nullptr;
11424 UpdateCounterOprNum = 0;
11426 IsUpdatePriorComp =
true;
11430 if (Def->getParent() != LoopBB)
11432 if (Def->isCopy()) {
11434 if (Def->getOperand(0).getSubReg() || Def->getOperand(1).getSubReg())
11436 CurReg = Def->getOperand(1).getReg();
11437 }
else if (Def->isPHI()) {
11441 IsUpdatePriorComp =
false;
11446 switch (Def->getOpcode()) {
11447 case AArch64::ADDSXri:
11448 case AArch64::ADDSWri:
11449 case AArch64::SUBSXri:
11450 case AArch64::SUBSWri:
11451 case AArch64::ADDXri:
11452 case AArch64::ADDWri:
11453 case AArch64::SUBXri:
11454 case AArch64::SUBWri:
11456 UpdateCounterOprNum = 1;
11458 case AArch64::ADDSXrr:
11459 case AArch64::ADDSWrr:
11460 case AArch64::SUBSXrr:
11461 case AArch64::SUBSWrr:
11462 case AArch64::ADDXrr:
11463 case AArch64::ADDWrr:
11464 case AArch64::SUBXrr:
11465 case AArch64::SUBWrr:
11468 UpdateCounterOprNum = 1;
11470 UpdateCounterOprNum = 2;
11477 CurReg = Def->getOperand(UpdateCounterOprNum).getReg();
11492std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
11503 if (
MI.isCall() ||
MI.hasUnmodeledSideEffects())
11514 if (
TBB == LoopBB && FBB == LoopBB)
11518 if (
TBB != LoopBB && FBB ==
nullptr)
11521 assert((
TBB == LoopBB || FBB == LoopBB) &&
11522 "The Loop must be a single-basic-block loop");
11527 if (CondBranch->
getOpcode() != AArch64::Bcc)
11535 unsigned CompCounterOprNum = 0;
11537 if (
MI.modifiesRegister(AArch64::NZCV, &
TRI)) {
11541 switch (
MI.getOpcode()) {
11542 case AArch64::SUBSXri:
11543 case AArch64::SUBSWri:
11544 case AArch64::ADDSXri:
11545 case AArch64::ADDSWri:
11547 CompCounterOprNum = 1;
11549 case AArch64::ADDSWrr:
11550 case AArch64::ADDSXrr:
11551 case AArch64::SUBSWrr:
11552 case AArch64::SUBSXrr:
11556 if (isWhileOpcode(
MI.getOpcode())) {
11563 if (CompCounterOprNum == 0) {
11565 CompCounterOprNum = 2;
11567 CompCounterOprNum = 1;
11579 bool IsUpdatePriorComp;
11580 unsigned UpdateCounterOprNum;
11582 Update, UpdateCounterOprNum,
Init, IsUpdatePriorComp))
11585 return std::make_unique<AArch64PipelinerLoopInfo>(
11586 LoopBB, CondBranch, Comp, CompCounterOprNum, Update, UpdateCounterOprNum,
11596 TypeSize Scale(0U,
false), Width(0U,
false);
11597 int64_t MinOffset, MaxOffset;
11598 if (
getMemOpInfo(
MI.getOpcode(), Scale, Width, MinOffset, MaxOffset)) {
11600 if (
MI.getOperand(ImmIdx).isImm() && !
MI.getOperand(ImmIdx - 1).isFI()) {
11601 int64_t Imm =
MI.getOperand(ImmIdx).getImm();
11602 if (Imm < MinOffset || Imm > MaxOffset) {
11603 ErrInfo =
"Unexpected immediate on load/store instruction";
11609 const MCInstrDesc &MCID =
MI.getDesc();
11611 const MachineOperand &MO =
MI.getOperand(
Op);
11615 ErrInfo =
"OPERAND_IMPLICIT_IMM_0 should be 0";
11624 ErrInfo =
"OPERAND_SHIFT_MSL should be msl shift of 8 or 16";
11635#define GET_INSTRINFO_HELPERS
11636#define GET_INSTRMAP_INFO
11637#include "AArch64GenInstrInfo.inc"
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder & UseMI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static cl::opt< unsigned > BCCDisplacementBits("aarch64-bcc-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of Bcc instructions (DEBUG)"))
static Register genNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned MnegOpc, const TargetRegisterClass *RC)
genNeg - Helper to generate an intermediate negation of the second operand of Root
static bool isFrameStoreOpcode(int Opcode)
static cl::opt< unsigned > GatherOptSearchLimit("aarch64-search-limit", cl::Hidden, cl::init(2048), cl::desc("Restrict range of instructions to search for the " "machine-combiner gather pattern optimization"))
static bool getMaddPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find instructions that can be turned into madd.
static AArch64CC::CondCode findCondCodeUsedByInstr(const MachineInstr &Instr)
Find a condition code used by the instruction.
static MachineInstr * genFusedMultiplyAcc(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyAcc - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genFusedMultiplyAccNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate64(unsigned Opc)
static bool isFrameLoadOpcode(int Opcode)
static unsigned removeCopies(const MachineRegisterInfo &MRI, unsigned VReg)
static bool areCFlagsAccessedBetweenInstrs(MachineBasicBlock::iterator From, MachineBasicBlock::iterator To, const TargetRegisterInfo *TRI, const AccessKind AccessToCheck=AK_All)
True when condition flags are accessed (either by writing or reading) on the instruction trace starti...
static bool getFMAPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Floating-Point Support.
static bool isADDSRegImm(unsigned Opcode)
static bool isCheapCopy(const MachineInstr &MI, const AArch64RegisterInfo &RI)
static bool isANDOpcode(MachineInstr &MI)
static void appendOffsetComment(int NumBytes, llvm::raw_string_ostream &Comment, StringRef RegScale={})
static unsigned sForm(MachineInstr &Instr)
Get opcode of S version of Instr.
static bool isCombineInstrSettingFlag(unsigned Opc)
static bool getFNEGPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getIndVarInfo(Register Reg, const MachineBasicBlock *LoopBB, MachineInstr *&UpdateInst, unsigned &UpdateCounterOprNum, Register &InitReg, bool &IsUpdatePriorComp)
If Reg is an induction variable, return true and set some parameters.
static const MachineInstrBuilder & AddSubReg(const MachineInstrBuilder &MIB, MCRegister Reg, unsigned SubIdx, unsigned State, const TargetRegisterInfo *TRI)
static bool canPairLdStOpc(unsigned FirstOpc, unsigned SecondOpc)
static int findCondCodeUseOperandIdxForBranchOrSelect(const MachineInstr &Instr)
static bool isPostIndexLdStOpcode(unsigned Opcode)
Return true if the opcode is a post-index ld/st instruction, which really loads from base+0.
static unsigned getBranchDisplacementBits(unsigned Opc)
static cl::opt< unsigned > CBDisplacementBits("aarch64-cb-offset-bits", cl::Hidden, cl::init(9), cl::desc("Restrict range of CB instructions (DEBUG)"))
static std::optional< ParamLoadedValue > describeORRLoadedValue(const MachineInstr &MI, Register DescribedReg, const TargetInstrInfo *TII, const TargetRegisterInfo *TRI)
If the given ORR instruction is a copy, and DescribedReg overlaps with the destination register then,...
static bool getFMULPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static void appendReadRegExpr(SmallVectorImpl< char > &Expr, unsigned RegNum)
static MachineInstr * genMaddR(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, unsigned VR, const TargetRegisterClass *RC)
genMaddR - Generate madd instruction and combine mul and add using an extra virtual register Example ...
static Register cloneInstr(const MachineInstr *MI, unsigned ReplaceOprNum, Register ReplaceReg, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertTo)
Clone an instruction from MI.
static bool scaleOffset(unsigned Opc, int64_t &Offset)
static bool canCombineWithFMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc)
unsigned scaledOffsetOpcode(unsigned Opcode, unsigned &Scale)
static MachineInstr * genFusedMultiplyIdx(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC)
genFusedMultiplyIdx - Helper to generate fused multiply accumulate instructions.
static MachineInstr * genIndexedMultiply(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxDupOp, unsigned MulOpc, const TargetRegisterClass *RC, MachineRegisterInfo &MRI)
Fold (FMUL x (DUP y lane)) into (FMUL_indexed x y lane)
static bool isSUBSRegImm(unsigned Opcode)
static bool UpdateOperandRegClass(MachineInstr &Instr)
static const TargetRegisterClass * getRegClass(const MachineInstr &MI, Register Reg)
static bool canCmpInstrBeRemoved(MachineInstr &MI, MachineInstr &CmpInstr, int CmpValue, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > &CCUseInstrs, bool &IsInvertCC)
unsigned unscaledOffsetOpcode(unsigned Opcode)
static bool getLoadPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Search for patterns of LD instructions we can optimize.
static bool canInstrSubstituteCmpInstr(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI)
Check if CmpInstr can be substituted by MI.
static UsedNZCV getUsedNZCV(AArch64CC::CondCode CC)
static bool isCombineInstrCandidateFP(const MachineInstr &Inst)
static void appendLoadRegExpr(SmallVectorImpl< char > &Expr, int64_t OffsetFromDefCFA)
static void appendConstantExpr(SmallVectorImpl< char > &Expr, int64_t Constant, dwarf::LocationAtom Operation)
static unsigned convertToNonFlagSettingOpc(const MachineInstr &MI)
Return the opcode that does not set flags when possible - otherwise return the original opcode.
static bool outliningCandidatesV8_3OpsConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool isCombineInstrCandidate32(unsigned Opc)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
static unsigned offsetExtendOpcode(unsigned Opcode)
static void loadRegPairFromStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register DestReg, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
static void generateGatherLanePattern(MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned Pattern, unsigned NumLanes)
Generate optimized instruction sequence for gather load patterns to improve Memory-Level Parallelism ...
static bool getMiscPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
Find other MI combine patterns.
static bool outliningCandidatesSigningKeyConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool outliningCandidatesSigningScopeConsensus(const outliner::Candidate &a, const outliner::Candidate &b)
static bool shouldClusterFI(const MachineFrameInfo &MFI, int FI1, int64_t Offset1, unsigned Opcode1, int FI2, int64_t Offset2, unsigned Opcode2)
static cl::opt< unsigned > TBZDisplacementBits("aarch64-tbz-offset-bits", cl::Hidden, cl::init(14), cl::desc("Restrict range of TB[N]Z instructions (DEBUG)"))
static void extractPhiReg(const MachineInstr &Phi, const MachineBasicBlock *MBB, Register &RegMBB, Register &RegOther)
static MCCFIInstruction createDefCFAExpression(const TargetRegisterInfo &TRI, unsigned Reg, const StackOffset &Offset)
static bool isDefinedOutside(Register Reg, const MachineBasicBlock *BB)
static MachineInstr * genFusedMultiply(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, unsigned IdxMulOpd, unsigned MaddOpc, const TargetRegisterClass *RC, FMAInstKind kind=FMAInstKind::Default, const Register *ReplacedAddend=nullptr)
genFusedMultiply - Generate fused multiply instructions.
static bool getGatherLanePattern(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, unsigned LoadLaneOpCode, unsigned NumLanes)
Check if the given instruction forms a gather load pattern that can be optimized for better Memory-Le...
static MachineInstr * genFusedMultiplyIdxNeg(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg, unsigned IdxMulOpd, unsigned MaddOpc, unsigned MnegOpc, const TargetRegisterClass *RC)
genFusedMultiplyAccNeg - Helper to generate fused multiply accumulate instructions with an additional...
static bool isCombineInstrCandidate(unsigned Opc)
static unsigned regOffsetOpcode(unsigned Opcode)
MachineOutlinerClass
Constants defining how certain sequences should be outlined.
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerRegSave
Emit a call and tail-call.
@ MachineOutlinerNoLRSave
Only emit a branch.
@ MachineOutlinerThunk
Emit a call and return.
static cl::opt< unsigned > BDisplacementBits("aarch64-b-offset-bits", cl::Hidden, cl::init(26), cl::desc("Restrict range of B instructions (DEBUG)"))
static bool areCFlagsAliveInSuccessors(const MachineBasicBlock *MBB)
Check if AArch64::NZCV should be alive in successors of MBB.
static void emitFrameOffsetAdj(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, int64_t Offset, unsigned Opc, const TargetInstrInfo *TII, MachineInstr::MIFlag Flag, bool NeedsWinCFI, bool *HasWinCFI, bool EmitCFAOffset, StackOffset CFAOffset, unsigned FrameReg)
static bool isCheapImmediate(const MachineInstr &MI, unsigned BitSize)
static cl::opt< unsigned > CBZDisplacementBits("aarch64-cbz-offset-bits", cl::Hidden, cl::init(19), cl::desc("Restrict range of CB[N]Z instructions (DEBUG)"))
static void genSubAdd2SubSub(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, unsigned IdxOpd1, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
Do the following transformation A - (B + C) ==> (A - B) - C A - (B + C) ==> (A - C) - B.
static unsigned canFoldIntoCSel(const MachineRegisterInfo &MRI, unsigned VReg, unsigned *NewReg=nullptr)
static void signOutlinedFunction(MachineFunction &MF, MachineBasicBlock &MBB, const AArch64InstrInfo *TII, bool ShouldSignReturnAddr)
static MachineInstr * genFNegatedMAD(MachineFunction &MF, MachineRegisterInfo &MRI, const TargetInstrInfo *TII, MachineInstr &Root, SmallVectorImpl< MachineInstr * > &InsInstrs)
static bool canCombineWithMUL(MachineBasicBlock &MBB, MachineOperand &MO, unsigned MulOpc, unsigned ZeroReg)
static void storeRegPairToStackSlot(const TargetRegisterInfo &TRI, MachineBasicBlock &MBB, MachineBasicBlock::iterator InsertBefore, const MCInstrDesc &MCID, Register SrcReg, bool IsKill, unsigned SubIdx0, unsigned SubIdx1, int FI, MachineMemOperand *MMO)
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
static const Function * getParent(const Value *V)
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
DXIL Forward Handle Accesses
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
This file implements the LivePhysRegs utility for tracking liveness of physical registers.
Machine Check Debug Module
Register const TargetRegisterInfo * TRI
Promote Memory to Register
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
MachineInstr unsigned OpIdx
ConstantRange Range(APInt(BitWidth, Low), APInt(BitWidth, High))
PowerPC Reduce CR logical Operation
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
This file defines the SmallSet class.
This file defines the SmallVector class.
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
AArch64FunctionInfo - This class is derived from MachineFunctionInfo and contains private AArch64-spe...
SignReturnAddress getSignReturnAddressCondition() const
bool hasStreamingModeChanges() const
void setOutliningStyle(const std::string &Style)
std::optional< bool > hasRedZone() const
static bool shouldSignReturnAddress(SignReturnAddress Condition, bool IsLRSpilled)
bool shouldSignWithBKey() const
static bool isHForm(const MachineInstr &MI)
Returns whether the instruction is in H form (16 bit operands)
void insertSelect(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI, const DebugLoc &DL, Register DstReg, ArrayRef< MachineOperand > Cond, Register TrueReg, Register FalseReg) const override
static bool hasBTISemantics(const MachineInstr &MI)
Returns whether the instruction can be compatible with non-zero BTYPE.
static bool isQForm(const MachineInstr &MI)
Returns whether the instruction is in Q form (128 bit operands)
static bool getMemOpInfo(unsigned Opcode, TypeSize &Scale, TypeSize &Width, int64_t &MinOffset, int64_t &MaxOffset)
Returns true if opcode Opc is a memory operation.
static bool isTailCallReturnInst(const MachineInstr &MI)
Returns true if MI is one of the TCRETURN* instructions.
static bool isFPRCopy(const MachineInstr &MI)
Does this instruction rename an FPR without modifying bits?
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
If the specific machine instruction is an instruction that moves/copies value from one register to an...
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
GetInstSize - Return the number of bytes of code the specified instruction may be.
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, Register DestReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
static bool isGPRCopy(const MachineInstr &MI)
Does this instruction rename a GPR without modifying bits?
static unsigned convertToFlagSettingOpc(unsigned Opc)
Return the opcode that set flags when possible.
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
bool canInsertSelect(const MachineBasicBlock &, ArrayRef< MachineOperand > Cond, Register, Register, Register, int &, int &, int &) const override
Register isLoadFromStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
static const MachineOperand & getLdStOffsetOp(const MachineInstr &MI)
Returns the immediate offset operator of a load/store.
bool isCoalescableExtInstr(const MachineInstr &MI, Register &SrcReg, Register &DstReg, unsigned &SubIdx) const override
static std::optional< unsigned > getUnscaledLdSt(unsigned Opc)
Returns the unscaled load/store for the scaled load/store opcode, if there is a corresponding unscale...
static bool hasUnscaledLdStOffset(unsigned Opc)
Return true if it has an unscaled load/store offset.
static const MachineOperand & getLdStAmountOp(const MachineInstr &MI)
Returns the shift amount operator of a load/store.
static bool isPreLdSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load/store.
std::optional< ExtAddrMode > getAddrModeFromMemoryOp(const MachineInstr &MemI, const TargetRegisterInfo *TRI) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
bool analyzeBranchPredicate(MachineBasicBlock &MBB, MachineBranchPredicate &MBP, bool AllowModify) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DestReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
static bool isPairableLdStInst(const MachineInstr &MI)
Return true if pairing the given load or store may be paired with another.
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool isKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
const AArch64RegisterInfo & getRegisterInfo() const
getRegisterInfo - TargetInstrInfo is a superset of MRegister info.
static bool isPreSt(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed store.
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
void insertNoop(MachineBasicBlock &MBB, MachineBasicBlock::iterator MI) const override
AArch64InstrInfo(const AArch64Subtarget &STI)
static bool isPairedLdSt(const MachineInstr &MI)
Returns whether the instruction is a paired load/store.
bool getMemOperandWithOffsetWidth(const MachineInstr &MI, const MachineOperand *&BaseOp, int64_t &Offset, bool &OffsetIsScalable, TypeSize &Width, const TargetRegisterInfo *TRI) const
If OffsetIsScalable is set to 'true', the offset is scaled by vscale.
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
static bool isStridedAccess(const MachineInstr &MI)
Return true if the given load or store is a strided memory access.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
Detect opportunities for ldp/stp formation.
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
bool isThroughputPattern(unsigned Pattern) const override
Return true when a code sequence can improve throughput.
MachineOperand & getMemOpBaseRegImmOfsOffsetOperand(MachineInstr &LdSt) const
Return the immediate offset of the base register in a load/store LdSt.
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify=false) const override
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
static bool isLdStPairSuppressed(const MachineInstr &MI)
Return true if pairing the given load or store is hinted to be unprofitable.
Register isStoreToStackSlotPostFE(const MachineInstr &MI, int &FrameIndex) const override
Check for post-frame ptr elimination stack locations as well.
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &DL, int *BytesAdded=nullptr) const override
MachineBasicBlock::iterator probedStackAlloc(MachineBasicBlock::iterator MBBI, Register TargetReg, bool FrameSetup) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
bool optimizeCompareInstr(MachineInstr &CmpInstr, Register SrcReg, Register SrcReg2, int64_t CmpMask, int64_t CmpValue, const MachineRegisterInfo *MRI) const override
optimizeCompareInstr - Convert the instruction supplying the argument to the comparison into one that...
static unsigned getLoadStoreImmIdx(unsigned Opc)
Returns the index for the immediate for a given instruction.
static bool isGPRZero(const MachineInstr &MI)
Does this instruction set its full destination register to zero?
void copyGPRRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, unsigned ZeroReg, llvm::ArrayRef< unsigned > Indices) const
bool analyzeCompare(const MachineInstr &MI, Register &SrcReg, Register &SrcReg2, int64_t &CmpMask, int64_t &CmpValue) const override
analyzeCompare - For a comparison instruction, return the source registers in SrcReg and SrcReg2,...
CombinerObjective getCombinerObjective(unsigned Pattern) const override
static bool isFpOrNEON(Register Reg)
Returns whether the physical register is FP or NEON.
bool isAsCheapAsAMove(const MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyLikeInstrImpl(const MachineInstr &MI) const override
static void suppressLdStPair(MachineInstr &MI)
Hint that pairing the given load or store is unprofitable.
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
static bool isPreLd(const MachineInstr &MI)
Returns whether the instruction is a pre-indexed load.
void copyPhysRegTuple(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc, unsigned Opcode, llvm::ArrayRef< unsigned > Indices) const
bool optimizeCondBranch(MachineInstr &MI) const override
Replace csincr-branch sequence by simple conditional branch.
static int getMemScale(unsigned Opc)
Scaling factor for (scaled or unscaled) load or store.
bool isCandidateToMergeOrPair(const MachineInstr &MI) const
Return true if this is a load/store that can be potentially paired/merged.
MCInst getNop() const override
static const MachineOperand & getLdStBaseOp(const MachineInstr &MI)
Returns the base register operator of a load/store.
bool isReservedReg(const MachineFunction &MF, MCRegister Reg) const
const AArch64RegisterInfo * getRegisterInfo() const override
bool isNeonAvailable() const
Returns true if the target has NEON and the function at runtime is known to have NEON enabled (e....
bool isSVEorStreamingSVEAvailable() const
Returns true if the target has access to either the full range of SVE instructions,...
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
const T & front() const
front - Get the first element.
size_t size() const
size - Get the array size.
This is an important base class in LLVM.
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
bool hasOptSize() const
Optimize this function for size (-Os) or minimum size (-Oz).
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI InstListType::iterator eraseFromParent()
This method unlinks 'this' from the containing basic block and deletes it.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
static LocationSize precise(uint64_t Value)
This class is intended to be used as a base class for asm properties and features specific to the tar...
bool usesWindowsCFI() const
static MCCFIInstruction cfiDefCfa(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa defines a rule for computing CFA as: take address from Register and add Offset to it.
static MCCFIInstruction createOffset(MCSymbol *L, unsigned Register, int64_t Offset, SMLoc Loc={})
.cfi_offset Previous value of Register is saved at offset Offset from CFA.
static MCCFIInstruction cfiDefCfaOffset(MCSymbol *L, int64_t Offset, SMLoc Loc={})
.cfi_def_cfa_offset modifies a rule for computing CFA.
static MCCFIInstruction createEscape(MCSymbol *L, StringRef Vals, SMLoc Loc={}, StringRef Comment="")
.cfi_escape Allows the user to add arbitrary bytes to the unwind info.
MCInstBuilder & addImm(int64_t Val)
Add a new integer immediate operand.
Instances of this class represent a single low-level machine instruction.
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
ArrayRef< MCOperandInfo > operands() const
Wrapper class representing physical registers. Should be passed by value.
constexpr bool isValid() const
static constexpr unsigned NoRegister
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
bool isInlineAsmBrIndirectTarget() const
Returns true if this is the indirect dest of an INLINEASM_BR.
unsigned pred_size() const
LLVM_ABI void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
reverse_instr_iterator instr_rbegin()
LLVM_ABI iterator getFirstTerminator()
Returns an iterator to the first terminator instruction of this basic block.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
reverse_instr_iterator instr_rend()
Instructions::iterator instr_iterator
instr_iterator instr_end()
Instructions::const_iterator const_instr_iterator
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
LLVM_ABI instr_iterator erase(instr_iterator I)
Remove an instruction from the instruction list and delete it.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
void setMachineBlockAddressTaken()
Set this block to indicate that its address is used as something other than the target of a terminato...
LLVM_ABI bool isLiveIn(MCRegister Reg, LaneBitmask LaneMask=LaneBitmask::getAll()) const
Return true if the specified register is in the live in set.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
uint64_t getStackSize() const
Return the number of bytes that must be allocated to hold all of the fixed size frame objects.
void setStackID(int ObjectIdx, uint8_t ID)
bool isCalleeSavedInfoValid() const
Has the callee saved info been calculated yet?
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
unsigned getNumObjects() const
Return the number of objects.
int64_t getObjectOffset(int ObjectIdx) const
Return the assigned stack offset of the specified object from the incoming stack pointer.
bool isFixedObjectIndex(int ObjectIdx) const
Returns true if the specified index corresponds to a fixed stack object.
unsigned addFrameInst(const MCCFIInstruction &Inst)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineJumpTableInfo * getJumpTableInfo() const
getJumpTableInfo - Return the jump table info object for the current function.
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & addCFIIndex(unsigned CFIIndex) const
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
const MachineInstrBuilder & addDef(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register definition operand.
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool isCall(QueryType Type=AnyInBundle) const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI uint32_t mergeFlagsWith(const MachineInstr &Other) const
Return the MIFlags which represent both MachineInstrs.
unsigned getNumOperands() const
Retuns the total number of operands.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool registerDefIsDead(Register Reg, const TargetRegisterInfo *TRI) const
Returns true if the register is dead in this machine instruction.
bool definesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr fully defines the specified register.
LLVM_ABI void setDesc(const MCInstrDesc &TID)
Replace the instruction descriptor (thus opcode) of the current instruction with a new one.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
LLVM_ABI bool isLoadFoldBarrier() const
Returns true if it is illegal to fold a load across this instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
LLVM_ABI void eraseFromParent()
Unlink 'this' from the containing basic block and delete it.
LLVM_ABI void removeOperand(unsigned OpNo)
Erase an operand from an instruction, leaving it with one fewer operand than it started with.
LLVM_ABI void addRegisterDefined(Register Reg, const TargetRegisterInfo *RegInfo=nullptr)
We have determined MI defines a register.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI int findRegisterDefOperandIdx(Register Reg, const TargetRegisterInfo *TRI, bool isDead=false, bool Overlap=false) const
Returns the operand index that is a def of the specified register or -1 if it is not found.
const std::vector< MachineJumpTableEntry > & getJumpTables() const
A description of a memory reference used in the backend.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
LLVM_ABI MachineFunction * getMachineFunction(const Function &F) const
Returns the MachineFunction associated to IR function F if there is one, otherwise nullptr.
MachineOperand class - Representation of each machine instruction operand.
void setSubReg(unsigned subReg)
unsigned getSubReg() const
void setImm(int64_t immVal)
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
void setIsDead(bool Val=true)
LLVM_ABI void setReg(Register Reg)
Change the register this operand corresponds to.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
void setIsKill(bool Val=true)
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
unsigned getTargetFlags() const
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
void setIsUndef(bool Val=true)
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
bool tracksLiveness() const
tracksLiveness - Returns true when tracking register liveness accurately.
bool reservedRegsFrozen() const
reservedRegsFrozen - Returns true after freezeReservedRegs() was called to ensure the set of reserved...
const TargetRegisterClass * getRegClassOrNull(Register Reg) const
Return the register class of Reg, or null if Reg has not been assigned a register class yet.
LLVM_ABI const TargetRegisterClass * constrainRegClass(Register Reg, const TargetRegisterClass *RC, unsigned MinNumRegs=0)
constrainRegClass - Constrain the register class of the specified virtual register to be a common sub...
LLVM_ABI MachineInstr * getUniqueVRegDef(Register Reg) const
getUniqueVRegDef - Return the unique machine instr that defines the specified virtual register or nul...
A Module instance is used to store all the information related to an LLVM module.
MI-level patchpoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
Wrapper class representing virtual and physical registers.
MCRegister asMCReg() const
Utility to check-convert this value to a MCRegister.
constexpr bool isValid() const
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
static constexpr bool isVirtualRegister(unsigned Reg)
Return true if the specified register number is in the virtual register namespace.
constexpr bool isPhysical() const
Return true if the specified register number is in the physical register namespace.
Represents a location in source code.
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
SmallPtrSet - This class implements a set which is optimized for holding SmallSize or less elements.
SmallSet - This maintains a set of unique values, optimizing for the case when the set is small (less...
SmallString - A SmallString is just a SmallVector with methods and accessors that make it work better...
void append(StringRef RHS)
Append from a StringRef.
StringRef str() const
Explicit conversion to StringRef.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
StackOffset holds a fixed and a scalable offset in bytes.
int64_t getFixed() const
Returns the fixed component of the stack.
int64_t getScalable() const
Returns the scalable component of the stack.
static StackOffset get(int64_t Fixed, int64_t Scalable)
static StackOffset getScalable(int64_t Scalable)
static StackOffset getFixed(int64_t Fixed)
MI-level Statepoint operands.
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
StringRef - Represent a constant reference to a string, i.e.
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual std::optional< ParamLoadedValue > describeLoadedValue(const MachineInstr &MI, Register Reg) const
Produce the expression describing the MI loading a value into the physical register Reg.
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isSchedulingBoundary(const MachineInstr &MI, const MachineBasicBlock *MBB, const MachineFunction &MF) const
Test if the given instruction should be considered a scheduling boundary.
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual bool isFunctionSafeToSplit(const MachineFunction &MF) const
Return true if the function is a viable candidate for machine function splitting.
CodeModel::Model getCodeModel() const
Returns the code model.
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
bool contains(Register Reg) const
Return true if the specified register is included in this register class.
bool hasSubClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a sub-class of or equal to this class.
bool hasSuperClassEq(const TargetRegisterClass *RC) const
Returns true if RC is a super-class of or equal to this class.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Value * getOperand(unsigned i) const
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
self_iterator getIterator()
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
static CondCode getInvertedCondCode(CondCode Code)
@ MO_DLLIMPORT
MO_DLLIMPORT - On a symbol operand, this represents that the reference to the symbol is for an import...
@ MO_NC
MO_NC - Indicates whether the linker is expected to check the symbol reference for overflow.
@ MO_G1
MO_G1 - A symbol operand with this flag (granule 1) represents the bits 16-31 of a 64-bit address,...
@ MO_S
MO_S - Indicates that the bits of the symbol operand represented by MO_G0 etc are signed.
@ MO_PAGEOFF
MO_PAGEOFF - A symbol operand with this flag represents the offset of that symbol within a 4K page.
@ MO_GOT
MO_GOT - This flag indicates that a symbol operand represents the address of the GOT entry for the sy...
@ MO_PREL
MO_PREL - Indicates that the bits of the symbol operand represented by MO_G0 etc are PC relative.
@ MO_G0
MO_G0 - A symbol operand with this flag (granule 0) represents the bits 0-15 of a 64-bit address,...
@ MO_ARM64EC_CALLMANGLE
MO_ARM64EC_CALLMANGLE - Operand refers to the Arm64EC-mangled version of a symbol,...
@ MO_PAGE
MO_PAGE - A symbol operand with this flag represents the pc-relative offset of the 4K page containing...
@ MO_HI12
MO_HI12 - This flag indicates that a symbol operand represents the bits 13-24 of a 64-bit address,...
@ MO_TLS
MO_TLS - Indicates that the operand being accessed is some kind of thread-local symbol.
@ MO_G2
MO_G2 - A symbol operand with this flag (granule 2) represents the bits 32-47 of a 64-bit address,...
@ MO_TAGGED
MO_TAGGED - With MO_PAGE, indicates that the page includes a memory tag in bits 56-63.
@ MO_G3
MO_G3 - A symbol operand with this flag (granule 3) represents the high 16-bits of a 64-bit address,...
@ MO_COFFSTUB
MO_COFFSTUB - On a symbol operand "FOO", this indicates that the reference is actually to the "....
unsigned getCheckerSizeInBytes(AuthCheckMethod Method)
Returns the number of bytes added by checkAuthenticatedRegister.
static uint64_t decodeLogicalImmediate(uint64_t val, unsigned regSize)
decodeLogicalImmediate - Decode a logical immediate value in the form "N:immr:imms" (where the immr a...
static unsigned getShiftValue(unsigned Imm)
getShiftValue - Extract the shift value.
static unsigned getArithExtendImm(AArch64_AM::ShiftExtendType ET, unsigned Imm)
getArithExtendImm - Encode the extend type and shift amount for an arithmetic instruction: imm: 3-bit...
static unsigned getArithShiftValue(unsigned Imm)
getArithShiftValue - get the arithmetic shift value.
static uint64_t encodeLogicalImmediate(uint64_t imm, unsigned regSize)
encodeLogicalImmediate - Return the encoded immediate value for a logical immediate instruction of th...
static AArch64_AM::ShiftExtendType getExtendType(unsigned Imm)
getExtendType - Extract the extend type for operands of arithmetic ops.
static AArch64_AM::ShiftExtendType getArithExtendType(unsigned Imm)
static AArch64_AM::ShiftExtendType getShiftType(unsigned Imm)
getShiftType - Extract the shift type.
static unsigned getShifterImm(AArch64_AM::ShiftExtendType ST, unsigned Imm)
getShifterImm - Encode the shift type and amount: imm: 6-bit shift amount shifter: 000 ==> lsl 001 ==...
void expandMOVImm(uint64_t Imm, unsigned BitSize, SmallVectorImpl< ImmInsnModel > &Insn)
Expand a MOVi32imm or MOVi64imm pseudo instruction to one or more real move-immediate instructions to...
static const uint64_t InstrFlagIsWhile
static const uint64_t InstrFlagIsPTestLike
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
@ C
The default llvm calling convention, compatible with C.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Renamable
Register that may be renamed.
@ Define
Register definition.
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
@ ScalablePredicateVector
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
LLVM_ABI Instruction & back() const
This is an optimization pass for GlobalISel generic memory operations.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
static bool isCondBranchOpcode(int Opc)
MCCFIInstruction createDefCFA(const TargetRegisterInfo &TRI, unsigned FrameReg, unsigned Reg, const StackOffset &Offset, bool LastAdjustmentWasScalable=true)
static bool isPTrueOpcode(unsigned Opc)
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
bool succeeded(LogicalResult Result)
Utility function that returns true if the provided LogicalResult corresponds to a success value.
int isAArch64FrameOffsetLegal(const MachineInstr &MI, StackOffset &Offset, bool *OutUseUnscaledOp=nullptr, unsigned *OutUnscaledOp=nullptr, int64_t *EmittableOffset=nullptr)
Check if the Offset is a valid frame offset for MI.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
static bool isIndirectBranchOpcode(int Opc)
iterator_range< T > make_range(T x, T y)
Convenience function for iterating over sub-ranges.
unsigned getBLRCallOpcode(const MachineFunction &MF)
Return opcode to be used for indirect calls.
@ AArch64FrameOffsetIsLegal
Offset is legal.
@ AArch64FrameOffsetCanUpdate
Offset can apply, at least partly.
@ AArch64FrameOffsetCannotUpdate
Offset cannot apply.
constexpr bool isPowerOf2_64(uint64_t Value)
Return true if the argument is a power of two > 0 (64 bit edition.)
unsigned Log2_64(uint64_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
static bool isSEHInstruction(const MachineInstr &MI)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
void sort(IteratorTy Start, IteratorTy End)
AArch64MachineCombinerPattern
@ MULSUBv2i32_indexed_OP1
@ MULADDv4i16_indexed_OP2
@ MULSUBv8i16_indexed_OP2
@ MULSUBv4i16_indexed_OP2
@ MULSUBv4i32_indexed_OP2
@ MULADDv2i32_indexed_OP1
@ MULADDv4i32_indexed_OP1
@ MULADDv2i32_indexed_OP2
@ MULSUBv4i16_indexed_OP1
@ MULADDv4i32_indexed_OP2
@ MULSUBv8i16_indexed_OP1
@ MULSUBv2i32_indexed_OP2
@ MULADDv8i16_indexed_OP1
@ MULSUBv4i32_indexed_OP1
@ MULADDv8i16_indexed_OP2
@ MULADDv4i16_indexed_OP1
void emitFrameOffset(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, unsigned DestReg, unsigned SrcReg, StackOffset Offset, const TargetInstrInfo *TII, MachineInstr::MIFlag=MachineInstr::NoFlags, bool SetNZCV=false, bool NeedsWinCFI=false, bool *HasWinCFI=nullptr, bool EmitCFAOffset=false, StackOffset InitialOffset={}, unsigned FrameReg=AArch64::SP)
emitFrameOffset - Emit instructions as needed to set DestReg to SrcReg plus Offset.
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
std::optional< UsedNZCV > examineCFlagsUse(MachineInstr &MI, MachineInstr &CmpInstr, const TargetRegisterInfo &TRI, SmallVectorImpl< MachineInstr * > *CCUseInstrs=nullptr)
CodeGenOptLevel
Code generation optimization level.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
auto instructionsWithoutDebug(IterT It, IterT End, bool SkipPseudoOp=true)
Construct a range iterator which begins at It and moves forwards until End is reached,...
LLVM_ABI raw_fd_ostream & errs()
This returns a reference to a raw_ostream for standard error.
auto drop_end(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the last N elements excluded.
unsigned getUndefRegState(bool B)
static MCRegister getXRegFromWReg(MCRegister Reg)
unsigned getDefRegState(bool B)
MCCFIInstruction createCFAOffset(const TargetRegisterInfo &MRI, unsigned Reg, const StackOffset &OffsetFromDefCFA, std::optional< int64_t > IncomingVGOffsetFromDefCFA)
unsigned getKillRegState(bool B)
FunctionAddr VTableAddr Next
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
static bool isUncondBranchOpcode(int Opc)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
constexpr bool isIntN(unsigned N, int64_t x)
Checks if an signed integer fits into the given (dynamic) bit width.
bool rewriteAArch64FrameIndex(MachineInstr &MI, unsigned FrameRegIdx, unsigned FrameReg, StackOffset &Offset, const AArch64InstrInfo *TII)
rewriteAArch64FrameIndex - Rewrite MI to access 'Offset' bytes from the FP.
bool is_contained(R &&Range, const E &Element)
Returns true if Element is found in Range.
static const MachineMemOperand::Flags MOSuppressPair
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
void appendLEB128(SmallVectorImpl< U > &Buffer, T Value)
std::pair< MachineOperand, DIExpression * > ParamLoadedValue
bool isNZCVTouchedInInstructionRange(const MachineInstr &DefMI, const MachineInstr &UseMI, const TargetRegisterInfo *TRI)
Return true if there is an instruction /after/ DefMI and before UseMI which either reads or clobbers ...
auto seq(T Begin, T End)
Iterate over an integral type from Begin up to - but not including - End.
static const MachineMemOperand::Flags MOStridedAccess
void fullyRecomputeLiveIns(ArrayRef< MachineBasicBlock * > MBBs)
Convenience function for recomputing live-in's for a set of MBBs until the computation converges.
LLVM_ABI Printable printReg(Register Reg, const TargetRegisterInfo *TRI=nullptr, unsigned SubIdx=0, const MachineRegisterInfo *MRI=nullptr)
Prints virtual and physical registers with or without a TRI instance.
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
LLVM_ABI static const MBBSectionID ColdSectionID
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.