LLVM 22.0.0git
RISCVInstrInfo.cpp
Go to the documentation of this file.
1//===-- RISCVInstrInfo.cpp - RISC-V Instruction Information -----*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file contains the RISC-V implementation of the TargetInstrInfo class.
10//
11//===----------------------------------------------------------------------===//
12
13#include "RISCVInstrInfo.h"
16#include "RISCV.h"
18#include "RISCVSubtarget.h"
19#include "llvm/ADT/STLExtras.h"
21#include "llvm/ADT/Statistic.h"
33#include "llvm/IR/Module.h"
34#include "llvm/MC/MCDwarf.h"
38
39using namespace llvm;
40
41#define GEN_CHECK_COMPRESS_INSTR
42#include "RISCVGenCompressInstEmitter.inc"
43
44#define GET_INSTRINFO_CTOR_DTOR
45#define GET_INSTRINFO_NAMED_OPS
46#include "RISCVGenInstrInfo.inc"
47
48#define DEBUG_TYPE "riscv-instr-info"
49STATISTIC(NumVRegSpilled,
50 "Number of registers within vector register groups spilled");
51STATISTIC(NumVRegReloaded,
52 "Number of registers within vector register groups reloaded");
53
55 "riscv-prefer-whole-register-move", cl::init(false), cl::Hidden,
56 cl::desc("Prefer whole register move for vector registers."));
57
59 "riscv-force-machine-combiner-strategy", cl::Hidden,
60 cl::desc("Force machine combiner to use a specific strategy for machine "
61 "trace metrics evaluation."),
64 "Local strategy."),
66 "MinInstrCount strategy.")));
67
69
70using namespace RISCV;
71
72#define GET_RISCVVPseudosTable_IMPL
73#include "RISCVGenSearchableTables.inc"
74
75} // namespace llvm::RISCVVPseudosTable
76
77namespace llvm::RISCV {
78
79#define GET_RISCVMaskedPseudosTable_IMPL
80#include "RISCVGenSearchableTables.inc"
81
82} // end namespace llvm::RISCV
83
85 : RISCVGenInstrInfo(STI, RegInfo, RISCV::ADJCALLSTACKDOWN,
86 RISCV::ADJCALLSTACKUP),
87 RegInfo(STI.getHwMode()), STI(STI) {}
88
89#define GET_INSTRINFO_HELPERS
90#include "RISCVGenInstrInfo.inc"
91
93 if (STI.hasStdExtZca())
94 return MCInstBuilder(RISCV::C_NOP);
95 return MCInstBuilder(RISCV::ADDI)
96 .addReg(RISCV::X0)
97 .addReg(RISCV::X0)
98 .addImm(0);
99}
100
102 int &FrameIndex) const {
103 TypeSize Dummy = TypeSize::getZero();
104 return isLoadFromStackSlot(MI, FrameIndex, Dummy);
105}
106
107static std::optional<unsigned> getLMULForRVVWholeLoadStore(unsigned Opcode) {
108 switch (Opcode) {
109 default:
110 return std::nullopt;
111 case RISCV::VS1R_V:
112 case RISCV::VL1RE8_V:
113 case RISCV::VL1RE16_V:
114 case RISCV::VL1RE32_V:
115 case RISCV::VL1RE64_V:
116 return 1;
117 case RISCV::VS2R_V:
118 case RISCV::VL2RE8_V:
119 case RISCV::VL2RE16_V:
120 case RISCV::VL2RE32_V:
121 case RISCV::VL2RE64_V:
122 return 2;
123 case RISCV::VS4R_V:
124 case RISCV::VL4RE8_V:
125 case RISCV::VL4RE16_V:
126 case RISCV::VL4RE32_V:
127 case RISCV::VL4RE64_V:
128 return 4;
129 case RISCV::VS8R_V:
130 case RISCV::VL8RE8_V:
131 case RISCV::VL8RE16_V:
132 case RISCV::VL8RE32_V:
133 case RISCV::VL8RE64_V:
134 return 8;
135 }
136}
137
139 int &FrameIndex,
140 TypeSize &MemBytes) const {
141 switch (MI.getOpcode()) {
142 default:
143 return 0;
144 case RISCV::LB:
145 case RISCV::LBU:
146 MemBytes = TypeSize::getFixed(1);
147 break;
148 case RISCV::LH:
149 case RISCV::LH_INX:
150 case RISCV::LHU:
151 case RISCV::FLH:
152 MemBytes = TypeSize::getFixed(2);
153 break;
154 case RISCV::LW:
155 case RISCV::LW_INX:
156 case RISCV::FLW:
157 case RISCV::LWU:
158 MemBytes = TypeSize::getFixed(4);
159 break;
160 case RISCV::LD:
161 case RISCV::LD_RV32:
162 case RISCV::FLD:
163 MemBytes = TypeSize::getFixed(8);
164 break;
165 case RISCV::VL1RE8_V:
166 case RISCV::VL2RE8_V:
167 case RISCV::VL4RE8_V:
168 case RISCV::VL8RE8_V:
169 if (!MI.getOperand(1).isFI())
170 return Register();
171 FrameIndex = MI.getOperand(1).getIndex();
172 unsigned LMUL = *getLMULForRVVWholeLoadStore(MI.getOpcode());
174 return MI.getOperand(0).getReg();
175 }
176
177 if (MI.getOperand(1).isFI() && MI.getOperand(2).isImm() &&
178 MI.getOperand(2).getImm() == 0) {
179 FrameIndex = MI.getOperand(1).getIndex();
180 return MI.getOperand(0).getReg();
181 }
182
183 return 0;
184}
185
187 int &FrameIndex) const {
188 TypeSize Dummy = TypeSize::getZero();
189 return isStoreToStackSlot(MI, FrameIndex, Dummy);
190}
191
193 int &FrameIndex,
194 TypeSize &MemBytes) const {
195 switch (MI.getOpcode()) {
196 default:
197 return 0;
198 case RISCV::SB:
199 MemBytes = TypeSize::getFixed(1);
200 break;
201 case RISCV::SH:
202 case RISCV::SH_INX:
203 case RISCV::FSH:
204 MemBytes = TypeSize::getFixed(2);
205 break;
206 case RISCV::SW:
207 case RISCV::SW_INX:
208 case RISCV::FSW:
209 MemBytes = TypeSize::getFixed(4);
210 break;
211 case RISCV::SD:
212 case RISCV::SD_RV32:
213 case RISCV::FSD:
214 MemBytes = TypeSize::getFixed(8);
215 break;
216 case RISCV::VS1R_V:
217 case RISCV::VS2R_V:
218 case RISCV::VS4R_V:
219 case RISCV::VS8R_V:
220 if (!MI.getOperand(1).isFI())
221 return Register();
222 FrameIndex = MI.getOperand(1).getIndex();
223 unsigned LMUL = *getLMULForRVVWholeLoadStore(MI.getOpcode());
225 return MI.getOperand(0).getReg();
226 }
227
228 if (MI.getOperand(1).isFI() && MI.getOperand(2).isImm() &&
229 MI.getOperand(2).getImm() == 0) {
230 FrameIndex = MI.getOperand(1).getIndex();
231 return MI.getOperand(0).getReg();
232 }
233
234 return 0;
235}
236
238 const MachineInstr &MI) const {
239 switch (RISCV::getRVVMCOpcode(MI.getOpcode())) {
240 case RISCV::VMV_V_X:
241 case RISCV::VFMV_V_F:
242 case RISCV::VMV_V_I:
243 case RISCV::VMV_S_X:
244 case RISCV::VFMV_S_F:
245 case RISCV::VID_V:
246 return MI.getOperand(1).isUndef();
247 default:
249 }
250}
251
252static bool forwardCopyWillClobberTuple(unsigned DstReg, unsigned SrcReg,
253 unsigned NumRegs) {
254 return DstReg > SrcReg && (DstReg - SrcReg) < NumRegs;
255}
256
258 const MachineBasicBlock &MBB,
261 RISCVVType::VLMUL LMul) {
263 return false;
264
265 assert(MBBI->getOpcode() == TargetOpcode::COPY &&
266 "Unexpected COPY instruction.");
267 Register SrcReg = MBBI->getOperand(1).getReg();
269
270 bool FoundDef = false;
271 bool FirstVSetVLI = false;
272 unsigned FirstSEW = 0;
273 while (MBBI != MBB.begin()) {
274 --MBBI;
275 if (MBBI->isMetaInstruction())
276 continue;
277
278 if (RISCVInstrInfo::isVectorConfigInstr(*MBBI)) {
279 // There is a vsetvli between COPY and source define instruction.
280 // vy = def_vop ... (producing instruction)
281 // ...
282 // vsetvli
283 // ...
284 // vx = COPY vy
285 if (!FoundDef) {
286 if (!FirstVSetVLI) {
287 FirstVSetVLI = true;
288 unsigned FirstVType = MBBI->getOperand(2).getImm();
289 RISCVVType::VLMUL FirstLMul = RISCVVType::getVLMUL(FirstVType);
290 FirstSEW = RISCVVType::getSEW(FirstVType);
291 // The first encountered vsetvli must have the same lmul as the
292 // register class of COPY.
293 if (FirstLMul != LMul)
294 return false;
295 }
296 // Only permit `vsetvli x0, x0, vtype` between COPY and the source
297 // define instruction.
298 if (!RISCVInstrInfo::isVLPreservingConfig(*MBBI))
299 return false;
300 continue;
301 }
302
303 // MBBI is the first vsetvli before the producing instruction.
304 unsigned VType = MBBI->getOperand(2).getImm();
305 // If there is a vsetvli between COPY and the producing instruction.
306 if (FirstVSetVLI) {
307 // If SEW is different, return false.
308 if (RISCVVType::getSEW(VType) != FirstSEW)
309 return false;
310 }
311
312 // If the vsetvli is tail undisturbed, keep the whole register move.
313 if (!RISCVVType::isTailAgnostic(VType))
314 return false;
315
316 // The checking is conservative. We only have register classes for
317 // LMUL = 1/2/4/8. We should be able to convert vmv1r.v to vmv.v.v
318 // for fractional LMUL operations. However, we could not use the vsetvli
319 // lmul for widening operations. The result of widening operation is
320 // 2 x LMUL.
321 return LMul == RISCVVType::getVLMUL(VType);
322 } else if (MBBI->isInlineAsm() || MBBI->isCall()) {
323 return false;
324 } else if (MBBI->getNumDefs()) {
325 // Check all the instructions which will change VL.
326 // For example, vleff has implicit def VL.
327 if (MBBI->modifiesRegister(RISCV::VL, /*TRI=*/nullptr))
328 return false;
329
330 // Only converting whole register copies to vmv.v.v when the defining
331 // value appears in the explicit operands.
332 for (const MachineOperand &MO : MBBI->explicit_operands()) {
333 if (!MO.isReg() || !MO.isDef())
334 continue;
335 if (!FoundDef && TRI->regsOverlap(MO.getReg(), SrcReg)) {
336 // We only permit the source of COPY has the same LMUL as the defined
337 // operand.
338 // There are cases we need to keep the whole register copy if the LMUL
339 // is different.
340 // For example,
341 // $x0 = PseudoVSETIVLI 4, 73 // vsetivli zero, 4, e16,m2,ta,m
342 // $v28m4 = PseudoVWADD_VV_M2 $v26m2, $v8m2
343 // # The COPY may be created by vlmul_trunc intrinsic.
344 // $v26m2 = COPY renamable $v28m2, implicit killed $v28m4
345 //
346 // After widening, the valid value will be 4 x e32 elements. If we
347 // convert the COPY to vmv.v.v, it will only copy 4 x e16 elements.
348 // FIXME: The COPY of subregister of Zvlsseg register will not be able
349 // to convert to vmv.v.[v|i] under the constraint.
350 if (MO.getReg() != SrcReg)
351 return false;
352
353 // In widening reduction instructions with LMUL_1 input vector case,
354 // only checking the LMUL is insufficient due to reduction result is
355 // always LMUL_1.
356 // For example,
357 // $x11 = PseudoVSETIVLI 1, 64 // vsetivli a1, 1, e8, m1, ta, mu
358 // $v8m1 = PseudoVWREDSUM_VS_M1 $v26, $v27
359 // $v26 = COPY killed renamable $v8
360 // After widening, The valid value will be 1 x e16 elements. If we
361 // convert the COPY to vmv.v.v, it will only copy 1 x e8 elements.
362 uint64_t TSFlags = MBBI->getDesc().TSFlags;
364 return false;
365
366 // If the producing instruction does not depend on vsetvli, do not
367 // convert COPY to vmv.v.v. For example, VL1R_V or PseudoVRELOAD.
368 if (!RISCVII::hasSEWOp(TSFlags) || !RISCVII::hasVLOp(TSFlags))
369 return false;
370
371 // Found the definition.
372 FoundDef = true;
373 DefMBBI = MBBI;
374 break;
375 }
376 }
377 }
378 }
379
380 return false;
381}
382
385 const DebugLoc &DL, MCRegister DstReg, MCRegister SrcReg, bool KillSrc,
386 const TargetRegisterClass *RegClass) const {
387 const RISCVRegisterInfo *TRI = STI.getRegisterInfo();
389 unsigned NF = RISCVRI::getNF(RegClass->TSFlags);
390
391 uint16_t SrcEncoding = TRI->getEncodingValue(SrcReg);
392 uint16_t DstEncoding = TRI->getEncodingValue(DstReg);
393 auto [LMulVal, Fractional] = RISCVVType::decodeVLMUL(LMul);
394 assert(!Fractional && "It is impossible be fractional lmul here.");
395 unsigned NumRegs = NF * LMulVal;
396 bool ReversedCopy =
397 forwardCopyWillClobberTuple(DstEncoding, SrcEncoding, NumRegs);
398 if (ReversedCopy) {
399 // If the src and dest overlap when copying a tuple, we need to copy the
400 // registers in reverse.
401 SrcEncoding += NumRegs - 1;
402 DstEncoding += NumRegs - 1;
403 }
404
405 unsigned I = 0;
406 auto GetCopyInfo = [&](uint16_t SrcEncoding, uint16_t DstEncoding)
407 -> std::tuple<RISCVVType::VLMUL, const TargetRegisterClass &, unsigned,
408 unsigned, unsigned> {
409 if (ReversedCopy) {
410 // For reversed copying, if there are enough aligned registers(8/4/2), we
411 // can do a larger copy(LMUL8/4/2).
412 // Besides, we have already known that DstEncoding is larger than
413 // SrcEncoding in forwardCopyWillClobberTuple, so the difference between
414 // DstEncoding and SrcEncoding should be >= LMUL value we try to use to
415 // avoid clobbering.
416 uint16_t Diff = DstEncoding - SrcEncoding;
417 if (I + 8 <= NumRegs && Diff >= 8 && SrcEncoding % 8 == 7 &&
418 DstEncoding % 8 == 7)
419 return {RISCVVType::LMUL_8, RISCV::VRM8RegClass, RISCV::VMV8R_V,
420 RISCV::PseudoVMV_V_V_M8, RISCV::PseudoVMV_V_I_M8};
421 if (I + 4 <= NumRegs && Diff >= 4 && SrcEncoding % 4 == 3 &&
422 DstEncoding % 4 == 3)
423 return {RISCVVType::LMUL_4, RISCV::VRM4RegClass, RISCV::VMV4R_V,
424 RISCV::PseudoVMV_V_V_M4, RISCV::PseudoVMV_V_I_M4};
425 if (I + 2 <= NumRegs && Diff >= 2 && SrcEncoding % 2 == 1 &&
426 DstEncoding % 2 == 1)
427 return {RISCVVType::LMUL_2, RISCV::VRM2RegClass, RISCV::VMV2R_V,
428 RISCV::PseudoVMV_V_V_M2, RISCV::PseudoVMV_V_I_M2};
429 // Or we should do LMUL1 copying.
430 return {RISCVVType::LMUL_1, RISCV::VRRegClass, RISCV::VMV1R_V,
431 RISCV::PseudoVMV_V_V_M1, RISCV::PseudoVMV_V_I_M1};
432 }
433
434 // For forward copying, if source register encoding and destination register
435 // encoding are aligned to 8/4/2, we can do a LMUL8/4/2 copying.
436 if (I + 8 <= NumRegs && SrcEncoding % 8 == 0 && DstEncoding % 8 == 0)
437 return {RISCVVType::LMUL_8, RISCV::VRM8RegClass, RISCV::VMV8R_V,
438 RISCV::PseudoVMV_V_V_M8, RISCV::PseudoVMV_V_I_M8};
439 if (I + 4 <= NumRegs && SrcEncoding % 4 == 0 && DstEncoding % 4 == 0)
440 return {RISCVVType::LMUL_4, RISCV::VRM4RegClass, RISCV::VMV4R_V,
441 RISCV::PseudoVMV_V_V_M4, RISCV::PseudoVMV_V_I_M4};
442 if (I + 2 <= NumRegs && SrcEncoding % 2 == 0 && DstEncoding % 2 == 0)
443 return {RISCVVType::LMUL_2, RISCV::VRM2RegClass, RISCV::VMV2R_V,
444 RISCV::PseudoVMV_V_V_M2, RISCV::PseudoVMV_V_I_M2};
445 // Or we should do LMUL1 copying.
446 return {RISCVVType::LMUL_1, RISCV::VRRegClass, RISCV::VMV1R_V,
447 RISCV::PseudoVMV_V_V_M1, RISCV::PseudoVMV_V_I_M1};
448 };
449
450 while (I != NumRegs) {
451 // For non-segment copying, we only do this once as the registers are always
452 // aligned.
453 // For segment copying, we may do this several times. If the registers are
454 // aligned to larger LMUL, we can eliminate some copyings.
455 auto [LMulCopied, RegClass, Opc, VVOpc, VIOpc] =
456 GetCopyInfo(SrcEncoding, DstEncoding);
457 auto [NumCopied, _] = RISCVVType::decodeVLMUL(LMulCopied);
458
460 if (LMul == LMulCopied &&
461 isConvertibleToVMV_V_V(STI, MBB, MBBI, DefMBBI, LMul)) {
462 Opc = VVOpc;
463 if (DefMBBI->getOpcode() == VIOpc)
464 Opc = VIOpc;
465 }
466
467 // Emit actual copying.
468 // For reversed copying, the encoding should be decreased.
469 MCRegister ActualSrcReg = TRI->findVRegWithEncoding(
470 RegClass, ReversedCopy ? (SrcEncoding - NumCopied + 1) : SrcEncoding);
471 MCRegister ActualDstReg = TRI->findVRegWithEncoding(
472 RegClass, ReversedCopy ? (DstEncoding - NumCopied + 1) : DstEncoding);
473
474 auto MIB = BuildMI(MBB, MBBI, DL, get(Opc), ActualDstReg);
475 bool UseVMV_V_I = RISCV::getRVVMCOpcode(Opc) == RISCV::VMV_V_I;
476 bool UseVMV = UseVMV_V_I || RISCV::getRVVMCOpcode(Opc) == RISCV::VMV_V_V;
477 if (UseVMV)
478 MIB.addReg(ActualDstReg, RegState::Undef);
479 if (UseVMV_V_I)
480 MIB = MIB.add(DefMBBI->getOperand(2));
481 else
482 MIB = MIB.addReg(ActualSrcReg, getKillRegState(KillSrc));
483 if (UseVMV) {
484 const MCInstrDesc &Desc = DefMBBI->getDesc();
485 MIB.add(DefMBBI->getOperand(RISCVII::getVLOpNum(Desc))); // AVL
486 unsigned Log2SEW =
487 DefMBBI->getOperand(RISCVII::getSEWOpNum(Desc)).getImm();
488 MIB.addImm(Log2SEW ? Log2SEW : 3); // SEW
489 MIB.addImm(0); // tu, mu
490 MIB.addReg(RISCV::VL, RegState::Implicit);
491 MIB.addReg(RISCV::VTYPE, RegState::Implicit);
492 }
493 // Add an implicit read of the original source to silence the verifier
494 // in the cases where some of the smaller VRs we're copying from might be
495 // undef, caused by the fact that the original, larger source VR might not
496 // be fully initialized at the time this COPY happens.
497 MIB.addReg(SrcReg, RegState::Implicit);
498
499 // If we are copying reversely, we should decrease the encoding.
500 SrcEncoding += (ReversedCopy ? -NumCopied : NumCopied);
501 DstEncoding += (ReversedCopy ? -NumCopied : NumCopied);
502 I += NumCopied;
503 }
504}
505
508 const DebugLoc &DL, Register DstReg,
509 Register SrcReg, bool KillSrc,
510 bool RenamableDest, bool RenamableSrc) const {
511 const TargetRegisterInfo *TRI = STI.getRegisterInfo();
512 unsigned KillFlag = getKillRegState(KillSrc);
513
514 if (RISCV::GPRRegClass.contains(DstReg, SrcReg)) {
515 BuildMI(MBB, MBBI, DL, get(RISCV::ADDI), DstReg)
516 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc))
517 .addImm(0);
518 return;
519 }
520
521 if (RISCV::GPRF16RegClass.contains(DstReg, SrcReg)) {
522 BuildMI(MBB, MBBI, DL, get(RISCV::PseudoMV_FPR16INX), DstReg)
523 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc));
524 return;
525 }
526
527 if (RISCV::GPRF32RegClass.contains(DstReg, SrcReg)) {
528 BuildMI(MBB, MBBI, DL, get(RISCV::PseudoMV_FPR32INX), DstReg)
529 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc));
530 return;
531 }
532
533 if (RISCV::GPRPairRegClass.contains(DstReg, SrcReg)) {
534 if (STI.isRV32() && STI.hasStdExtZdinx()) {
535 // On RV32_Zdinx, FMV.D will move a pair of registers to another pair of
536 // registers, in one instruction.
537 BuildMI(MBB, MBBI, DL, get(RISCV::FSGNJ_D_IN32X), DstReg)
538 .addReg(SrcReg, getRenamableRegState(RenamableSrc))
539 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc));
540 return;
541 }
542
543 MCRegister EvenReg = TRI->getSubReg(SrcReg, RISCV::sub_gpr_even);
544 MCRegister OddReg = TRI->getSubReg(SrcReg, RISCV::sub_gpr_odd);
545 // We need to correct the odd register of X0_Pair.
546 if (OddReg == RISCV::DUMMY_REG_PAIR_WITH_X0)
547 OddReg = RISCV::X0;
548 assert(DstReg != RISCV::X0_Pair && "Cannot write to X0_Pair");
549
550 // Emit an ADDI for both parts of GPRPair.
551 BuildMI(MBB, MBBI, DL, get(RISCV::ADDI),
552 TRI->getSubReg(DstReg, RISCV::sub_gpr_even))
553 .addReg(EvenReg, KillFlag)
554 .addImm(0);
555 BuildMI(MBB, MBBI, DL, get(RISCV::ADDI),
556 TRI->getSubReg(DstReg, RISCV::sub_gpr_odd))
557 .addReg(OddReg, KillFlag)
558 .addImm(0);
559 return;
560 }
561
562 // Handle copy from csr
563 if (RISCV::VCSRRegClass.contains(SrcReg) &&
564 RISCV::GPRRegClass.contains(DstReg)) {
565 BuildMI(MBB, MBBI, DL, get(RISCV::CSRRS), DstReg)
566 .addImm(RISCVSysReg::lookupSysRegByName(TRI->getName(SrcReg))->Encoding)
567 .addReg(RISCV::X0);
568 return;
569 }
570
571 if (RISCV::FPR16RegClass.contains(DstReg, SrcReg)) {
572 unsigned Opc;
573 if (STI.hasStdExtZfh()) {
574 Opc = RISCV::FSGNJ_H;
575 } else {
576 assert(STI.hasStdExtF() &&
577 (STI.hasStdExtZfhmin() || STI.hasStdExtZfbfmin()) &&
578 "Unexpected extensions");
579 // Zfhmin/Zfbfmin doesn't have FSGNJ_H, replace FSGNJ_H with FSGNJ_S.
580 DstReg = TRI->getMatchingSuperReg(DstReg, RISCV::sub_16,
581 &RISCV::FPR32RegClass);
582 SrcReg = TRI->getMatchingSuperReg(SrcReg, RISCV::sub_16,
583 &RISCV::FPR32RegClass);
584 Opc = RISCV::FSGNJ_S;
585 }
586 BuildMI(MBB, MBBI, DL, get(Opc), DstReg)
587 .addReg(SrcReg, KillFlag)
588 .addReg(SrcReg, KillFlag);
589 return;
590 }
591
592 if (RISCV::FPR32RegClass.contains(DstReg, SrcReg)) {
593 BuildMI(MBB, MBBI, DL, get(RISCV::FSGNJ_S), DstReg)
594 .addReg(SrcReg, KillFlag)
595 .addReg(SrcReg, KillFlag);
596 return;
597 }
598
599 if (RISCV::FPR64RegClass.contains(DstReg, SrcReg)) {
600 BuildMI(MBB, MBBI, DL, get(RISCV::FSGNJ_D), DstReg)
601 .addReg(SrcReg, KillFlag)
602 .addReg(SrcReg, KillFlag);
603 return;
604 }
605
606 if (RISCV::FPR32RegClass.contains(DstReg) &&
607 RISCV::GPRRegClass.contains(SrcReg)) {
608 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_W_X), DstReg)
609 .addReg(SrcReg, KillFlag);
610 return;
611 }
612
613 if (RISCV::GPRRegClass.contains(DstReg) &&
614 RISCV::FPR32RegClass.contains(SrcReg)) {
615 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_X_W), DstReg)
616 .addReg(SrcReg, KillFlag);
617 return;
618 }
619
620 if (RISCV::FPR64RegClass.contains(DstReg) &&
621 RISCV::GPRRegClass.contains(SrcReg)) {
622 assert(STI.getXLen() == 64 && "Unexpected GPR size");
623 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_D_X), DstReg)
624 .addReg(SrcReg, KillFlag);
625 return;
626 }
627
628 if (RISCV::GPRRegClass.contains(DstReg) &&
629 RISCV::FPR64RegClass.contains(SrcReg)) {
630 assert(STI.getXLen() == 64 && "Unexpected GPR size");
631 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_X_D), DstReg)
632 .addReg(SrcReg, KillFlag);
633 return;
634 }
635
636 // VR->VR copies.
637 const TargetRegisterClass *RegClass =
638 TRI->getCommonMinimalPhysRegClass(SrcReg, DstReg);
639 if (RISCVRegisterInfo::isRVVRegClass(RegClass)) {
640 copyPhysRegVector(MBB, MBBI, DL, DstReg, SrcReg, KillSrc, RegClass);
641 return;
642 }
643
644 llvm_unreachable("Impossible reg-to-reg copy");
645}
646
649 Register SrcReg, bool IsKill, int FI,
650 const TargetRegisterClass *RC,
651 Register VReg,
652 MachineInstr::MIFlag Flags) const {
653 MachineFunction *MF = MBB.getParent();
654 MachineFrameInfo &MFI = MF->getFrameInfo();
655 Align Alignment = MFI.getObjectAlign(FI);
656
657 unsigned Opcode;
658 if (RISCV::GPRRegClass.hasSubClassEq(RC)) {
659 Opcode = RegInfo.getRegSizeInBits(RISCV::GPRRegClass) == 32 ? RISCV::SW
660 : RISCV::SD;
661 } else if (RISCV::GPRF16RegClass.hasSubClassEq(RC)) {
662 Opcode = RISCV::SH_INX;
663 } else if (RISCV::GPRF32RegClass.hasSubClassEq(RC)) {
664 Opcode = RISCV::SW_INX;
665 } else if (RISCV::GPRPairRegClass.hasSubClassEq(RC)) {
666 if (!STI.is64Bit() && STI.hasStdExtZilsd() &&
667 Alignment >= STI.getZilsdAlign()) {
668 Opcode = RISCV::SD_RV32;
669 } else {
670 Opcode = RISCV::PseudoRV32ZdinxSD;
671 }
672 } else if (RISCV::FPR16RegClass.hasSubClassEq(RC)) {
673 Opcode = RISCV::FSH;
674 } else if (RISCV::FPR32RegClass.hasSubClassEq(RC)) {
675 Opcode = RISCV::FSW;
676 } else if (RISCV::FPR64RegClass.hasSubClassEq(RC)) {
677 Opcode = RISCV::FSD;
678 } else if (RISCV::VRRegClass.hasSubClassEq(RC)) {
679 Opcode = RISCV::VS1R_V;
680 } else if (RISCV::VRM2RegClass.hasSubClassEq(RC)) {
681 Opcode = RISCV::VS2R_V;
682 } else if (RISCV::VRM4RegClass.hasSubClassEq(RC)) {
683 Opcode = RISCV::VS4R_V;
684 } else if (RISCV::VRM8RegClass.hasSubClassEq(RC)) {
685 Opcode = RISCV::VS8R_V;
686 } else if (RISCV::VRN2M1RegClass.hasSubClassEq(RC))
687 Opcode = RISCV::PseudoVSPILL2_M1;
688 else if (RISCV::VRN2M2RegClass.hasSubClassEq(RC))
689 Opcode = RISCV::PseudoVSPILL2_M2;
690 else if (RISCV::VRN2M4RegClass.hasSubClassEq(RC))
691 Opcode = RISCV::PseudoVSPILL2_M4;
692 else if (RISCV::VRN3M1RegClass.hasSubClassEq(RC))
693 Opcode = RISCV::PseudoVSPILL3_M1;
694 else if (RISCV::VRN3M2RegClass.hasSubClassEq(RC))
695 Opcode = RISCV::PseudoVSPILL3_M2;
696 else if (RISCV::VRN4M1RegClass.hasSubClassEq(RC))
697 Opcode = RISCV::PseudoVSPILL4_M1;
698 else if (RISCV::VRN4M2RegClass.hasSubClassEq(RC))
699 Opcode = RISCV::PseudoVSPILL4_M2;
700 else if (RISCV::VRN5M1RegClass.hasSubClassEq(RC))
701 Opcode = RISCV::PseudoVSPILL5_M1;
702 else if (RISCV::VRN6M1RegClass.hasSubClassEq(RC))
703 Opcode = RISCV::PseudoVSPILL6_M1;
704 else if (RISCV::VRN7M1RegClass.hasSubClassEq(RC))
705 Opcode = RISCV::PseudoVSPILL7_M1;
706 else if (RISCV::VRN8M1RegClass.hasSubClassEq(RC))
707 Opcode = RISCV::PseudoVSPILL8_M1;
708 else
709 llvm_unreachable("Can't store this register to stack slot");
710
714 TypeSize::getScalable(MFI.getObjectSize(FI)), Alignment);
715
717 BuildMI(MBB, I, DebugLoc(), get(Opcode))
718 .addReg(SrcReg, getKillRegState(IsKill))
719 .addFrameIndex(FI)
720 .addMemOperand(MMO)
721 .setMIFlag(Flags);
722 NumVRegSpilled += RegInfo.getRegSizeInBits(*RC) / RISCV::RVVBitsPerBlock;
723 } else {
726 MFI.getObjectSize(FI), Alignment);
727
728 BuildMI(MBB, I, DebugLoc(), get(Opcode))
729 .addReg(SrcReg, getKillRegState(IsKill))
730 .addFrameIndex(FI)
731 .addImm(0)
732 .addMemOperand(MMO)
733 .setMIFlag(Flags);
734 }
735}
736
739 Register DstReg, int FI,
740 const TargetRegisterClass *RC,
741 Register VReg, unsigned SubReg,
742 MachineInstr::MIFlag Flags) const {
743 MachineFunction *MF = MBB.getParent();
744 MachineFrameInfo &MFI = MF->getFrameInfo();
745 Align Alignment = MFI.getObjectAlign(FI);
746 DebugLoc DL =
747 Flags & MachineInstr::FrameDestroy ? MBB.findDebugLoc(I) : DebugLoc();
748
749 unsigned Opcode;
750 if (RISCV::GPRRegClass.hasSubClassEq(RC)) {
751 Opcode = RegInfo.getRegSizeInBits(RISCV::GPRRegClass) == 32 ? RISCV::LW
752 : RISCV::LD;
753 } else if (RISCV::GPRF16RegClass.hasSubClassEq(RC)) {
754 Opcode = RISCV::LH_INX;
755 } else if (RISCV::GPRF32RegClass.hasSubClassEq(RC)) {
756 Opcode = RISCV::LW_INX;
757 } else if (RISCV::GPRPairRegClass.hasSubClassEq(RC)) {
758 if (!STI.is64Bit() && STI.hasStdExtZilsd() &&
759 Alignment >= STI.getZilsdAlign()) {
760 Opcode = RISCV::LD_RV32;
761 } else {
762 Opcode = RISCV::PseudoRV32ZdinxLD;
763 }
764 } else if (RISCV::FPR16RegClass.hasSubClassEq(RC)) {
765 Opcode = RISCV::FLH;
766 } else if (RISCV::FPR32RegClass.hasSubClassEq(RC)) {
767 Opcode = RISCV::FLW;
768 } else if (RISCV::FPR64RegClass.hasSubClassEq(RC)) {
769 Opcode = RISCV::FLD;
770 } else if (RISCV::VRRegClass.hasSubClassEq(RC)) {
771 Opcode = RISCV::VL1RE8_V;
772 } else if (RISCV::VRM2RegClass.hasSubClassEq(RC)) {
773 Opcode = RISCV::VL2RE8_V;
774 } else if (RISCV::VRM4RegClass.hasSubClassEq(RC)) {
775 Opcode = RISCV::VL4RE8_V;
776 } else if (RISCV::VRM8RegClass.hasSubClassEq(RC)) {
777 Opcode = RISCV::VL8RE8_V;
778 } else if (RISCV::VRN2M1RegClass.hasSubClassEq(RC))
779 Opcode = RISCV::PseudoVRELOAD2_M1;
780 else if (RISCV::VRN2M2RegClass.hasSubClassEq(RC))
781 Opcode = RISCV::PseudoVRELOAD2_M2;
782 else if (RISCV::VRN2M4RegClass.hasSubClassEq(RC))
783 Opcode = RISCV::PseudoVRELOAD2_M4;
784 else if (RISCV::VRN3M1RegClass.hasSubClassEq(RC))
785 Opcode = RISCV::PseudoVRELOAD3_M1;
786 else if (RISCV::VRN3M2RegClass.hasSubClassEq(RC))
787 Opcode = RISCV::PseudoVRELOAD3_M2;
788 else if (RISCV::VRN4M1RegClass.hasSubClassEq(RC))
789 Opcode = RISCV::PseudoVRELOAD4_M1;
790 else if (RISCV::VRN4M2RegClass.hasSubClassEq(RC))
791 Opcode = RISCV::PseudoVRELOAD4_M2;
792 else if (RISCV::VRN5M1RegClass.hasSubClassEq(RC))
793 Opcode = RISCV::PseudoVRELOAD5_M1;
794 else if (RISCV::VRN6M1RegClass.hasSubClassEq(RC))
795 Opcode = RISCV::PseudoVRELOAD6_M1;
796 else if (RISCV::VRN7M1RegClass.hasSubClassEq(RC))
797 Opcode = RISCV::PseudoVRELOAD7_M1;
798 else if (RISCV::VRN8M1RegClass.hasSubClassEq(RC))
799 Opcode = RISCV::PseudoVRELOAD8_M1;
800 else
801 llvm_unreachable("Can't load this register from stack slot");
802
806 TypeSize::getScalable(MFI.getObjectSize(FI)), Alignment);
807
809 BuildMI(MBB, I, DL, get(Opcode), DstReg)
810 .addFrameIndex(FI)
811 .addMemOperand(MMO)
812 .setMIFlag(Flags);
813 NumVRegReloaded += RegInfo.getRegSizeInBits(*RC) / RISCV::RVVBitsPerBlock;
814 } else {
817 MFI.getObjectSize(FI), Alignment);
818
819 BuildMI(MBB, I, DL, get(Opcode), DstReg)
820 .addFrameIndex(FI)
821 .addImm(0)
822 .addMemOperand(MMO)
823 .setMIFlag(Flags);
824 }
825}
826std::optional<unsigned> getFoldedOpcode(MachineFunction &MF, MachineInstr &MI,
828 const RISCVSubtarget &ST) {
829
830 // The below optimizations narrow the load so they are only valid for little
831 // endian.
832 // TODO: Support big endian by adding an offset into the frame object?
833 if (MF.getDataLayout().isBigEndian())
834 return std::nullopt;
835
836 // Fold load from stack followed by sext.b/sext.h/sext.w/zext.b/zext.h/zext.w.
837 if (Ops.size() != 1 || Ops[0] != 1)
838 return std::nullopt;
839
840 switch (MI.getOpcode()) {
841 default:
842 if (RISCVInstrInfo::isSEXT_W(MI))
843 return RISCV::LW;
844 if (RISCVInstrInfo::isZEXT_W(MI))
845 return RISCV::LWU;
846 if (RISCVInstrInfo::isZEXT_B(MI))
847 return RISCV::LBU;
848 break;
849 case RISCV::SEXT_H:
850 return RISCV::LH;
851 case RISCV::SEXT_B:
852 return RISCV::LB;
853 case RISCV::ZEXT_H_RV32:
854 case RISCV::ZEXT_H_RV64:
855 return RISCV::LHU;
856 }
857
858 switch (RISCV::getRVVMCOpcode(MI.getOpcode())) {
859 default:
860 return std::nullopt;
861 case RISCV::VMV_X_S: {
862 unsigned Log2SEW =
863 MI.getOperand(RISCVII::getSEWOpNum(MI.getDesc())).getImm();
864 if (ST.getXLen() < (1U << Log2SEW))
865 return std::nullopt;
866 switch (Log2SEW) {
867 case 3:
868 return RISCV::LB;
869 case 4:
870 return RISCV::LH;
871 case 5:
872 return RISCV::LW;
873 case 6:
874 return RISCV::LD;
875 default:
876 llvm_unreachable("Unexpected SEW");
877 }
878 }
879 case RISCV::VFMV_F_S: {
880 unsigned Log2SEW =
881 MI.getOperand(RISCVII::getSEWOpNum(MI.getDesc())).getImm();
882 switch (Log2SEW) {
883 case 4:
884 return RISCV::FLH;
885 case 5:
886 return RISCV::FLW;
887 case 6:
888 return RISCV::FLD;
889 default:
890 llvm_unreachable("Unexpected SEW");
891 }
892 }
893 }
894}
895
896// This is the version used during InlineSpiller::spillAroundUses
899 MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS,
900 VirtRegMap *VRM) const {
901
902 std::optional<unsigned> LoadOpc = getFoldedOpcode(MF, MI, Ops, STI);
903 if (!LoadOpc)
904 return nullptr;
905 Register DstReg = MI.getOperand(0).getReg();
906 return BuildMI(*MI.getParent(), InsertPt, MI.getDebugLoc(), get(*LoadOpc),
907 DstReg)
908 .addFrameIndex(FrameIndex)
909 .addImm(0);
910}
911
912static unsigned getLoadPredicatedOpcode(unsigned Opcode) {
913 switch (Opcode) {
914 case RISCV::LB:
915 return RISCV::PseudoCCLB;
916 case RISCV::LBU:
917 return RISCV::PseudoCCLBU;
918 case RISCV::LH:
919 return RISCV::PseudoCCLH;
920 case RISCV::LHU:
921 return RISCV::PseudoCCLHU;
922 case RISCV::LW:
923 return RISCV::PseudoCCLW;
924 case RISCV::LWU:
925 return RISCV::PseudoCCLWU;
926 case RISCV::LD:
927 return RISCV::PseudoCCLD;
928 case RISCV::QC_E_LB:
929 return RISCV::PseudoCCQC_E_LB;
930 case RISCV::QC_E_LBU:
931 return RISCV::PseudoCCQC_E_LBU;
932 case RISCV::QC_E_LH:
933 return RISCV::PseudoCCQC_E_LH;
934 case RISCV::QC_E_LHU:
935 return RISCV::PseudoCCQC_E_LHU;
936 case RISCV::QC_E_LW:
937 return RISCV::PseudoCCQC_E_LW;
938 default:
939 return 0;
940 }
941}
942
946 LiveIntervals *LIS) const {
947 // For now, only handle RISCV::PseudoCCMOVGPR.
948 if (MI.getOpcode() != RISCV::PseudoCCMOVGPR)
949 return nullptr;
950
951 unsigned PredOpc = getLoadPredicatedOpcode(LoadMI.getOpcode());
952
953 if (!STI.hasShortForwardBranchILoad() || !PredOpc)
954 return nullptr;
955
957 if (Ops.size() != 1 || (Ops[0] != 4 && Ops[0] != 5))
958 return nullptr;
959
960 bool Invert = Ops[0] == 5;
961 const MachineOperand &FalseReg = MI.getOperand(!Invert ? 5 : 4);
962 Register DestReg = MI.getOperand(0).getReg();
963 const TargetRegisterClass *PreviousClass = MRI.getRegClass(FalseReg.getReg());
964 if (!MRI.constrainRegClass(DestReg, PreviousClass))
965 return nullptr;
966
967 // Create a new predicated version of DefMI.
968 MachineInstrBuilder NewMI = BuildMI(*MI.getParent(), InsertPt,
969 MI.getDebugLoc(), get(PredOpc), DestReg)
970 .add({MI.getOperand(1), MI.getOperand(2)});
971
972 // Add condition code, inverting if necessary.
973 auto CC = static_cast<RISCVCC::CondCode>(MI.getOperand(3).getImm());
974 if (!Invert)
976 NewMI.addImm(CC);
977
978 // Copy the false register.
979 NewMI.add(FalseReg);
980
981 // Copy all the DefMI operands.
982 const MCInstrDesc &DefDesc = LoadMI.getDesc();
983 for (unsigned i = 1, e = DefDesc.getNumOperands(); i != e; ++i)
984 NewMI.add(LoadMI.getOperand(i));
985
986 NewMI.cloneMemRefs(LoadMI);
987 return NewMI;
988}
989
992 const DebugLoc &DL, Register DstReg, uint64_t Val,
993 MachineInstr::MIFlag Flag, bool DstRenamable,
994 bool DstIsDead) const {
995 Register SrcReg = RISCV::X0;
996
997 // For RV32, allow a sign or unsigned 32 bit value.
998 if (!STI.is64Bit() && !isInt<32>(Val)) {
999 // If have a uimm32 it will still fit in a register so we can allow it.
1000 if (!isUInt<32>(Val))
1001 report_fatal_error("Should only materialize 32-bit constants for RV32");
1002
1003 // Sign extend for generateInstSeq.
1004 Val = SignExtend64<32>(Val);
1005 }
1006
1008 assert(!Seq.empty());
1009
1010 bool SrcRenamable = false;
1011 unsigned Num = 0;
1012
1013 for (const RISCVMatInt::Inst &Inst : Seq) {
1014 bool LastItem = ++Num == Seq.size();
1015 unsigned DstRegState = getDeadRegState(DstIsDead && LastItem) |
1016 getRenamableRegState(DstRenamable);
1017 unsigned SrcRegState = getKillRegState(SrcReg != RISCV::X0) |
1018 getRenamableRegState(SrcRenamable);
1019 switch (Inst.getOpndKind()) {
1020 case RISCVMatInt::Imm:
1021 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1022 .addReg(DstReg, RegState::Define | DstRegState)
1023 .addImm(Inst.getImm())
1024 .setMIFlag(Flag);
1025 break;
1026 case RISCVMatInt::RegX0:
1027 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1028 .addReg(DstReg, RegState::Define | DstRegState)
1029 .addReg(SrcReg, SrcRegState)
1030 .addReg(RISCV::X0)
1031 .setMIFlag(Flag);
1032 break;
1034 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1035 .addReg(DstReg, RegState::Define | DstRegState)
1036 .addReg(SrcReg, SrcRegState)
1037 .addReg(SrcReg, SrcRegState)
1038 .setMIFlag(Flag);
1039 break;
1041 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1042 .addReg(DstReg, RegState::Define | DstRegState)
1043 .addReg(SrcReg, SrcRegState)
1044 .addImm(Inst.getImm())
1045 .setMIFlag(Flag);
1046 break;
1047 }
1048
1049 // Only the first instruction has X0 as its source.
1050 SrcReg = DstReg;
1051 SrcRenamable = DstRenamable;
1052 }
1053}
1054
1056 switch (Opc) {
1057 default:
1058 return RISCVCC::COND_INVALID;
1059 case RISCV::BEQ:
1060 case RISCV::BEQI:
1061 case RISCV::CV_BEQIMM:
1062 case RISCV::QC_BEQI:
1063 case RISCV::QC_E_BEQI:
1064 case RISCV::NDS_BBC:
1065 case RISCV::NDS_BEQC:
1066 return RISCVCC::COND_EQ;
1067 case RISCV::BNE:
1068 case RISCV::BNEI:
1069 case RISCV::QC_BNEI:
1070 case RISCV::QC_E_BNEI:
1071 case RISCV::CV_BNEIMM:
1072 case RISCV::NDS_BBS:
1073 case RISCV::NDS_BNEC:
1074 return RISCVCC::COND_NE;
1075 case RISCV::BLT:
1076 case RISCV::QC_BLTI:
1077 case RISCV::QC_E_BLTI:
1078 return RISCVCC::COND_LT;
1079 case RISCV::BGE:
1080 case RISCV::QC_BGEI:
1081 case RISCV::QC_E_BGEI:
1082 return RISCVCC::COND_GE;
1083 case RISCV::BLTU:
1084 case RISCV::QC_BLTUI:
1085 case RISCV::QC_E_BLTUI:
1086 return RISCVCC::COND_LTU;
1087 case RISCV::BGEU:
1088 case RISCV::QC_BGEUI:
1089 case RISCV::QC_E_BGEUI:
1090 return RISCVCC::COND_GEU;
1091 }
1092}
1093
1095 int64_t C1) {
1096 switch (CC) {
1097 default:
1098 llvm_unreachable("Unexpected CC");
1099 case RISCVCC::COND_EQ:
1100 return C0 == C1;
1101 case RISCVCC::COND_NE:
1102 return C0 != C1;
1103 case RISCVCC::COND_LT:
1104 return C0 < C1;
1105 case RISCVCC::COND_GE:
1106 return C0 >= C1;
1107 case RISCVCC::COND_LTU:
1108 return (uint64_t)C0 < (uint64_t)C1;
1109 case RISCVCC::COND_GEU:
1110 return (uint64_t)C0 >= (uint64_t)C1;
1111 }
1112}
1113
1114// The contents of values added to Cond are not examined outside of
1115// RISCVInstrInfo, giving us flexibility in what to push to it. For RISCV, we
1116// push BranchOpcode, Reg1, Reg2.
1119 // Block ends with fall-through condbranch.
1120 assert(LastInst.getDesc().isConditionalBranch() &&
1121 "Unknown conditional branch");
1122 Target = LastInst.getOperand(2).getMBB();
1123 Cond.push_back(MachineOperand::CreateImm(LastInst.getOpcode()));
1124 Cond.push_back(LastInst.getOperand(0));
1125 Cond.push_back(LastInst.getOperand(1));
1126}
1127
1128static unsigned getInverseXqcicmOpcode(unsigned Opcode) {
1129 switch (Opcode) {
1130 default:
1131 llvm_unreachable("Unexpected Opcode");
1132 case RISCV::QC_MVEQ:
1133 return RISCV::QC_MVNE;
1134 case RISCV::QC_MVNE:
1135 return RISCV::QC_MVEQ;
1136 case RISCV::QC_MVLT:
1137 return RISCV::QC_MVGE;
1138 case RISCV::QC_MVGE:
1139 return RISCV::QC_MVLT;
1140 case RISCV::QC_MVLTU:
1141 return RISCV::QC_MVGEU;
1142 case RISCV::QC_MVGEU:
1143 return RISCV::QC_MVLTU;
1144 case RISCV::QC_MVEQI:
1145 return RISCV::QC_MVNEI;
1146 case RISCV::QC_MVNEI:
1147 return RISCV::QC_MVEQI;
1148 case RISCV::QC_MVLTI:
1149 return RISCV::QC_MVGEI;
1150 case RISCV::QC_MVGEI:
1151 return RISCV::QC_MVLTI;
1152 case RISCV::QC_MVLTUI:
1153 return RISCV::QC_MVGEUI;
1154 case RISCV::QC_MVGEUI:
1155 return RISCV::QC_MVLTUI;
1156 }
1157}
1158
1159unsigned RISCVCC::getBrCond(RISCVCC::CondCode CC, unsigned SelectOpc) {
1160 switch (SelectOpc) {
1161 default:
1162 switch (CC) {
1163 default:
1164 llvm_unreachable("Unexpected condition code!");
1165 case RISCVCC::COND_EQ:
1166 return RISCV::BEQ;
1167 case RISCVCC::COND_NE:
1168 return RISCV::BNE;
1169 case RISCVCC::COND_LT:
1170 return RISCV::BLT;
1171 case RISCVCC::COND_GE:
1172 return RISCV::BGE;
1173 case RISCVCC::COND_LTU:
1174 return RISCV::BLTU;
1175 case RISCVCC::COND_GEU:
1176 return RISCV::BGEU;
1177 }
1178 break;
1179 case RISCV::Select_GPR_Using_CC_Imm5_Zibi:
1180 switch (CC) {
1181 default:
1182 llvm_unreachable("Unexpected condition code!");
1183 case RISCVCC::COND_EQ:
1184 return RISCV::BEQI;
1185 case RISCVCC::COND_NE:
1186 return RISCV::BNEI;
1187 }
1188 break;
1189 case RISCV::Select_GPR_Using_CC_SImm5_CV:
1190 switch (CC) {
1191 default:
1192 llvm_unreachable("Unexpected condition code!");
1193 case RISCVCC::COND_EQ:
1194 return RISCV::CV_BEQIMM;
1195 case RISCVCC::COND_NE:
1196 return RISCV::CV_BNEIMM;
1197 }
1198 break;
1199 case RISCV::Select_GPRNoX0_Using_CC_SImm5NonZero_QC:
1200 switch (CC) {
1201 default:
1202 llvm_unreachable("Unexpected condition code!");
1203 case RISCVCC::COND_EQ:
1204 return RISCV::QC_BEQI;
1205 case RISCVCC::COND_NE:
1206 return RISCV::QC_BNEI;
1207 case RISCVCC::COND_LT:
1208 return RISCV::QC_BLTI;
1209 case RISCVCC::COND_GE:
1210 return RISCV::QC_BGEI;
1211 }
1212 break;
1213 case RISCV::Select_GPRNoX0_Using_CC_UImm5NonZero_QC:
1214 switch (CC) {
1215 default:
1216 llvm_unreachable("Unexpected condition code!");
1217 case RISCVCC::COND_LTU:
1218 return RISCV::QC_BLTUI;
1219 case RISCVCC::COND_GEU:
1220 return RISCV::QC_BGEUI;
1221 }
1222 break;
1223 case RISCV::Select_GPRNoX0_Using_CC_SImm16NonZero_QC:
1224 switch (CC) {
1225 default:
1226 llvm_unreachable("Unexpected condition code!");
1227 case RISCVCC::COND_EQ:
1228 return RISCV::QC_E_BEQI;
1229 case RISCVCC::COND_NE:
1230 return RISCV::QC_E_BNEI;
1231 case RISCVCC::COND_LT:
1232 return RISCV::QC_E_BLTI;
1233 case RISCVCC::COND_GE:
1234 return RISCV::QC_E_BGEI;
1235 }
1236 break;
1237 case RISCV::Select_GPRNoX0_Using_CC_UImm16NonZero_QC:
1238 switch (CC) {
1239 default:
1240 llvm_unreachable("Unexpected condition code!");
1241 case RISCVCC::COND_LTU:
1242 return RISCV::QC_E_BLTUI;
1243 case RISCVCC::COND_GEU:
1244 return RISCV::QC_E_BGEUI;
1245 }
1246 break;
1247 case RISCV::Select_GPR_Using_CC_UImmLog2XLen_NDS:
1248 switch (CC) {
1249 default:
1250 llvm_unreachable("Unexpected condition code!");
1251 case RISCVCC::COND_EQ:
1252 return RISCV::NDS_BBC;
1253 case RISCVCC::COND_NE:
1254 return RISCV::NDS_BBS;
1255 }
1256 break;
1257 case RISCV::Select_GPR_Using_CC_UImm7_NDS:
1258 switch (CC) {
1259 default:
1260 llvm_unreachable("Unexpected condition code!");
1261 case RISCVCC::COND_EQ:
1262 return RISCV::NDS_BEQC;
1263 case RISCVCC::COND_NE:
1264 return RISCV::NDS_BNEC;
1265 }
1266 break;
1267 }
1268}
1269
1271 switch (CC) {
1272 default:
1273 llvm_unreachable("Unrecognized conditional branch");
1274 case RISCVCC::COND_EQ:
1275 return RISCVCC::COND_NE;
1276 case RISCVCC::COND_NE:
1277 return RISCVCC::COND_EQ;
1278 case RISCVCC::COND_LT:
1279 return RISCVCC::COND_GE;
1280 case RISCVCC::COND_GE:
1281 return RISCVCC::COND_LT;
1282 case RISCVCC::COND_LTU:
1283 return RISCVCC::COND_GEU;
1284 case RISCVCC::COND_GEU:
1285 return RISCVCC::COND_LTU;
1286 }
1287}
1288
1291 MachineBasicBlock *&FBB,
1293 bool AllowModify) const {
1294 TBB = FBB = nullptr;
1295 Cond.clear();
1296
1297 // If the block has no terminators, it just falls into the block after it.
1298 MachineBasicBlock::iterator I = MBB.getLastNonDebugInstr();
1299 if (I == MBB.end() || !isUnpredicatedTerminator(*I))
1300 return false;
1301
1302 // Count the number of terminators and find the first unconditional or
1303 // indirect branch.
1304 MachineBasicBlock::iterator FirstUncondOrIndirectBr = MBB.end();
1305 int NumTerminators = 0;
1306 for (auto J = I.getReverse(); J != MBB.rend() && isUnpredicatedTerminator(*J);
1307 J++) {
1308 NumTerminators++;
1309 if (J->getDesc().isUnconditionalBranch() ||
1310 J->getDesc().isIndirectBranch()) {
1311 FirstUncondOrIndirectBr = J.getReverse();
1312 }
1313 }
1314
1315 // If AllowModify is true, we can erase any terminators after
1316 // FirstUncondOrIndirectBR.
1317 if (AllowModify && FirstUncondOrIndirectBr != MBB.end()) {
1318 while (std::next(FirstUncondOrIndirectBr) != MBB.end()) {
1319 std::next(FirstUncondOrIndirectBr)->eraseFromParent();
1320 NumTerminators--;
1321 }
1322 I = FirstUncondOrIndirectBr;
1323 }
1324
1325 // We can't handle blocks that end in an indirect branch.
1326 if (I->getDesc().isIndirectBranch())
1327 return true;
1328
1329 // We can't handle Generic branch opcodes from Global ISel.
1330 if (I->isPreISelOpcode())
1331 return true;
1332
1333 // We can't handle blocks with more than 2 terminators.
1334 if (NumTerminators > 2)
1335 return true;
1336
1337 // Handle a single unconditional branch.
1338 if (NumTerminators == 1 && I->getDesc().isUnconditionalBranch()) {
1340 return false;
1341 }
1342
1343 // Handle a single conditional branch.
1344 if (NumTerminators == 1 && I->getDesc().isConditionalBranch()) {
1346 return false;
1347 }
1348
1349 // Handle a conditional branch followed by an unconditional branch.
1350 if (NumTerminators == 2 && std::prev(I)->getDesc().isConditionalBranch() &&
1351 I->getDesc().isUnconditionalBranch()) {
1352 parseCondBranch(*std::prev(I), TBB, Cond);
1353 FBB = getBranchDestBlock(*I);
1354 return false;
1355 }
1356
1357 // Otherwise, we can't handle this.
1358 return true;
1359}
1360
1362 int *BytesRemoved) const {
1363 if (BytesRemoved)
1364 *BytesRemoved = 0;
1365 MachineBasicBlock::iterator I = MBB.getLastNonDebugInstr();
1366 if (I == MBB.end())
1367 return 0;
1368
1369 if (!I->getDesc().isUnconditionalBranch() &&
1370 !I->getDesc().isConditionalBranch())
1371 return 0;
1372
1373 // Remove the branch.
1374 if (BytesRemoved)
1375 *BytesRemoved += getInstSizeInBytes(*I);
1376 I->eraseFromParent();
1377
1378 I = MBB.end();
1379
1380 if (I == MBB.begin())
1381 return 1;
1382 --I;
1383 if (!I->getDesc().isConditionalBranch())
1384 return 1;
1385
1386 // Remove the branch.
1387 if (BytesRemoved)
1388 *BytesRemoved += getInstSizeInBytes(*I);
1389 I->eraseFromParent();
1390 return 2;
1391}
1392
1393// Inserts a branch into the end of the specific MachineBasicBlock, returning
1394// the number of instructions inserted.
1397 ArrayRef<MachineOperand> Cond, const DebugLoc &DL, int *BytesAdded) const {
1398 if (BytesAdded)
1399 *BytesAdded = 0;
1400
1401 // Shouldn't be a fall through.
1402 assert(TBB && "insertBranch must not be told to insert a fallthrough");
1403 assert((Cond.size() == 3 || Cond.size() == 0) &&
1404 "RISC-V branch conditions have two components!");
1405
1406 // Unconditional branch.
1407 if (Cond.empty()) {
1408 MachineInstr &MI = *BuildMI(&MBB, DL, get(RISCV::PseudoBR)).addMBB(TBB);
1409 if (BytesAdded)
1410 *BytesAdded += getInstSizeInBytes(MI);
1411 return 1;
1412 }
1413
1414 // Either a one or two-way conditional branch.
1415 MachineInstr &CondMI = *BuildMI(&MBB, DL, get(Cond[0].getImm()))
1416 .add(Cond[1])
1417 .add(Cond[2])
1418 .addMBB(TBB);
1419 if (BytesAdded)
1420 *BytesAdded += getInstSizeInBytes(CondMI);
1421
1422 // One-way conditional branch.
1423 if (!FBB)
1424 return 1;
1425
1426 // Two-way conditional branch.
1427 MachineInstr &MI = *BuildMI(&MBB, DL, get(RISCV::PseudoBR)).addMBB(FBB);
1428 if (BytesAdded)
1429 *BytesAdded += getInstSizeInBytes(MI);
1430 return 2;
1431}
1432
1434 MachineBasicBlock &DestBB,
1435 MachineBasicBlock &RestoreBB,
1436 const DebugLoc &DL, int64_t BrOffset,
1437 RegScavenger *RS) const {
1438 assert(RS && "RegScavenger required for long branching");
1439 assert(MBB.empty() &&
1440 "new block should be inserted for expanding unconditional branch");
1441 assert(MBB.pred_size() == 1);
1442 assert(RestoreBB.empty() &&
1443 "restore block should be inserted for restoring clobbered registers");
1444
1445 MachineFunction *MF = MBB.getParent();
1449
1450 if (!isInt<32>(BrOffset))
1452 "Branch offsets outside of the signed 32-bit range not supported");
1453
1454 // FIXME: A virtual register must be used initially, as the register
1455 // scavenger won't work with empty blocks (SIInstrInfo::insertIndirectBranch
1456 // uses the same workaround).
1457 Register ScratchReg = MRI.createVirtualRegister(&RISCV::GPRJALRRegClass);
1458 auto II = MBB.end();
1459 // We may also update the jump target to RestoreBB later.
1460 MachineInstr &MI = *BuildMI(MBB, II, DL, get(RISCV::PseudoJump))
1461 .addReg(ScratchReg, RegState::Define | RegState::Dead)
1462 .addMBB(&DestBB, RISCVII::MO_CALL);
1463
1464 RS->enterBasicBlockEnd(MBB);
1465 const TargetRegisterClass *RC = &RISCV::GPRRegClass;
1466 if (STI.hasStdExtZicfilp())
1467 RC = &RISCV::GPRX7RegClass;
1468 Register TmpGPR =
1469 RS->scavengeRegisterBackwards(*RC, MI.getIterator(),
1470 /*RestoreAfter=*/false, /*SpAdj=*/0,
1471 /*AllowSpill=*/false);
1472 if (TmpGPR.isValid())
1473 RS->setRegUsed(TmpGPR);
1474 else {
1475 // The case when there is no scavenged register needs special handling.
1476
1477 // Pick s11(or s1 for rve) because it doesn't make a difference.
1478 TmpGPR = STI.hasStdExtE() ? RISCV::X9 : RISCV::X27;
1479 // Force t2 if Zicfilp is on
1480 if (STI.hasStdExtZicfilp())
1481 TmpGPR = RISCV::X7;
1482
1483 int FrameIndex = RVFI->getBranchRelaxationScratchFrameIndex();
1484 if (FrameIndex == -1)
1485 report_fatal_error("underestimated function size");
1486
1487 storeRegToStackSlot(MBB, MI, TmpGPR, /*IsKill=*/true, FrameIndex,
1488 &RISCV::GPRRegClass, Register());
1489 TRI->eliminateFrameIndex(std::prev(MI.getIterator()),
1490 /*SpAdj=*/0, /*FIOperandNum=*/1);
1491
1492 MI.getOperand(1).setMBB(&RestoreBB);
1493
1494 loadRegFromStackSlot(RestoreBB, RestoreBB.end(), TmpGPR, FrameIndex,
1495 &RISCV::GPRRegClass, Register());
1496 TRI->eliminateFrameIndex(RestoreBB.back(),
1497 /*SpAdj=*/0, /*FIOperandNum=*/1);
1498 }
1499
1500 MRI.replaceRegWith(ScratchReg, TmpGPR);
1501 MRI.clearVirtRegs();
1502}
1503
1506 assert((Cond.size() == 3) && "Invalid branch condition!");
1507 switch (Cond[0].getImm()) {
1508 default:
1509 llvm_unreachable("Unknown conditional branch!");
1510 case RISCV::BEQ:
1511 Cond[0].setImm(RISCV::BNE);
1512 break;
1513 case RISCV::BEQI:
1514 Cond[0].setImm(RISCV::BNEI);
1515 break;
1516 case RISCV::BNE:
1517 Cond[0].setImm(RISCV::BEQ);
1518 break;
1519 case RISCV::BNEI:
1520 Cond[0].setImm(RISCV::BEQI);
1521 break;
1522 case RISCV::BLT:
1523 Cond[0].setImm(RISCV::BGE);
1524 break;
1525 case RISCV::BGE:
1526 Cond[0].setImm(RISCV::BLT);
1527 break;
1528 case RISCV::BLTU:
1529 Cond[0].setImm(RISCV::BGEU);
1530 break;
1531 case RISCV::BGEU:
1532 Cond[0].setImm(RISCV::BLTU);
1533 break;
1534 case RISCV::CV_BEQIMM:
1535 Cond[0].setImm(RISCV::CV_BNEIMM);
1536 break;
1537 case RISCV::CV_BNEIMM:
1538 Cond[0].setImm(RISCV::CV_BEQIMM);
1539 break;
1540 case RISCV::QC_BEQI:
1541 Cond[0].setImm(RISCV::QC_BNEI);
1542 break;
1543 case RISCV::QC_BNEI:
1544 Cond[0].setImm(RISCV::QC_BEQI);
1545 break;
1546 case RISCV::QC_BGEI:
1547 Cond[0].setImm(RISCV::QC_BLTI);
1548 break;
1549 case RISCV::QC_BLTI:
1550 Cond[0].setImm(RISCV::QC_BGEI);
1551 break;
1552 case RISCV::QC_BGEUI:
1553 Cond[0].setImm(RISCV::QC_BLTUI);
1554 break;
1555 case RISCV::QC_BLTUI:
1556 Cond[0].setImm(RISCV::QC_BGEUI);
1557 break;
1558 case RISCV::QC_E_BEQI:
1559 Cond[0].setImm(RISCV::QC_E_BNEI);
1560 break;
1561 case RISCV::QC_E_BNEI:
1562 Cond[0].setImm(RISCV::QC_E_BEQI);
1563 break;
1564 case RISCV::QC_E_BGEI:
1565 Cond[0].setImm(RISCV::QC_E_BLTI);
1566 break;
1567 case RISCV::QC_E_BLTI:
1568 Cond[0].setImm(RISCV::QC_E_BGEI);
1569 break;
1570 case RISCV::QC_E_BGEUI:
1571 Cond[0].setImm(RISCV::QC_E_BLTUI);
1572 break;
1573 case RISCV::QC_E_BLTUI:
1574 Cond[0].setImm(RISCV::QC_E_BGEUI);
1575 break;
1576 case RISCV::NDS_BBC:
1577 Cond[0].setImm(RISCV::NDS_BBS);
1578 break;
1579 case RISCV::NDS_BBS:
1580 Cond[0].setImm(RISCV::NDS_BBC);
1581 break;
1582 case RISCV::NDS_BEQC:
1583 Cond[0].setImm(RISCV::NDS_BNEC);
1584 break;
1585 case RISCV::NDS_BNEC:
1586 Cond[0].setImm(RISCV::NDS_BEQC);
1587 break;
1588 }
1589
1590 return false;
1591}
1592
1593// Return true if the instruction is a load immediate instruction (i.e.
1594// ADDI x0, imm).
1595static bool isLoadImm(const MachineInstr *MI, int64_t &Imm) {
1596 if (MI->getOpcode() == RISCV::ADDI && MI->getOperand(1).isReg() &&
1597 MI->getOperand(1).getReg() == RISCV::X0) {
1598 Imm = MI->getOperand(2).getImm();
1599 return true;
1600 }
1601 return false;
1602}
1603
1605 const MachineOperand &Op, int64_t &Imm) {
1606 // Either a load from immediate instruction or X0.
1607 if (!Op.isReg())
1608 return false;
1609
1610 Register Reg = Op.getReg();
1611 if (Reg == RISCV::X0) {
1612 Imm = 0;
1613 return true;
1614 }
1615 return Reg.isVirtual() && isLoadImm(MRI.getVRegDef(Reg), Imm);
1616}
1617
1619 bool IsSigned = false;
1620 bool IsEquality = false;
1621 switch (MI.getOpcode()) {
1622 default:
1623 return false;
1624 case RISCV::BEQ:
1625 case RISCV::BNE:
1626 IsEquality = true;
1627 break;
1628 case RISCV::BGE:
1629 case RISCV::BLT:
1630 IsSigned = true;
1631 break;
1632 case RISCV::BGEU:
1633 case RISCV::BLTU:
1634 break;
1635 }
1636
1637 MachineBasicBlock *MBB = MI.getParent();
1638 MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo();
1639
1640 const MachineOperand &LHS = MI.getOperand(0);
1641 const MachineOperand &RHS = MI.getOperand(1);
1642 MachineBasicBlock *TBB = MI.getOperand(2).getMBB();
1643
1644 RISCVCC::CondCode CC = getCondFromBranchOpc(MI.getOpcode());
1646
1647 // Canonicalize conditional branches which can be constant folded into
1648 // beqz or bnez. We can't modify the CFG here.
1649 int64_t C0, C1;
1650 if (isFromLoadImm(MRI, LHS, C0) && isFromLoadImm(MRI, RHS, C1)) {
1651 unsigned NewOpc = evaluateCondBranch(CC, C0, C1) ? RISCV::BEQ : RISCV::BNE;
1652 // Build the new branch and remove the old one.
1653 BuildMI(*MBB, MI, MI.getDebugLoc(), get(NewOpc))
1654 .addReg(RISCV::X0)
1655 .addReg(RISCV::X0)
1656 .addMBB(TBB);
1657 MI.eraseFromParent();
1658 return true;
1659 }
1660
1661 if (IsEquality)
1662 return false;
1663
1664 // For two constants C0 and C1 from
1665 // ```
1666 // li Y, C0
1667 // li Z, C1
1668 // ```
1669 // 1. if C1 = C0 + 1
1670 // we can turn:
1671 // (a) blt Y, X -> bge X, Z
1672 // (b) bge Y, X -> blt X, Z
1673 //
1674 // 2. if C1 = C0 - 1
1675 // we can turn:
1676 // (a) blt X, Y -> bge Z, X
1677 // (b) bge X, Y -> blt Z, X
1678 //
1679 // To make sure this optimization is really beneficial, we only
1680 // optimize for cases where Y had only one use (i.e. only used by the branch).
1681 // Try to find the register for constant Z; return
1682 // invalid register otherwise.
1683 auto searchConst = [&](int64_t C1) -> Register {
1685 auto DefC1 = std::find_if(++II, E, [&](const MachineInstr &I) -> bool {
1686 int64_t Imm;
1687 return isLoadImm(&I, Imm) && Imm == C1 &&
1688 I.getOperand(0).getReg().isVirtual();
1689 });
1690 if (DefC1 != E)
1691 return DefC1->getOperand(0).getReg();
1692
1693 return Register();
1694 };
1695
1696 unsigned NewOpc = RISCVCC::getBrCond(getInverseBranchCondition(CC));
1697
1698 // Might be case 1.
1699 // Don't change 0 to 1 since we can use x0.
1700 // For unsigned cases changing -1U to 0 would be incorrect.
1701 // The incorrect case for signed would be INT_MAX, but isFromLoadImm can't
1702 // return that.
1703 if (isFromLoadImm(MRI, LHS, C0) && C0 != 0 && LHS.getReg().isVirtual() &&
1704 MRI.hasOneUse(LHS.getReg()) && (IsSigned || C0 != -1)) {
1705 assert(isInt<12>(C0) && "Unexpected immediate");
1706 if (Register RegZ = searchConst(C0 + 1)) {
1707 BuildMI(*MBB, MI, MI.getDebugLoc(), get(NewOpc))
1708 .add(RHS)
1709 .addReg(RegZ)
1710 .addMBB(TBB);
1711 // We might extend the live range of Z, clear its kill flag to
1712 // account for this.
1713 MRI.clearKillFlags(RegZ);
1714 MI.eraseFromParent();
1715 return true;
1716 }
1717 }
1718
1719 // Might be case 2.
1720 // For signed cases we don't want to change 0 since we can use x0.
1721 // For unsigned cases changing 0 to -1U would be incorrect.
1722 // The incorrect case for signed would be INT_MIN, but isFromLoadImm can't
1723 // return that.
1724 if (isFromLoadImm(MRI, RHS, C0) && C0 != 0 && RHS.getReg().isVirtual() &&
1725 MRI.hasOneUse(RHS.getReg())) {
1726 assert(isInt<12>(C0) && "Unexpected immediate");
1727 if (Register RegZ = searchConst(C0 - 1)) {
1728 BuildMI(*MBB, MI, MI.getDebugLoc(), get(NewOpc))
1729 .addReg(RegZ)
1730 .add(LHS)
1731 .addMBB(TBB);
1732 // We might extend the live range of Z, clear its kill flag to
1733 // account for this.
1734 MRI.clearKillFlags(RegZ);
1735 MI.eraseFromParent();
1736 return true;
1737 }
1738 }
1739
1740 return false;
1741}
1742
1745 assert(MI.getDesc().isBranch() && "Unexpected opcode!");
1746 // The branch target is always the last operand.
1747 int NumOp = MI.getNumExplicitOperands();
1748 return MI.getOperand(NumOp - 1).getMBB();
1749}
1750
1752 int64_t BrOffset) const {
1753 unsigned XLen = STI.getXLen();
1754 // Ideally we could determine the supported branch offset from the
1755 // RISCVII::FormMask, but this can't be used for Pseudo instructions like
1756 // PseudoBR.
1757 switch (BranchOp) {
1758 default:
1759 llvm_unreachable("Unexpected opcode!");
1760 case RISCV::NDS_BBC:
1761 case RISCV::NDS_BBS:
1762 case RISCV::NDS_BEQC:
1763 case RISCV::NDS_BNEC:
1764 return isInt<11>(BrOffset);
1765 case RISCV::BEQ:
1766 case RISCV::BNE:
1767 case RISCV::BLT:
1768 case RISCV::BGE:
1769 case RISCV::BLTU:
1770 case RISCV::BGEU:
1771 case RISCV::BEQI:
1772 case RISCV::BNEI:
1773 case RISCV::CV_BEQIMM:
1774 case RISCV::CV_BNEIMM:
1775 case RISCV::QC_BEQI:
1776 case RISCV::QC_BNEI:
1777 case RISCV::QC_BGEI:
1778 case RISCV::QC_BLTI:
1779 case RISCV::QC_BLTUI:
1780 case RISCV::QC_BGEUI:
1781 case RISCV::QC_E_BEQI:
1782 case RISCV::QC_E_BNEI:
1783 case RISCV::QC_E_BGEI:
1784 case RISCV::QC_E_BLTI:
1785 case RISCV::QC_E_BLTUI:
1786 case RISCV::QC_E_BGEUI:
1787 return isInt<13>(BrOffset);
1788 case RISCV::JAL:
1789 case RISCV::PseudoBR:
1790 return isInt<21>(BrOffset);
1791 case RISCV::PseudoJump:
1792 return isInt<32>(SignExtend64(BrOffset + 0x800, XLen));
1793 }
1794}
1795
1796// If the operation has a predicated pseudo instruction, return the pseudo
1797// instruction opcode. Otherwise, return RISCV::INSTRUCTION_LIST_END.
1798// TODO: Support more operations.
1799unsigned getPredicatedOpcode(unsigned Opcode) {
1800 // clang-format off
1801 switch (Opcode) {
1802 case RISCV::ADD: return RISCV::PseudoCCADD;
1803 case RISCV::SUB: return RISCV::PseudoCCSUB;
1804 case RISCV::SLL: return RISCV::PseudoCCSLL;
1805 case RISCV::SRL: return RISCV::PseudoCCSRL;
1806 case RISCV::SRA: return RISCV::PseudoCCSRA;
1807 case RISCV::AND: return RISCV::PseudoCCAND;
1808 case RISCV::OR: return RISCV::PseudoCCOR;
1809 case RISCV::XOR: return RISCV::PseudoCCXOR;
1810 case RISCV::MAX: return RISCV::PseudoCCMAX;
1811 case RISCV::MAXU: return RISCV::PseudoCCMAXU;
1812 case RISCV::MIN: return RISCV::PseudoCCMIN;
1813 case RISCV::MINU: return RISCV::PseudoCCMINU;
1814 case RISCV::MUL: return RISCV::PseudoCCMUL;
1815 case RISCV::LUI: return RISCV::PseudoCCLUI;
1816 case RISCV::QC_LI: return RISCV::PseudoCCQC_LI;
1817 case RISCV::QC_E_LI: return RISCV::PseudoCCQC_E_LI;
1818
1819 case RISCV::ADDI: return RISCV::PseudoCCADDI;
1820 case RISCV::SLLI: return RISCV::PseudoCCSLLI;
1821 case RISCV::SRLI: return RISCV::PseudoCCSRLI;
1822 case RISCV::SRAI: return RISCV::PseudoCCSRAI;
1823 case RISCV::ANDI: return RISCV::PseudoCCANDI;
1824 case RISCV::ORI: return RISCV::PseudoCCORI;
1825 case RISCV::XORI: return RISCV::PseudoCCXORI;
1826
1827 case RISCV::ADDW: return RISCV::PseudoCCADDW;
1828 case RISCV::SUBW: return RISCV::PseudoCCSUBW;
1829 case RISCV::SLLW: return RISCV::PseudoCCSLLW;
1830 case RISCV::SRLW: return RISCV::PseudoCCSRLW;
1831 case RISCV::SRAW: return RISCV::PseudoCCSRAW;
1832
1833 case RISCV::ADDIW: return RISCV::PseudoCCADDIW;
1834 case RISCV::SLLIW: return RISCV::PseudoCCSLLIW;
1835 case RISCV::SRLIW: return RISCV::PseudoCCSRLIW;
1836 case RISCV::SRAIW: return RISCV::PseudoCCSRAIW;
1837
1838 case RISCV::ANDN: return RISCV::PseudoCCANDN;
1839 case RISCV::ORN: return RISCV::PseudoCCORN;
1840 case RISCV::XNOR: return RISCV::PseudoCCXNOR;
1841
1842 case RISCV::NDS_BFOS: return RISCV::PseudoCCNDS_BFOS;
1843 case RISCV::NDS_BFOZ: return RISCV::PseudoCCNDS_BFOZ;
1844 }
1845 // clang-format on
1846
1847 return RISCV::INSTRUCTION_LIST_END;
1848}
1849
1850/// Identify instructions that can be folded into a CCMOV instruction, and
1851/// return the defining instruction.
1853 const MachineRegisterInfo &MRI,
1854 const TargetInstrInfo *TII,
1855 const RISCVSubtarget &STI) {
1856 if (!Reg.isVirtual())
1857 return nullptr;
1858 if (!MRI.hasOneNonDBGUse(Reg))
1859 return nullptr;
1860 MachineInstr *MI = MRI.getVRegDef(Reg);
1861 if (!MI)
1862 return nullptr;
1863
1864 if (!STI.hasShortForwardBranchIMinMax() &&
1865 (MI->getOpcode() == RISCV::MAX || MI->getOpcode() == RISCV::MIN ||
1866 MI->getOpcode() == RISCV::MINU || MI->getOpcode() == RISCV::MAXU))
1867 return nullptr;
1868
1869 if (!STI.hasShortForwardBranchIMul() && MI->getOpcode() == RISCV::MUL)
1870 return nullptr;
1871
1872 // Check if MI can be predicated and folded into the CCMOV.
1873 if (getPredicatedOpcode(MI->getOpcode()) == RISCV::INSTRUCTION_LIST_END)
1874 return nullptr;
1875 // Don't predicate li idiom.
1876 if (MI->getOpcode() == RISCV::ADDI && MI->getOperand(1).isReg() &&
1877 MI->getOperand(1).getReg() == RISCV::X0)
1878 return nullptr;
1879 // Check if MI has any other defs or physreg uses.
1880 for (const MachineOperand &MO : llvm::drop_begin(MI->operands())) {
1881 // Reject frame index operands, PEI can't handle the predicated pseudos.
1882 if (MO.isFI() || MO.isCPI() || MO.isJTI())
1883 return nullptr;
1884 if (!MO.isReg())
1885 continue;
1886 // MI can't have any tied operands, that would conflict with predication.
1887 if (MO.isTied())
1888 return nullptr;
1889 if (MO.isDef())
1890 return nullptr;
1891 // Allow constant physregs.
1892 if (MO.getReg().isPhysical() && !MRI.isConstantPhysReg(MO.getReg()))
1893 return nullptr;
1894 }
1895 bool DontMoveAcrossStores = true;
1896 if (!MI->isSafeToMove(DontMoveAcrossStores))
1897 return nullptr;
1898 return MI;
1899}
1900
1903 unsigned &TrueOp, unsigned &FalseOp,
1904 bool &Optimizable) const {
1905 assert(MI.getOpcode() == RISCV::PseudoCCMOVGPR &&
1906 "Unknown select instruction");
1907 // CCMOV operands:
1908 // 0: Def.
1909 // 1: LHS of compare.
1910 // 2: RHS of compare.
1911 // 3: Condition code.
1912 // 4: False use.
1913 // 5: True use.
1914 TrueOp = 5;
1915 FalseOp = 4;
1916 Cond.push_back(MI.getOperand(1));
1917 Cond.push_back(MI.getOperand(2));
1918 Cond.push_back(MI.getOperand(3));
1919 // We can only fold when we support short forward branch opt.
1920 Optimizable = STI.hasShortForwardBranchIALU();
1921 return false;
1922}
1923
1927 bool PreferFalse) const {
1928 assert(MI.getOpcode() == RISCV::PseudoCCMOVGPR &&
1929 "Unknown select instruction");
1930 if (!STI.hasShortForwardBranchIALU())
1931 return nullptr;
1932
1933 MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo();
1935 canFoldAsPredicatedOp(MI.getOperand(5).getReg(), MRI, this, STI);
1936 bool Invert = !DefMI;
1937 if (!DefMI)
1938 DefMI = canFoldAsPredicatedOp(MI.getOperand(4).getReg(), MRI, this, STI);
1939 if (!DefMI)
1940 return nullptr;
1941
1942 // Find new register class to use.
1943 MachineOperand FalseReg = MI.getOperand(Invert ? 5 : 4);
1944 Register DestReg = MI.getOperand(0).getReg();
1945 const TargetRegisterClass *PreviousClass = MRI.getRegClass(FalseReg.getReg());
1946 if (!MRI.constrainRegClass(DestReg, PreviousClass))
1947 return nullptr;
1948
1949 unsigned PredOpc = getPredicatedOpcode(DefMI->getOpcode());
1950 assert(PredOpc != RISCV::INSTRUCTION_LIST_END && "Unexpected opcode!");
1951
1952 // Create a new predicated version of DefMI.
1953 MachineInstrBuilder NewMI =
1954 BuildMI(*MI.getParent(), MI, MI.getDebugLoc(), get(PredOpc), DestReg);
1955
1956 // Copy the condition portion.
1957 NewMI.add(MI.getOperand(1));
1958 NewMI.add(MI.getOperand(2));
1959
1960 // Add condition code, inverting if necessary.
1961 auto CC = static_cast<RISCVCC::CondCode>(MI.getOperand(3).getImm());
1962 if (Invert)
1964 NewMI.addImm(CC);
1965
1966 // Copy the false register.
1967 NewMI.add(FalseReg);
1968
1969 // Copy all the DefMI operands.
1970 const MCInstrDesc &DefDesc = DefMI->getDesc();
1971 for (unsigned i = 1, e = DefDesc.getNumOperands(); i != e; ++i)
1972 NewMI.add(DefMI->getOperand(i));
1973
1974 // Update SeenMIs set: register newly created MI and erase removed DefMI.
1975 SeenMIs.insert(NewMI);
1976 SeenMIs.erase(DefMI);
1977
1978 // If MI is inside a loop, and DefMI is outside the loop, then kill flags on
1979 // DefMI would be invalid when transferred inside the loop. Checking for a
1980 // loop is expensive, but at least remove kill flags if they are in different
1981 // BBs.
1982 if (DefMI->getParent() != MI.getParent())
1983 NewMI->clearKillInfo();
1984
1985 // The caller will erase MI, but not DefMI.
1986 DefMI->eraseFromParent();
1987 return NewMI;
1988}
1989
1991 if (MI.isMetaInstruction())
1992 return 0;
1993
1994 unsigned Opcode = MI.getOpcode();
1995
1996 if (Opcode == TargetOpcode::INLINEASM ||
1997 Opcode == TargetOpcode::INLINEASM_BR) {
1998 const MachineFunction &MF = *MI.getParent()->getParent();
1999 return getInlineAsmLength(MI.getOperand(0).getSymbolName(),
2000 *MF.getTarget().getMCAsmInfo());
2001 }
2002
2003 if (!MI.memoperands_empty()) {
2004 MachineMemOperand *MMO = *(MI.memoperands_begin());
2005 if (STI.hasStdExtZihintntl() && MMO->isNonTemporal()) {
2006 if (STI.hasStdExtZca()) {
2007 if (isCompressibleInst(MI, STI))
2008 return 4; // c.ntl.all + c.load/c.store
2009 return 6; // c.ntl.all + load/store
2010 }
2011 return 8; // ntl.all + load/store
2012 }
2013 }
2014
2015 if (Opcode == TargetOpcode::BUNDLE)
2016 return getInstBundleLength(MI);
2017
2018 if (MI.getParent() && MI.getParent()->getParent()) {
2019 if (isCompressibleInst(MI, STI))
2020 return 2;
2021 }
2022
2023 switch (Opcode) {
2024 case RISCV::PseudoMV_FPR16INX:
2025 case RISCV::PseudoMV_FPR32INX:
2026 // MV is always compressible to either c.mv or c.li rd, 0.
2027 return STI.hasStdExtZca() ? 2 : 4;
2028 case TargetOpcode::STACKMAP:
2029 // The upper bound for a stackmap intrinsic is the full length of its shadow
2031 case TargetOpcode::PATCHPOINT:
2032 // The size of the patchpoint intrinsic is the number of bytes requested
2034 case TargetOpcode::STATEPOINT: {
2035 // The size of the statepoint intrinsic is the number of bytes requested
2036 unsigned NumBytes = StatepointOpers(&MI).getNumPatchBytes();
2037 // No patch bytes means at most a PseudoCall is emitted
2038 return std::max(NumBytes, 8U);
2039 }
2040 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
2041 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
2042 case TargetOpcode::PATCHABLE_TAIL_CALL: {
2043 const MachineFunction &MF = *MI.getParent()->getParent();
2044 const Function &F = MF.getFunction();
2045 if (Opcode == TargetOpcode::PATCHABLE_FUNCTION_ENTER &&
2046 F.hasFnAttribute("patchable-function-entry")) {
2047 unsigned Num;
2048 if (F.getFnAttribute("patchable-function-entry")
2049 .getValueAsString()
2050 .getAsInteger(10, Num))
2051 return get(Opcode).getSize();
2052
2053 // Number of C.NOP or NOP
2054 return (STI.hasStdExtZca() ? 2 : 4) * Num;
2055 }
2056 // XRay uses C.JAL + 21 or 33 C.NOP for each sled in RV32 and RV64,
2057 // respectively.
2058 return STI.is64Bit() ? 68 : 44;
2059 }
2060 default:
2061 return get(Opcode).getSize();
2062 }
2063}
2064
2065unsigned RISCVInstrInfo::getInstBundleLength(const MachineInstr &MI) const {
2066 unsigned Size = 0;
2068 MachineBasicBlock::const_instr_iterator E = MI.getParent()->instr_end();
2069 while (++I != E && I->isInsideBundle()) {
2070 assert(!I->isBundle() && "No nested bundle!");
2072 }
2073 return Size;
2074}
2075
2077 const unsigned Opcode = MI.getOpcode();
2078 switch (Opcode) {
2079 default:
2080 break;
2081 case RISCV::FSGNJ_D:
2082 case RISCV::FSGNJ_S:
2083 case RISCV::FSGNJ_H:
2084 case RISCV::FSGNJ_D_INX:
2085 case RISCV::FSGNJ_D_IN32X:
2086 case RISCV::FSGNJ_S_INX:
2087 case RISCV::FSGNJ_H_INX:
2088 // The canonical floating-point move is fsgnj rd, rs, rs.
2089 return MI.getOperand(1).isReg() && MI.getOperand(2).isReg() &&
2090 MI.getOperand(1).getReg() == MI.getOperand(2).getReg();
2091 case RISCV::ADDI:
2092 case RISCV::ORI:
2093 case RISCV::XORI:
2094 return (MI.getOperand(1).isReg() &&
2095 MI.getOperand(1).getReg() == RISCV::X0) ||
2096 (MI.getOperand(2).isImm() && MI.getOperand(2).getImm() == 0);
2097 }
2098 return MI.isAsCheapAsAMove();
2099}
2100
2101std::optional<DestSourcePair>
2103 if (MI.isMoveReg())
2104 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2105 switch (MI.getOpcode()) {
2106 default:
2107 break;
2108 case RISCV::ADD:
2109 case RISCV::OR:
2110 case RISCV::XOR:
2111 if (MI.getOperand(1).isReg() && MI.getOperand(1).getReg() == RISCV::X0 &&
2112 MI.getOperand(2).isReg())
2113 return DestSourcePair{MI.getOperand(0), MI.getOperand(2)};
2114 if (MI.getOperand(2).isReg() && MI.getOperand(2).getReg() == RISCV::X0 &&
2115 MI.getOperand(1).isReg())
2116 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2117 break;
2118 case RISCV::ADDI:
2119 // Operand 1 can be a frameindex but callers expect registers
2120 if (MI.getOperand(1).isReg() && MI.getOperand(2).isImm() &&
2121 MI.getOperand(2).getImm() == 0)
2122 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2123 break;
2124 case RISCV::SUB:
2125 if (MI.getOperand(2).isReg() && MI.getOperand(2).getReg() == RISCV::X0 &&
2126 MI.getOperand(1).isReg())
2127 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2128 break;
2129 case RISCV::SH1ADD:
2130 case RISCV::SH1ADD_UW:
2131 case RISCV::SH2ADD:
2132 case RISCV::SH2ADD_UW:
2133 case RISCV::SH3ADD:
2134 case RISCV::SH3ADD_UW:
2135 if (MI.getOperand(1).isReg() && MI.getOperand(1).getReg() == RISCV::X0 &&
2136 MI.getOperand(2).isReg())
2137 return DestSourcePair{MI.getOperand(0), MI.getOperand(2)};
2138 break;
2139 case RISCV::FSGNJ_D:
2140 case RISCV::FSGNJ_S:
2141 case RISCV::FSGNJ_H:
2142 case RISCV::FSGNJ_D_INX:
2143 case RISCV::FSGNJ_D_IN32X:
2144 case RISCV::FSGNJ_S_INX:
2145 case RISCV::FSGNJ_H_INX:
2146 // The canonical floating-point move is fsgnj rd, rs, rs.
2147 if (MI.getOperand(1).isReg() && MI.getOperand(2).isReg() &&
2148 MI.getOperand(1).getReg() == MI.getOperand(2).getReg())
2149 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2150 break;
2151 }
2152 return std::nullopt;
2153}
2154
2156 if (ForceMachineCombinerStrategy.getNumOccurrences() == 0) {
2157 // The option is unused. Choose Local strategy only for in-order cores. When
2158 // scheduling model is unspecified, use MinInstrCount strategy as more
2159 // generic one.
2160 const auto &SchedModel = STI.getSchedModel();
2161 return (!SchedModel.hasInstrSchedModel() || SchedModel.isOutOfOrder())
2164 }
2165 // The strategy was forced by the option.
2167}
2168
2170 MachineInstr &Root, unsigned &Pattern,
2171 SmallVectorImpl<MachineInstr *> &InsInstrs) const {
2172 int16_t FrmOpIdx =
2173 RISCV::getNamedOperandIdx(Root.getOpcode(), RISCV::OpName::frm);
2174 if (FrmOpIdx < 0) {
2175 assert(all_of(InsInstrs,
2176 [](MachineInstr *MI) {
2177 return RISCV::getNamedOperandIdx(MI->getOpcode(),
2178 RISCV::OpName::frm) < 0;
2179 }) &&
2180 "New instructions require FRM whereas the old one does not have it");
2181 return;
2182 }
2183
2184 const MachineOperand &FRM = Root.getOperand(FrmOpIdx);
2185 MachineFunction &MF = *Root.getMF();
2186
2187 for (auto *NewMI : InsInstrs) {
2188 // We'd already added the FRM operand.
2189 if (static_cast<unsigned>(RISCV::getNamedOperandIdx(
2190 NewMI->getOpcode(), RISCV::OpName::frm)) != NewMI->getNumOperands())
2191 continue;
2192 MachineInstrBuilder MIB(MF, NewMI);
2193 MIB.add(FRM);
2194 if (FRM.getImm() == RISCVFPRndMode::DYN)
2195 MIB.addUse(RISCV::FRM, RegState::Implicit);
2196 }
2197}
2198
2199static bool isFADD(unsigned Opc) {
2200 switch (Opc) {
2201 default:
2202 return false;
2203 case RISCV::FADD_H:
2204 case RISCV::FADD_S:
2205 case RISCV::FADD_D:
2206 return true;
2207 }
2208}
2209
2210static bool isFSUB(unsigned Opc) {
2211 switch (Opc) {
2212 default:
2213 return false;
2214 case RISCV::FSUB_H:
2215 case RISCV::FSUB_S:
2216 case RISCV::FSUB_D:
2217 return true;
2218 }
2219}
2220
2221static bool isFMUL(unsigned Opc) {
2222 switch (Opc) {
2223 default:
2224 return false;
2225 case RISCV::FMUL_H:
2226 case RISCV::FMUL_S:
2227 case RISCV::FMUL_D:
2228 return true;
2229 }
2230}
2231
2232bool RISCVInstrInfo::isVectorAssociativeAndCommutative(const MachineInstr &Inst,
2233 bool Invert) const {
2234#define OPCODE_LMUL_CASE(OPC) \
2235 case RISCV::OPC##_M1: \
2236 case RISCV::OPC##_M2: \
2237 case RISCV::OPC##_M4: \
2238 case RISCV::OPC##_M8: \
2239 case RISCV::OPC##_MF2: \
2240 case RISCV::OPC##_MF4: \
2241 case RISCV::OPC##_MF8
2242
2243#define OPCODE_LMUL_MASK_CASE(OPC) \
2244 case RISCV::OPC##_M1_MASK: \
2245 case RISCV::OPC##_M2_MASK: \
2246 case RISCV::OPC##_M4_MASK: \
2247 case RISCV::OPC##_M8_MASK: \
2248 case RISCV::OPC##_MF2_MASK: \
2249 case RISCV::OPC##_MF4_MASK: \
2250 case RISCV::OPC##_MF8_MASK
2251
2252 unsigned Opcode = Inst.getOpcode();
2253 if (Invert) {
2254 if (auto InvOpcode = getInverseOpcode(Opcode))
2255 Opcode = *InvOpcode;
2256 else
2257 return false;
2258 }
2259
2260 // clang-format off
2261 switch (Opcode) {
2262 default:
2263 return false;
2264 OPCODE_LMUL_CASE(PseudoVADD_VV):
2265 OPCODE_LMUL_MASK_CASE(PseudoVADD_VV):
2266 OPCODE_LMUL_CASE(PseudoVMUL_VV):
2267 OPCODE_LMUL_MASK_CASE(PseudoVMUL_VV):
2268 return true;
2269 }
2270 // clang-format on
2271
2272#undef OPCODE_LMUL_MASK_CASE
2273#undef OPCODE_LMUL_CASE
2274}
2275
2276bool RISCVInstrInfo::areRVVInstsReassociable(const MachineInstr &Root,
2277 const MachineInstr &Prev) const {
2278 if (!areOpcodesEqualOrInverse(Root.getOpcode(), Prev.getOpcode()))
2279 return false;
2280
2281 assert(Root.getMF() == Prev.getMF());
2282 const MachineRegisterInfo *MRI = &Root.getMF()->getRegInfo();
2283 const TargetRegisterInfo *TRI = MRI->getTargetRegisterInfo();
2284
2285 // Make sure vtype operands are also the same.
2286 const MCInstrDesc &Desc = get(Root.getOpcode());
2287 const uint64_t TSFlags = Desc.TSFlags;
2288
2289 auto checkImmOperand = [&](unsigned OpIdx) {
2290 return Root.getOperand(OpIdx).getImm() == Prev.getOperand(OpIdx).getImm();
2291 };
2292
2293 auto checkRegOperand = [&](unsigned OpIdx) {
2294 return Root.getOperand(OpIdx).getReg() == Prev.getOperand(OpIdx).getReg();
2295 };
2296
2297 // PassThru
2298 // TODO: Potentially we can loosen the condition to consider Root to be
2299 // associable with Prev if Root has NoReg as passthru. In which case we
2300 // also need to loosen the condition on vector policies between these.
2301 if (!checkRegOperand(1))
2302 return false;
2303
2304 // SEW
2305 if (RISCVII::hasSEWOp(TSFlags) &&
2306 !checkImmOperand(RISCVII::getSEWOpNum(Desc)))
2307 return false;
2308
2309 // Mask
2310 if (RISCVII::usesMaskPolicy(TSFlags)) {
2311 const MachineBasicBlock *MBB = Root.getParent();
2314 Register MI1VReg;
2315
2316 bool SeenMI2 = false;
2317 for (auto End = MBB->rend(), It = It1; It != End; ++It) {
2318 if (It == It2) {
2319 SeenMI2 = true;
2320 if (!MI1VReg.isValid())
2321 // There is no V0 def between Root and Prev; they're sharing the
2322 // same V0.
2323 break;
2324 }
2325
2326 if (It->modifiesRegister(RISCV::V0, TRI)) {
2327 Register SrcReg = It->getOperand(1).getReg();
2328 // If it's not VReg it'll be more difficult to track its defs, so
2329 // bailing out here just to be safe.
2330 if (!SrcReg.isVirtual())
2331 return false;
2332
2333 if (!MI1VReg.isValid()) {
2334 // This is the V0 def for Root.
2335 MI1VReg = SrcReg;
2336 continue;
2337 }
2338
2339 // Some random mask updates.
2340 if (!SeenMI2)
2341 continue;
2342
2343 // This is the V0 def for Prev; check if it's the same as that of
2344 // Root.
2345 if (MI1VReg != SrcReg)
2346 return false;
2347 else
2348 break;
2349 }
2350 }
2351
2352 // If we haven't encountered Prev, it's likely that this function was
2353 // called in a wrong way (e.g. Root is before Prev).
2354 assert(SeenMI2 && "Prev is expected to appear before Root");
2355 }
2356
2357 // Tail / Mask policies
2358 if (RISCVII::hasVecPolicyOp(TSFlags) &&
2359 !checkImmOperand(RISCVII::getVecPolicyOpNum(Desc)))
2360 return false;
2361
2362 // VL
2363 if (RISCVII::hasVLOp(TSFlags)) {
2364 unsigned OpIdx = RISCVII::getVLOpNum(Desc);
2365 const MachineOperand &Op1 = Root.getOperand(OpIdx);
2366 const MachineOperand &Op2 = Prev.getOperand(OpIdx);
2367 if (Op1.getType() != Op2.getType())
2368 return false;
2369 switch (Op1.getType()) {
2371 if (Op1.getReg() != Op2.getReg())
2372 return false;
2373 break;
2375 if (Op1.getImm() != Op2.getImm())
2376 return false;
2377 break;
2378 default:
2379 llvm_unreachable("Unrecognized VL operand type");
2380 }
2381 }
2382
2383 // Rounding modes
2384 if (RISCVII::hasRoundModeOp(TSFlags) &&
2385 !checkImmOperand(RISCVII::getVLOpNum(Desc) - 1))
2386 return false;
2387
2388 return true;
2389}
2390
2391// Most of our RVV pseudos have passthru operand, so the real operands
2392// start from index = 2.
2393bool RISCVInstrInfo::hasReassociableVectorSibling(const MachineInstr &Inst,
2394 bool &Commuted) const {
2395 const MachineBasicBlock *MBB = Inst.getParent();
2396 const MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo();
2398 "Expect the present of passthrough operand.");
2399 MachineInstr *MI1 = MRI.getUniqueVRegDef(Inst.getOperand(2).getReg());
2400 MachineInstr *MI2 = MRI.getUniqueVRegDef(Inst.getOperand(3).getReg());
2401
2402 // If only one operand has the same or inverse opcode and it's the second
2403 // source operand, the operands must be commuted.
2404 Commuted = !areRVVInstsReassociable(Inst, *MI1) &&
2405 areRVVInstsReassociable(Inst, *MI2);
2406 if (Commuted)
2407 std::swap(MI1, MI2);
2408
2409 return areRVVInstsReassociable(Inst, *MI1) &&
2410 (isVectorAssociativeAndCommutative(*MI1) ||
2411 isVectorAssociativeAndCommutative(*MI1, /* Invert */ true)) &&
2413 MRI.hasOneNonDBGUse(MI1->getOperand(0).getReg());
2414}
2415
2417 const MachineInstr &Inst, const MachineBasicBlock *MBB) const {
2418 if (!isVectorAssociativeAndCommutative(Inst) &&
2419 !isVectorAssociativeAndCommutative(Inst, /*Invert=*/true))
2421
2422 const MachineOperand &Op1 = Inst.getOperand(2);
2423 const MachineOperand &Op2 = Inst.getOperand(3);
2424 const MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo();
2425
2426 // We need virtual register definitions for the operands that we will
2427 // reassociate.
2428 MachineInstr *MI1 = nullptr;
2429 MachineInstr *MI2 = nullptr;
2430 if (Op1.isReg() && Op1.getReg().isVirtual())
2431 MI1 = MRI.getUniqueVRegDef(Op1.getReg());
2432 if (Op2.isReg() && Op2.getReg().isVirtual())
2433 MI2 = MRI.getUniqueVRegDef(Op2.getReg());
2434
2435 // And at least one operand must be defined in MBB.
2436 return MI1 && MI2 && (MI1->getParent() == MBB || MI2->getParent() == MBB);
2437}
2438
2440 const MachineInstr &Root, unsigned Pattern,
2441 std::array<unsigned, 5> &OperandIndices) const {
2443 if (RISCV::getRVVMCOpcode(Root.getOpcode())) {
2444 // Skip the passthrough operand, so increment all indices by one.
2445 for (unsigned I = 0; I < 5; ++I)
2446 ++OperandIndices[I];
2447 }
2448}
2449
2451 bool &Commuted) const {
2452 if (isVectorAssociativeAndCommutative(Inst) ||
2453 isVectorAssociativeAndCommutative(Inst, /*Invert=*/true))
2454 return hasReassociableVectorSibling(Inst, Commuted);
2455
2456 if (!TargetInstrInfo::hasReassociableSibling(Inst, Commuted))
2457 return false;
2458
2459 const MachineRegisterInfo &MRI = Inst.getMF()->getRegInfo();
2460 unsigned OperandIdx = Commuted ? 2 : 1;
2461 const MachineInstr &Sibling =
2462 *MRI.getVRegDef(Inst.getOperand(OperandIdx).getReg());
2463
2464 int16_t InstFrmOpIdx =
2465 RISCV::getNamedOperandIdx(Inst.getOpcode(), RISCV::OpName::frm);
2466 int16_t SiblingFrmOpIdx =
2467 RISCV::getNamedOperandIdx(Sibling.getOpcode(), RISCV::OpName::frm);
2468
2469 return (InstFrmOpIdx < 0 && SiblingFrmOpIdx < 0) ||
2470 RISCV::hasEqualFRM(Inst, Sibling);
2471}
2472
2474 bool Invert) const {
2475 if (isVectorAssociativeAndCommutative(Inst, Invert))
2476 return true;
2477
2478 unsigned Opc = Inst.getOpcode();
2479 if (Invert) {
2480 auto InverseOpcode = getInverseOpcode(Opc);
2481 if (!InverseOpcode)
2482 return false;
2483 Opc = *InverseOpcode;
2484 }
2485
2486 if (isFADD(Opc) || isFMUL(Opc))
2489
2490 switch (Opc) {
2491 default:
2492 return false;
2493 case RISCV::ADD:
2494 case RISCV::ADDW:
2495 case RISCV::AND:
2496 case RISCV::OR:
2497 case RISCV::XOR:
2498 // From RISC-V ISA spec, if both the high and low bits of the same product
2499 // are required, then the recommended code sequence is:
2500 //
2501 // MULH[[S]U] rdh, rs1, rs2
2502 // MUL rdl, rs1, rs2
2503 // (source register specifiers must be in same order and rdh cannot be the
2504 // same as rs1 or rs2)
2505 //
2506 // Microarchitectures can then fuse these into a single multiply operation
2507 // instead of performing two separate multiplies.
2508 // MachineCombiner may reassociate MUL operands and lose the fusion
2509 // opportunity.
2510 case RISCV::MUL:
2511 case RISCV::MULW:
2512 case RISCV::MIN:
2513 case RISCV::MINU:
2514 case RISCV::MAX:
2515 case RISCV::MAXU:
2516 case RISCV::FMIN_H:
2517 case RISCV::FMIN_S:
2518 case RISCV::FMIN_D:
2519 case RISCV::FMAX_H:
2520 case RISCV::FMAX_S:
2521 case RISCV::FMAX_D:
2522 return true;
2523 }
2524
2525 return false;
2526}
2527
2528std::optional<unsigned>
2529RISCVInstrInfo::getInverseOpcode(unsigned Opcode) const {
2530#define RVV_OPC_LMUL_CASE(OPC, INV) \
2531 case RISCV::OPC##_M1: \
2532 return RISCV::INV##_M1; \
2533 case RISCV::OPC##_M2: \
2534 return RISCV::INV##_M2; \
2535 case RISCV::OPC##_M4: \
2536 return RISCV::INV##_M4; \
2537 case RISCV::OPC##_M8: \
2538 return RISCV::INV##_M8; \
2539 case RISCV::OPC##_MF2: \
2540 return RISCV::INV##_MF2; \
2541 case RISCV::OPC##_MF4: \
2542 return RISCV::INV##_MF4; \
2543 case RISCV::OPC##_MF8: \
2544 return RISCV::INV##_MF8
2545
2546#define RVV_OPC_LMUL_MASK_CASE(OPC, INV) \
2547 case RISCV::OPC##_M1_MASK: \
2548 return RISCV::INV##_M1_MASK; \
2549 case RISCV::OPC##_M2_MASK: \
2550 return RISCV::INV##_M2_MASK; \
2551 case RISCV::OPC##_M4_MASK: \
2552 return RISCV::INV##_M4_MASK; \
2553 case RISCV::OPC##_M8_MASK: \
2554 return RISCV::INV##_M8_MASK; \
2555 case RISCV::OPC##_MF2_MASK: \
2556 return RISCV::INV##_MF2_MASK; \
2557 case RISCV::OPC##_MF4_MASK: \
2558 return RISCV::INV##_MF4_MASK; \
2559 case RISCV::OPC##_MF8_MASK: \
2560 return RISCV::INV##_MF8_MASK
2561
2562 switch (Opcode) {
2563 default:
2564 return std::nullopt;
2565 case RISCV::FADD_H:
2566 return RISCV::FSUB_H;
2567 case RISCV::FADD_S:
2568 return RISCV::FSUB_S;
2569 case RISCV::FADD_D:
2570 return RISCV::FSUB_D;
2571 case RISCV::FSUB_H:
2572 return RISCV::FADD_H;
2573 case RISCV::FSUB_S:
2574 return RISCV::FADD_S;
2575 case RISCV::FSUB_D:
2576 return RISCV::FADD_D;
2577 case RISCV::ADD:
2578 return RISCV::SUB;
2579 case RISCV::SUB:
2580 return RISCV::ADD;
2581 case RISCV::ADDW:
2582 return RISCV::SUBW;
2583 case RISCV::SUBW:
2584 return RISCV::ADDW;
2585 // clang-format off
2586 RVV_OPC_LMUL_CASE(PseudoVADD_VV, PseudoVSUB_VV);
2587 RVV_OPC_LMUL_MASK_CASE(PseudoVADD_VV, PseudoVSUB_VV);
2588 RVV_OPC_LMUL_CASE(PseudoVSUB_VV, PseudoVADD_VV);
2589 RVV_OPC_LMUL_MASK_CASE(PseudoVSUB_VV, PseudoVADD_VV);
2590 // clang-format on
2591 }
2592
2593#undef RVV_OPC_LMUL_MASK_CASE
2594#undef RVV_OPC_LMUL_CASE
2595}
2596
2598 const MachineOperand &MO,
2599 bool DoRegPressureReduce) {
2600 if (!MO.isReg() || !MO.getReg().isVirtual())
2601 return false;
2602 const MachineRegisterInfo &MRI = Root.getMF()->getRegInfo();
2603 MachineInstr *MI = MRI.getVRegDef(MO.getReg());
2604 if (!MI || !isFMUL(MI->getOpcode()))
2605 return false;
2606
2609 return false;
2610
2611 // Try combining even if fmul has more than one use as it eliminates
2612 // dependency between fadd(fsub) and fmul. However, it can extend liveranges
2613 // for fmul operands, so reject the transformation in register pressure
2614 // reduction mode.
2615 if (DoRegPressureReduce && !MRI.hasOneNonDBGUse(MI->getOperand(0).getReg()))
2616 return false;
2617
2618 // Do not combine instructions from different basic blocks.
2619 if (Root.getParent() != MI->getParent())
2620 return false;
2621 return RISCV::hasEqualFRM(Root, *MI);
2622}
2623
2625 SmallVectorImpl<unsigned> &Patterns,
2626 bool DoRegPressureReduce) {
2627 unsigned Opc = Root.getOpcode();
2628 bool IsFAdd = isFADD(Opc);
2629 if (!IsFAdd && !isFSUB(Opc))
2630 return false;
2631 bool Added = false;
2632 if (canCombineFPFusedMultiply(Root, Root.getOperand(1),
2633 DoRegPressureReduce)) {
2636 Added = true;
2637 }
2638 if (canCombineFPFusedMultiply(Root, Root.getOperand(2),
2639 DoRegPressureReduce)) {
2642 Added = true;
2643 }
2644 return Added;
2645}
2646
2647static bool getFPPatterns(MachineInstr &Root,
2648 SmallVectorImpl<unsigned> &Patterns,
2649 bool DoRegPressureReduce) {
2650 return getFPFusedMultiplyPatterns(Root, Patterns, DoRegPressureReduce);
2651}
2652
2653/// Utility routine that checks if \param MO is defined by an
2654/// \param CombineOpc instruction in the basic block \param MBB
2656 const MachineOperand &MO,
2657 unsigned CombineOpc) {
2658 const MachineRegisterInfo &MRI = MBB.getParent()->getRegInfo();
2659 const MachineInstr *MI = nullptr;
2660
2661 if (MO.isReg() && MO.getReg().isVirtual())
2662 MI = MRI.getUniqueVRegDef(MO.getReg());
2663 // And it needs to be in the trace (otherwise, it won't have a depth).
2664 if (!MI || MI->getParent() != &MBB || MI->getOpcode() != CombineOpc)
2665 return nullptr;
2666 // Must only used by the user we combine with.
2667 if (!MRI.hasOneNonDBGUse(MI->getOperand(0).getReg()))
2668 return nullptr;
2669
2670 return MI;
2671}
2672
2673/// Utility routine that checks if \param MO is defined by a SLLI in \param
2674/// MBB that can be combined by splitting across 2 SHXADD instructions. The
2675/// first SHXADD shift amount is given by \param OuterShiftAmt.
2677 const MachineOperand &MO,
2678 unsigned OuterShiftAmt) {
2679 const MachineInstr *ShiftMI = canCombine(MBB, MO, RISCV::SLLI);
2680 if (!ShiftMI)
2681 return false;
2682
2683 unsigned InnerShiftAmt = ShiftMI->getOperand(2).getImm();
2684 if (InnerShiftAmt < OuterShiftAmt || (InnerShiftAmt - OuterShiftAmt) > 3)
2685 return false;
2686
2687 return true;
2688}
2689
2690// Returns the shift amount from a SHXADD instruction. Returns 0 if the
2691// instruction is not a SHXADD.
2692static unsigned getSHXADDShiftAmount(unsigned Opc) {
2693 switch (Opc) {
2694 default:
2695 return 0;
2696 case RISCV::SH1ADD:
2697 return 1;
2698 case RISCV::SH2ADD:
2699 return 2;
2700 case RISCV::SH3ADD:
2701 return 3;
2702 }
2703}
2704
2705// Returns the shift amount from a SHXADD.UW instruction. Returns 0 if the
2706// instruction is not a SHXADD.UW.
2707static unsigned getSHXADDUWShiftAmount(unsigned Opc) {
2708 switch (Opc) {
2709 default:
2710 return 0;
2711 case RISCV::SH1ADD_UW:
2712 return 1;
2713 case RISCV::SH2ADD_UW:
2714 return 2;
2715 case RISCV::SH3ADD_UW:
2716 return 3;
2717 }
2718}
2719
2720// Look for opportunities to combine (sh3add Z, (add X, (slli Y, 5))) into
2721// (sh3add (sh2add Y, Z), X).
2722static bool getSHXADDPatterns(const MachineInstr &Root,
2723 SmallVectorImpl<unsigned> &Patterns) {
2724 unsigned ShiftAmt = getSHXADDShiftAmount(Root.getOpcode());
2725 if (!ShiftAmt)
2726 return false;
2727
2728 const MachineBasicBlock &MBB = *Root.getParent();
2729
2730 const MachineInstr *AddMI = canCombine(MBB, Root.getOperand(2), RISCV::ADD);
2731 if (!AddMI)
2732 return false;
2733
2734 bool Found = false;
2735 if (canCombineShiftIntoShXAdd(MBB, AddMI->getOperand(1), ShiftAmt)) {
2737 Found = true;
2738 }
2739 if (canCombineShiftIntoShXAdd(MBB, AddMI->getOperand(2), ShiftAmt)) {
2741 Found = true;
2742 }
2743
2744 return Found;
2745}
2746
2758
2760 MachineInstr &Root, SmallVectorImpl<unsigned> &Patterns,
2761 bool DoRegPressureReduce) const {
2762
2763 if (getFPPatterns(Root, Patterns, DoRegPressureReduce))
2764 return true;
2765
2766 if (getSHXADDPatterns(Root, Patterns))
2767 return true;
2768
2769 return TargetInstrInfo::getMachineCombinerPatterns(Root, Patterns,
2770 DoRegPressureReduce);
2771}
2772
2773static unsigned getFPFusedMultiplyOpcode(unsigned RootOpc, unsigned Pattern) {
2774 switch (RootOpc) {
2775 default:
2776 llvm_unreachable("Unexpected opcode");
2777 case RISCV::FADD_H:
2778 return RISCV::FMADD_H;
2779 case RISCV::FADD_S:
2780 return RISCV::FMADD_S;
2781 case RISCV::FADD_D:
2782 return RISCV::FMADD_D;
2783 case RISCV::FSUB_H:
2784 return Pattern == RISCVMachineCombinerPattern::FMSUB ? RISCV::FMSUB_H
2785 : RISCV::FNMSUB_H;
2786 case RISCV::FSUB_S:
2787 return Pattern == RISCVMachineCombinerPattern::FMSUB ? RISCV::FMSUB_S
2788 : RISCV::FNMSUB_S;
2789 case RISCV::FSUB_D:
2790 return Pattern == RISCVMachineCombinerPattern::FMSUB ? RISCV::FMSUB_D
2791 : RISCV::FNMSUB_D;
2792 }
2793}
2794
2795static unsigned getAddendOperandIdx(unsigned Pattern) {
2796 switch (Pattern) {
2797 default:
2798 llvm_unreachable("Unexpected pattern");
2801 return 2;
2804 return 1;
2805 }
2806}
2807
2809 unsigned Pattern,
2812 MachineFunction *MF = Root.getMF();
2815
2816 MachineOperand &Mul1 = Prev.getOperand(1);
2817 MachineOperand &Mul2 = Prev.getOperand(2);
2818 MachineOperand &Dst = Root.getOperand(0);
2820
2821 Register DstReg = Dst.getReg();
2822 unsigned FusedOpc = getFPFusedMultiplyOpcode(Root.getOpcode(), Pattern);
2823 uint32_t IntersectedFlags = Root.getFlags() & Prev.getFlags();
2824 DebugLoc MergedLoc =
2826
2827 bool Mul1IsKill = Mul1.isKill();
2828 bool Mul2IsKill = Mul2.isKill();
2829 bool AddendIsKill = Addend.isKill();
2830
2831 // We need to clear kill flags since we may be extending the live range past
2832 // a kill. If the mul had kill flags, we can preserve those since we know
2833 // where the previous range stopped.
2834 MRI.clearKillFlags(Mul1.getReg());
2835 MRI.clearKillFlags(Mul2.getReg());
2836
2838 BuildMI(*MF, MergedLoc, TII->get(FusedOpc), DstReg)
2839 .addReg(Mul1.getReg(), getKillRegState(Mul1IsKill))
2840 .addReg(Mul2.getReg(), getKillRegState(Mul2IsKill))
2841 .addReg(Addend.getReg(), getKillRegState(AddendIsKill))
2842 .setMIFlags(IntersectedFlags);
2843
2844 InsInstrs.push_back(MIB);
2845 if (MRI.hasOneNonDBGUse(Prev.getOperand(0).getReg()))
2846 DelInstrs.push_back(&Prev);
2847 DelInstrs.push_back(&Root);
2848}
2849
2850// Combine patterns like (sh3add Z, (add X, (slli Y, 5))) to
2851// (sh3add (sh2add Y, Z), X) if the shift amount can be split across two
2852// shXadd instructions. The outer shXadd keeps its original opcode.
2853static void
2854genShXAddAddShift(MachineInstr &Root, unsigned AddOpIdx,
2857 DenseMap<Register, unsigned> &InstrIdxForVirtReg) {
2858 MachineFunction *MF = Root.getMF();
2861
2862 unsigned OuterShiftAmt = getSHXADDShiftAmount(Root.getOpcode());
2863 assert(OuterShiftAmt != 0 && "Unexpected opcode");
2864
2865 MachineInstr *AddMI = MRI.getUniqueVRegDef(Root.getOperand(2).getReg());
2866 MachineInstr *ShiftMI =
2867 MRI.getUniqueVRegDef(AddMI->getOperand(AddOpIdx).getReg());
2868
2869 unsigned InnerShiftAmt = ShiftMI->getOperand(2).getImm();
2870 assert(InnerShiftAmt >= OuterShiftAmt && "Unexpected shift amount");
2871
2872 unsigned InnerOpc;
2873 switch (InnerShiftAmt - OuterShiftAmt) {
2874 default:
2875 llvm_unreachable("Unexpected shift amount");
2876 case 0:
2877 InnerOpc = RISCV::ADD;
2878 break;
2879 case 1:
2880 InnerOpc = RISCV::SH1ADD;
2881 break;
2882 case 2:
2883 InnerOpc = RISCV::SH2ADD;
2884 break;
2885 case 3:
2886 InnerOpc = RISCV::SH3ADD;
2887 break;
2888 }
2889
2890 const MachineOperand &X = AddMI->getOperand(3 - AddOpIdx);
2891 const MachineOperand &Y = ShiftMI->getOperand(1);
2892 const MachineOperand &Z = Root.getOperand(1);
2893
2894 Register NewVR = MRI.createVirtualRegister(&RISCV::GPRRegClass);
2895
2896 auto MIB1 = BuildMI(*MF, MIMetadata(Root), TII->get(InnerOpc), NewVR)
2897 .addReg(Y.getReg(), getKillRegState(Y.isKill()))
2898 .addReg(Z.getReg(), getKillRegState(Z.isKill()));
2899 auto MIB2 = BuildMI(*MF, MIMetadata(Root), TII->get(Root.getOpcode()),
2900 Root.getOperand(0).getReg())
2901 .addReg(NewVR, RegState::Kill)
2902 .addReg(X.getReg(), getKillRegState(X.isKill()));
2903
2904 InstrIdxForVirtReg.insert(std::make_pair(NewVR, 0));
2905 InsInstrs.push_back(MIB1);
2906 InsInstrs.push_back(MIB2);
2907 DelInstrs.push_back(ShiftMI);
2908 DelInstrs.push_back(AddMI);
2909 DelInstrs.push_back(&Root);
2910}
2911
2913 MachineInstr &Root, unsigned Pattern,
2916 DenseMap<Register, unsigned> &InstrIdxForVirtReg) const {
2918 switch (Pattern) {
2919 default:
2921 DelInstrs, InstrIdxForVirtReg);
2922 return;
2925 MachineInstr &Prev = *MRI.getVRegDef(Root.getOperand(1).getReg());
2926 combineFPFusedMultiply(Root, Prev, Pattern, InsInstrs, DelInstrs);
2927 return;
2928 }
2931 MachineInstr &Prev = *MRI.getVRegDef(Root.getOperand(2).getReg());
2932 combineFPFusedMultiply(Root, Prev, Pattern, InsInstrs, DelInstrs);
2933 return;
2934 }
2936 genShXAddAddShift(Root, 1, InsInstrs, DelInstrs, InstrIdxForVirtReg);
2937 return;
2939 genShXAddAddShift(Root, 2, InsInstrs, DelInstrs, InstrIdxForVirtReg);
2940 return;
2941 }
2942}
2943
2945 StringRef &ErrInfo) const {
2946 MCInstrDesc const &Desc = MI.getDesc();
2947
2948 for (const auto &[Index, Operand] : enumerate(Desc.operands())) {
2949 const MachineOperand &MO = MI.getOperand(Index);
2950 unsigned OpType = Operand.OperandType;
2951 switch (OpType) {
2952 default:
2953 if (OpType >= RISCVOp::OPERAND_FIRST_RISCV_IMM &&
2955 if (!MO.isImm()) {
2956 ErrInfo = "Expected an immediate operand.";
2957 return false;
2958 }
2959 int64_t Imm = MO.getImm();
2960 bool Ok;
2961 switch (OpType) {
2962 default:
2963 llvm_unreachable("Unexpected operand type");
2964
2965 // clang-format off
2966#define CASE_OPERAND_UIMM(NUM) \
2967 case RISCVOp::OPERAND_UIMM##NUM: \
2968 Ok = isUInt<NUM>(Imm); \
2969 break;
2970#define CASE_OPERAND_SIMM(NUM) \
2971 case RISCVOp::OPERAND_SIMM##NUM: \
2972 Ok = isInt<NUM>(Imm); \
2973 break;
2989 // clang-format on
2991 Ok = isShiftedUInt<1, 1>(Imm);
2992 break;
2994 Ok = isShiftedUInt<4, 1>(Imm);
2995 break;
2997 Ok = isUInt<5>(Imm) && (Imm != 0);
2998 break;
3000 Ok = isUInt<5>(Imm) && (Imm > 3);
3001 break;
3003 Ok = Imm >= 1 && Imm <= 32;
3004 break;
3006 Ok = isShiftedUInt<5, 1>(Imm);
3007 break;
3009 Ok = isShiftedUInt<5, 2>(Imm);
3010 break;
3012 Ok = isShiftedUInt<4, 3>(Imm);
3013 break;
3015 Ok = isShiftedUInt<6, 2>(Imm);
3016 break;
3018 Ok = isShiftedUInt<5, 3>(Imm);
3019 break;
3021 Ok = isUInt<8>(Imm) && Imm >= 32;
3022 break;
3024 Ok = isShiftedUInt<6, 3>(Imm);
3025 break;
3027 Ok = isInt<8>(Imm);
3028 break;
3030 Ok = isShiftedInt<6, 4>(Imm) && (Imm != 0);
3031 break;
3033 Ok = isShiftedUInt<8, 2>(Imm) && (Imm != 0);
3034 break;
3036 Ok = isUInt<16>(Imm) && (Imm != 0);
3037 break;
3039 Ok = Imm == 3;
3040 break;
3042 Ok = Imm == 4;
3043 break;
3045 Ok = (isUInt<5>(Imm) && Imm != 0) || Imm == -1;
3046 break;
3047 // clang-format off
3053 // clang-format on
3055 Ok = Imm >= -15 && Imm <= 16;
3056 break;
3058 Ok = isInt<5>(Imm) && (Imm != 0);
3059 break;
3061 Ok = Imm != 0 && isInt<6>(Imm);
3062 break;
3064 Ok = isUInt<10>(Imm);
3065 break;
3067 Ok = isUInt<11>(Imm);
3068 break;
3070 Ok = isShiftedInt<7, 5>(Imm);
3071 break;
3073 Ok = isInt<16>(Imm) && (Imm != 0);
3074 break;
3076 Ok = isInt<20>(Imm);
3077 break;
3079 Ok = STI.is64Bit() ? isUInt<6>(Imm) : isUInt<5>(Imm);
3080 break;
3082 Ok = STI.is64Bit() ? isUInt<6>(Imm) : isUInt<5>(Imm);
3083 Ok = Ok && Imm != 0;
3084 break;
3086 Ok = (isUInt<5>(Imm) && Imm != 0) || (Imm >= 0xfffe0 && Imm <= 0xfffff);
3087 break;
3089 Ok = Imm >= 0 && Imm <= 10;
3090 break;
3092 Ok = Imm >= 0 && Imm <= 7;
3093 break;
3095 Ok = Imm >= 1 && Imm <= 10;
3096 break;
3098 Ok = Imm >= 2 && Imm <= 14;
3099 break;
3101 Ok = Imm >= RISCVZC::RA && Imm <= RISCVZC::RA_S0_S11;
3102 break;
3104 Ok = Imm >= RISCVZC::RA_S0 && Imm <= RISCVZC::RA_S0_S11;
3105 break;
3107 Ok = Imm >= 0 && Imm <= 48 && Imm % 16 == 0;
3108 break;
3111 break;
3113 Ok = Imm == RISCVFPRndMode::RTZ;
3114 break;
3116 Ok = Imm >= 0 && Imm < RISCVCC::COND_INVALID;
3117 break;
3119 Ok = isValidAtomicOrdering(Imm);
3120 break;
3123 Imm;
3124 break;
3126 Ok = (isUInt<5>(Imm) && RISCVVType::isValidSEW(1 << Imm));
3127 break;
3129 Ok = Imm == 0;
3130 break;
3133 if (RISCVII::usesVXRM(Desc.TSFlags))
3134 Ok = isUInt<2>(Imm);
3135 else
3137 break;
3140 break;
3142 Ok = Imm == 1 || Imm == 2 || Imm == 4;
3143 break;
3144 }
3145 if (!Ok) {
3146 ErrInfo = "Invalid immediate";
3147 return false;
3148 }
3149 }
3150 break;
3152 // TODO: We could be stricter about what non-register operands are
3153 // allowed.
3154 if (MO.isReg()) {
3155 ErrInfo = "Expected a non-register operand.";
3156 return false;
3157 }
3158 if (MO.isImm() && !isInt<12>(MO.getImm())) {
3159 ErrInfo = "Invalid immediate";
3160 return false;
3161 }
3162 break;
3165 // TODO: We could be stricter about what non-register operands are
3166 // allowed.
3167 if (MO.isReg()) {
3168 ErrInfo = "Expected a non-register operand.";
3169 return false;
3170 }
3171 if (MO.isImm() && !isUInt<20>(MO.getImm())) {
3172 ErrInfo = "Invalid immediate";
3173 return false;
3174 }
3175 break;
3177 // TODO: We could be stricter about what non-register operands are
3178 // allowed.
3179 if (MO.isReg()) {
3180 ErrInfo = "Expected a non-register operand.";
3181 return false;
3182 }
3183 if (MO.isImm() && !isInt<32>(MO.getImm())) {
3184 ErrInfo = "Invalid immediate";
3185 return false;
3186 }
3187 break;
3189 if (MO.isImm()) {
3190 int64_t Imm = MO.getImm();
3191 // VLMAX is represented as -1.
3192 if (!isUInt<5>(Imm) && Imm != -1) {
3193 ErrInfo = "Invalid immediate";
3194 return false;
3195 }
3196 } else if (!MO.isReg()) {
3197 ErrInfo = "Expected a register or immediate operand.";
3198 return false;
3199 }
3200 break;
3201 }
3202 }
3203
3204 const uint64_t TSFlags = Desc.TSFlags;
3205 if (RISCVII::hasVLOp(TSFlags)) {
3206 const MachineOperand &Op = MI.getOperand(RISCVII::getVLOpNum(Desc));
3207 if (!Op.isImm() && !Op.isReg()) {
3208 ErrInfo = "Invalid operand type for VL operand";
3209 return false;
3210 }
3211 if (Op.isReg() && Op.getReg().isValid()) {
3212 const MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo();
3213 auto *RC = MRI.getRegClass(Op.getReg());
3214 if (!RISCV::GPRNoX0RegClass.hasSubClassEq(RC)) {
3215 ErrInfo = "Invalid register class for VL operand";
3216 return false;
3217 }
3218 }
3219 if (!RISCVII::hasSEWOp(TSFlags)) {
3220 ErrInfo = "VL operand w/o SEW operand?";
3221 return false;
3222 }
3223 }
3224 if (RISCVII::hasSEWOp(TSFlags)) {
3225 unsigned OpIdx = RISCVII::getSEWOpNum(Desc);
3226 if (!MI.getOperand(OpIdx).isImm()) {
3227 ErrInfo = "SEW value expected to be an immediate";
3228 return false;
3229 }
3230 uint64_t Log2SEW = MI.getOperand(OpIdx).getImm();
3231 if (Log2SEW > 31) {
3232 ErrInfo = "Unexpected SEW value";
3233 return false;
3234 }
3235 unsigned SEW = Log2SEW ? 1 << Log2SEW : 8;
3236 if (!RISCVVType::isValidSEW(SEW)) {
3237 ErrInfo = "Unexpected SEW value";
3238 return false;
3239 }
3240 }
3241 if (RISCVII::hasVecPolicyOp(TSFlags)) {
3243 if (!MI.getOperand(OpIdx).isImm()) {
3244 ErrInfo = "Policy operand expected to be an immediate";
3245 return false;
3246 }
3247 uint64_t Policy = MI.getOperand(OpIdx).getImm();
3249 ErrInfo = "Invalid Policy Value";
3250 return false;
3251 }
3252 if (!RISCVII::hasVLOp(TSFlags)) {
3253 ErrInfo = "policy operand w/o VL operand?";
3254 return false;
3255 }
3256
3257 // VecPolicy operands can only exist on instructions with passthru/merge
3258 // arguments. Note that not all arguments with passthru have vec policy
3259 // operands- some instructions have implicit policies.
3260 unsigned UseOpIdx;
3261 if (!MI.isRegTiedToUseOperand(0, &UseOpIdx)) {
3262 ErrInfo = "policy operand w/o tied operand?";
3263 return false;
3264 }
3265 }
3266
3267 if (int Idx = RISCVII::getFRMOpNum(Desc);
3268 Idx >= 0 && MI.getOperand(Idx).getImm() == RISCVFPRndMode::DYN &&
3269 !MI.readsRegister(RISCV::FRM, /*TRI=*/nullptr)) {
3270 ErrInfo = "dynamic rounding mode should read FRM";
3271 return false;
3272 }
3273
3274 return true;
3275}
3276
3278 const MachineInstr &AddrI,
3279 ExtAddrMode &AM) const {
3280 switch (MemI.getOpcode()) {
3281 default:
3282 return false;
3283 case RISCV::LB:
3284 case RISCV::LBU:
3285 case RISCV::LH:
3286 case RISCV::LH_INX:
3287 case RISCV::LHU:
3288 case RISCV::LW:
3289 case RISCV::LW_INX:
3290 case RISCV::LWU:
3291 case RISCV::LD:
3292 case RISCV::LD_RV32:
3293 case RISCV::FLH:
3294 case RISCV::FLW:
3295 case RISCV::FLD:
3296 case RISCV::SB:
3297 case RISCV::SH:
3298 case RISCV::SH_INX:
3299 case RISCV::SW:
3300 case RISCV::SW_INX:
3301 case RISCV::SD:
3302 case RISCV::SD_RV32:
3303 case RISCV::FSH:
3304 case RISCV::FSW:
3305 case RISCV::FSD:
3306 break;
3307 }
3308
3309 if (MemI.getOperand(0).getReg() == Reg)
3310 return false;
3311
3312 if (AddrI.getOpcode() != RISCV::ADDI || !AddrI.getOperand(1).isReg() ||
3313 !AddrI.getOperand(2).isImm())
3314 return false;
3315
3316 int64_t OldOffset = MemI.getOperand(2).getImm();
3317 int64_t Disp = AddrI.getOperand(2).getImm();
3318 int64_t NewOffset = OldOffset + Disp;
3319 if (!STI.is64Bit())
3320 NewOffset = SignExtend64<32>(NewOffset);
3321
3322 if (!isInt<12>(NewOffset))
3323 return false;
3324
3325 AM.BaseReg = AddrI.getOperand(1).getReg();
3326 AM.ScaledReg = 0;
3327 AM.Scale = 0;
3328 AM.Displacement = NewOffset;
3330 return true;
3331}
3332
3334 const ExtAddrMode &AM) const {
3335
3336 const DebugLoc &DL = MemI.getDebugLoc();
3337 MachineBasicBlock &MBB = *MemI.getParent();
3338
3339 assert(AM.ScaledReg == 0 && AM.Scale == 0 &&
3340 "Addressing mode not supported for folding");
3341
3342 return BuildMI(MBB, MemI, DL, get(MemI.getOpcode()))
3343 .addReg(MemI.getOperand(0).getReg(),
3344 MemI.mayLoad() ? RegState::Define : 0)
3345 .addReg(AM.BaseReg)
3346 .addImm(AM.Displacement)
3347 .setMemRefs(MemI.memoperands())
3348 .setMIFlags(MemI.getFlags());
3349}
3350
3351// TODO: At the moment, MIPS introduced paring of instructions operating with
3352// word or double word. This should be extended with more instructions when more
3353// vendors support load/store pairing.
3355 switch (Opc) {
3356 default:
3357 return false;
3358 case RISCV::SW:
3359 case RISCV::SD:
3360 case RISCV::LD:
3361 case RISCV::LW:
3362 return true;
3363 }
3364}
3365
3367 const TargetRegisterInfo *TRI) {
3368 // If this is a volatile load/store, don't mess with it.
3369 if (LdSt.hasOrderedMemoryRef() || LdSt.getNumExplicitOperands() != 3)
3370 return false;
3371
3372 if (LdSt.getOperand(1).isFI())
3373 return true;
3374
3375 assert(LdSt.getOperand(1).isReg() && "Expected a reg operand.");
3376 // Can't cluster if the instruction modifies the base register
3377 // or it is update form. e.g. ld x5,8(x5)
3378 if (LdSt.modifiesRegister(LdSt.getOperand(1).getReg(), TRI))
3379 return false;
3380
3381 if (!LdSt.getOperand(2).isImm())
3382 return false;
3383
3384 return true;
3385}
3386
3389 int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width,
3390 const TargetRegisterInfo *TRI) const {
3391 if (!LdSt.mayLoadOrStore())
3392 return false;
3393
3394 // Conservatively, only handle scalar loads/stores for now.
3395 switch (LdSt.getOpcode()) {
3396 case RISCV::LB:
3397 case RISCV::LBU:
3398 case RISCV::SB:
3399 case RISCV::LH:
3400 case RISCV::LH_INX:
3401 case RISCV::LHU:
3402 case RISCV::FLH:
3403 case RISCV::SH:
3404 case RISCV::SH_INX:
3405 case RISCV::FSH:
3406 case RISCV::LW:
3407 case RISCV::LW_INX:
3408 case RISCV::LWU:
3409 case RISCV::FLW:
3410 case RISCV::SW:
3411 case RISCV::SW_INX:
3412 case RISCV::FSW:
3413 case RISCV::LD:
3414 case RISCV::LD_RV32:
3415 case RISCV::FLD:
3416 case RISCV::SD:
3417 case RISCV::SD_RV32:
3418 case RISCV::FSD:
3419 break;
3420 default:
3421 return false;
3422 }
3423 const MachineOperand *BaseOp;
3424 OffsetIsScalable = false;
3425 if (!getMemOperandWithOffsetWidth(LdSt, BaseOp, Offset, Width, TRI))
3426 return false;
3427 BaseOps.push_back(BaseOp);
3428 return true;
3429}
3430
3431// TODO: This was copied from SIInstrInfo. Could it be lifted to a common
3432// helper?
3435 const MachineInstr &MI2,
3437 // Only examine the first "base" operand of each instruction, on the
3438 // assumption that it represents the real base address of the memory access.
3439 // Other operands are typically offsets or indices from this base address.
3440 if (BaseOps1.front()->isIdenticalTo(*BaseOps2.front()))
3441 return true;
3442
3443 if (!MI1.hasOneMemOperand() || !MI2.hasOneMemOperand())
3444 return false;
3445
3446 auto MO1 = *MI1.memoperands_begin();
3447 auto MO2 = *MI2.memoperands_begin();
3448 if (MO1->getAddrSpace() != MO2->getAddrSpace())
3449 return false;
3450
3451 auto Base1 = MO1->getValue();
3452 auto Base2 = MO2->getValue();
3453 if (!Base1 || !Base2)
3454 return false;
3455 Base1 = getUnderlyingObject(Base1);
3456 Base2 = getUnderlyingObject(Base2);
3457
3458 if (isa<UndefValue>(Base1) || isa<UndefValue>(Base2))
3459 return false;
3460
3461 return Base1 == Base2;
3462}
3463
3465 ArrayRef<const MachineOperand *> BaseOps1, int64_t Offset1,
3466 bool OffsetIsScalable1, ArrayRef<const MachineOperand *> BaseOps2,
3467 int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize,
3468 unsigned NumBytes) const {
3469 // If the mem ops (to be clustered) do not have the same base ptr, then they
3470 // should not be clustered
3471 if (!BaseOps1.empty() && !BaseOps2.empty()) {
3472 const MachineInstr &FirstLdSt = *BaseOps1.front()->getParent();
3473 const MachineInstr &SecondLdSt = *BaseOps2.front()->getParent();
3474 if (!memOpsHaveSameBasePtr(FirstLdSt, BaseOps1, SecondLdSt, BaseOps2))
3475 return false;
3476 } else if (!BaseOps1.empty() || !BaseOps2.empty()) {
3477 // If only one base op is empty, they do not have the same base ptr
3478 return false;
3479 }
3480
3481 unsigned CacheLineSize =
3482 BaseOps1.front()->getParent()->getMF()->getSubtarget().getCacheLineSize();
3483 // Assume a cache line size of 64 bytes if no size is set in RISCVSubtarget.
3485 // Cluster if the memory operations are on the same or a neighbouring cache
3486 // line, but limit the maximum ClusterSize to avoid creating too much
3487 // additional register pressure.
3488 return ClusterSize <= 4 && std::abs(Offset1 - Offset2) < CacheLineSize;
3489}
3490
3491// Set BaseReg (the base register operand), Offset (the byte offset being
3492// accessed) and the access Width of the passed instruction that reads/writes
3493// memory. Returns false if the instruction does not read/write memory or the
3494// BaseReg/Offset/Width can't be determined. Is not guaranteed to always
3495// recognise base operands and offsets in all cases.
3496// TODO: Add an IsScalable bool ref argument (like the equivalent AArch64
3497// function) and set it as appropriate.
3499 const MachineInstr &LdSt, const MachineOperand *&BaseReg, int64_t &Offset,
3500 LocationSize &Width, const TargetRegisterInfo *TRI) const {
3501 if (!LdSt.mayLoadOrStore())
3502 return false;
3503
3504 // Here we assume the standard RISC-V ISA, which uses a base+offset
3505 // addressing mode. You'll need to relax these conditions to support custom
3506 // load/store instructions.
3507 if (LdSt.getNumExplicitOperands() != 3)
3508 return false;
3509 if ((!LdSt.getOperand(1).isReg() && !LdSt.getOperand(1).isFI()) ||
3510 !LdSt.getOperand(2).isImm())
3511 return false;
3512
3513 if (!LdSt.hasOneMemOperand())
3514 return false;
3515
3516 Width = (*LdSt.memoperands_begin())->getSize();
3517 BaseReg = &LdSt.getOperand(1);
3518 Offset = LdSt.getOperand(2).getImm();
3519 return true;
3520}
3521
3523 const MachineInstr &MIa, const MachineInstr &MIb) const {
3524 assert(MIa.mayLoadOrStore() && "MIa must be a load or store.");
3525 assert(MIb.mayLoadOrStore() && "MIb must be a load or store.");
3526
3529 return false;
3530
3531 // Retrieve the base register, offset from the base register and width. Width
3532 // is the size of memory that is being loaded/stored (e.g. 1, 2, 4). If
3533 // base registers are identical, and the offset of a lower memory access +
3534 // the width doesn't overlap the offset of a higher memory access,
3535 // then the memory accesses are different.
3536 const TargetRegisterInfo *TRI = STI.getRegisterInfo();
3537 const MachineOperand *BaseOpA = nullptr, *BaseOpB = nullptr;
3538 int64_t OffsetA = 0, OffsetB = 0;
3540 WidthB = LocationSize::precise(0);
3541 if (getMemOperandWithOffsetWidth(MIa, BaseOpA, OffsetA, WidthA, TRI) &&
3542 getMemOperandWithOffsetWidth(MIb, BaseOpB, OffsetB, WidthB, TRI)) {
3543 if (BaseOpA->isIdenticalTo(*BaseOpB)) {
3544 int LowOffset = std::min(OffsetA, OffsetB);
3545 int HighOffset = std::max(OffsetA, OffsetB);
3546 LocationSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
3547 if (LowWidth.hasValue() &&
3548 LowOffset + (int)LowWidth.getValue() <= HighOffset)
3549 return true;
3550 }
3551 }
3552 return false;
3553}
3554
3555std::pair<unsigned, unsigned>
3557 const unsigned Mask = RISCVII::MO_DIRECT_FLAG_MASK;
3558 return std::make_pair(TF & Mask, TF & ~Mask);
3559}
3560
3563 using namespace RISCVII;
3564 static const std::pair<unsigned, const char *> TargetFlags[] = {
3565 {MO_CALL, "riscv-call"},
3566 {MO_LO, "riscv-lo"},
3567 {MO_HI, "riscv-hi"},
3568 {MO_PCREL_LO, "riscv-pcrel-lo"},
3569 {MO_PCREL_HI, "riscv-pcrel-hi"},
3570 {MO_GOT_HI, "riscv-got-hi"},
3571 {MO_TPREL_LO, "riscv-tprel-lo"},
3572 {MO_TPREL_HI, "riscv-tprel-hi"},
3573 {MO_TPREL_ADD, "riscv-tprel-add"},
3574 {MO_TLS_GOT_HI, "riscv-tls-got-hi"},
3575 {MO_TLS_GD_HI, "riscv-tls-gd-hi"},
3576 {MO_TLSDESC_HI, "riscv-tlsdesc-hi"},
3577 {MO_TLSDESC_LOAD_LO, "riscv-tlsdesc-load-lo"},
3578 {MO_TLSDESC_ADD_LO, "riscv-tlsdesc-add-lo"},
3579 {MO_TLSDESC_CALL, "riscv-tlsdesc-call"}};
3580 return ArrayRef(TargetFlags);
3581}
3583 MachineFunction &MF, bool OutlineFromLinkOnceODRs) const {
3584 const Function &F = MF.getFunction();
3585
3586 // Can F be deduplicated by the linker? If it can, don't outline from it.
3587 if (!OutlineFromLinkOnceODRs && F.hasLinkOnceODRLinkage())
3588 return false;
3589
3590 // Don't outline from functions with section markings; the program could
3591 // expect that all the code is in the named section.
3592 if (F.hasSection())
3593 return false;
3594
3595 // It's safe to outline from MF.
3596 return true;
3597}
3598
3600 unsigned &Flags) const {
3601 // More accurate safety checking is done in getOutliningCandidateInfo.
3603}
3604
3605// Enum values indicating how an outlined call should be constructed.
3610
3615
3617 const MachineFunction *MF = MBB.getParent();
3618 const Function &F = MF->getFunction();
3619 return F.getFnAttribute("fentry-call").getValueAsBool() ||
3620 F.hasFnAttribute("patchable-function-entry");
3621}
3622
3624 MCRegister RegNo) {
3625 return MI.readsRegister(RegNo, TRI) ||
3626 MI.getDesc().hasImplicitUseOfPhysReg(RegNo);
3627}
3628
3630 const TargetRegisterInfo *TRI, MCRegister RegNo) {
3631 return MI.modifiesRegister(RegNo, TRI) ||
3632 MI.getDesc().hasImplicitDefOfPhysReg(RegNo);
3633}
3634
3636 if (!MBB.back().isReturn())
3637 return true;
3639 return true;
3640
3641 // If the candidate reads the pre-set register
3642 // that can be used for expanding PseudoTAIL instruction,
3643 // then we cannot insert tail call.
3644 const TargetSubtargetInfo &STI = MBB.getParent()->getSubtarget();
3645 MCRegister TailExpandUseRegNo =
3647 for (const MachineInstr &MI : MBB) {
3648 if (isMIReadsReg(MI, STI.getRegisterInfo(), TailExpandUseRegNo))
3649 return true;
3650 if (isMIModifiesReg(MI, STI.getRegisterInfo(), TailExpandUseRegNo))
3651 break;
3652 }
3653 return false;
3654}
3655
3657 // If last instruction is return then we can rely on
3658 // the verification already performed in the getOutliningTypeImpl.
3659 if (C.back().isReturn()) {
3660 assert(!cannotInsertTailCall(*C.getMBB()) &&
3661 "The candidate who uses return instruction must be outlined "
3662 "using tail call");
3663 return false;
3664 }
3665
3666 // Filter out candidates where the X5 register (t0) can't be used to setup
3667 // the function call.
3668 const TargetRegisterInfo *TRI = C.getMF()->getSubtarget().getRegisterInfo();
3669 if (llvm::any_of(C, [TRI](const MachineInstr &MI) {
3670 return isMIModifiesReg(MI, TRI, RISCV::X5);
3671 }))
3672 return true;
3673
3674 return !C.isAvailableAcrossAndOutOfSeq(RISCV::X5, *TRI);
3675}
3676
3677std::optional<std::unique_ptr<outliner::OutlinedFunction>>
3679 const MachineModuleInfo &MMI,
3680 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
3681 unsigned MinRepeats) const {
3682
3683 // Analyze each candidate and erase the ones that are not viable.
3684 llvm::erase_if(RepeatedSequenceLocs, analyzeCandidate);
3685
3686 // If the sequence doesn't have enough candidates left, then we're done.
3687 if (RepeatedSequenceLocs.size() < MinRepeats)
3688 return std::nullopt;
3689
3690 // Each RepeatedSequenceLoc is identical.
3691 outliner::Candidate &Candidate = RepeatedSequenceLocs[0];
3692 unsigned InstrSizeCExt =
3693 Candidate.getMF()->getSubtarget<RISCVSubtarget>().hasStdExtZca() ? 2 : 4;
3694 unsigned CallOverhead = 0, FrameOverhead = 0;
3695
3696 // Count the number of CFI instructions in the candidate, if present.
3697 unsigned CFICount = 0;
3698 for (auto &I : Candidate) {
3699 if (I.isCFIInstruction())
3700 CFICount++;
3701 }
3702
3703 // Ensure CFI coverage matches: comparing the number of CFIs in the candidate
3704 // with the total number of CFIs in the parent function for each candidate.
3705 // Outlining only a subset of a function’s CFIs would split the unwind state
3706 // across two code regions and lead to incorrect address offsets between the
3707 // outlined body and the remaining code. To preserve correct unwind info, we
3708 // only outline when all CFIs in the function can be outlined together.
3709 for (outliner::Candidate &C : RepeatedSequenceLocs) {
3710 std::vector<MCCFIInstruction> CFIInstructions =
3711 C.getMF()->getFrameInstructions();
3712
3713 if (CFICount > 0 && CFICount != CFIInstructions.size())
3714 return std::nullopt;
3715 }
3716
3718 if (Candidate.back().isReturn()) {
3720 // tail call = auipc + jalr in the worst case without linker relaxation.
3721 // FIXME: This code suggests the JALR can be compressed - how?
3722 CallOverhead = 4 + InstrSizeCExt;
3723 // Using tail call we move ret instruction from caller to callee.
3724 FrameOverhead = 0;
3725 } else {
3726 // call t0, function = 8 bytes.
3727 CallOverhead = 8;
3728 // jr t0 = 4 bytes, 2 bytes if compressed instructions are enabled.
3729 FrameOverhead = InstrSizeCExt;
3730 }
3731
3732 // If we have CFI instructions, we can only outline if the outlined section
3733 // can be a tail call.
3734 if (MOCI != MachineOutlinerTailCall && CFICount > 0)
3735 return std::nullopt;
3736
3737 for (auto &C : RepeatedSequenceLocs)
3738 C.setCallInfo(MOCI, CallOverhead);
3739
3740 unsigned SequenceSize = 0;
3741 for (auto &MI : Candidate)
3742 SequenceSize += getInstSizeInBytes(MI);
3743
3744 return std::make_unique<outliner::OutlinedFunction>(
3745 RepeatedSequenceLocs, SequenceSize, FrameOverhead, MOCI);
3746}
3747
3751 unsigned Flags) const {
3752 MachineInstr &MI = *MBBI;
3753 MachineBasicBlock *MBB = MI.getParent();
3754 const TargetRegisterInfo *TRI =
3755 MBB->getParent()->getSubtarget().getRegisterInfo();
3756 const auto &F = MI.getMF()->getFunction();
3757
3758 // We can only outline CFI instructions if we will tail call the outlined
3759 // function, or fix up the CFI offsets. Currently, CFI instructions are
3760 // outlined only if in a tail call.
3761 if (MI.isCFIInstruction())
3763
3764 if (cannotInsertTailCall(*MBB) &&
3765 (MI.isReturn() || isMIModifiesReg(MI, TRI, RISCV::X5)))
3767
3768 // Make sure the operands don't reference something unsafe.
3769 for (const auto &MO : MI.operands()) {
3770
3771 // pcrel-hi and pcrel-lo can't put in separate sections, filter that out
3772 // if any possible.
3773 if (MO.getTargetFlags() == RISCVII::MO_PCREL_LO &&
3774 (MI.getMF()->getTarget().getFunctionSections() || F.hasComdat() ||
3775 F.hasSection() || F.getSectionPrefix()))
3777 }
3778
3779 if (isLPAD(MI))
3781
3783}
3784
3787 const outliner::OutlinedFunction &OF) const {
3788
3789 if (OF.FrameConstructionID == MachineOutlinerTailCall)
3790 return;
3791
3792 MBB.addLiveIn(RISCV::X5);
3793
3794 // Add in a return instruction to the end of the outlined frame.
3795 MBB.insert(MBB.end(), BuildMI(MF, DebugLoc(), get(RISCV::JALR))
3796 .addReg(RISCV::X0, RegState::Define)
3797 .addReg(RISCV::X5)
3798 .addImm(0));
3799}
3800
3804
3805 if (C.CallConstructionID == MachineOutlinerTailCall) {
3806 It = MBB.insert(It, BuildMI(MF, DebugLoc(), get(RISCV::PseudoTAIL))
3807 .addGlobalAddress(M.getNamedValue(MF.getName()),
3808 /*Offset=*/0, RISCVII::MO_CALL));
3809 return It;
3810 }
3811
3812 // Add in a call instruction to the outlined function at the given location.
3813 It = MBB.insert(It,
3814 BuildMI(MF, DebugLoc(), get(RISCV::PseudoCALLReg), RISCV::X5)
3815 .addGlobalAddress(M.getNamedValue(MF.getName()), 0,
3817 return It;
3818}
3819
3820std::optional<RegImmPair> RISCVInstrInfo::isAddImmediate(const MachineInstr &MI,
3821 Register Reg) const {
3822 // TODO: Handle cases where Reg is a super- or sub-register of the
3823 // destination register.
3824 const MachineOperand &Op0 = MI.getOperand(0);
3825 if (!Op0.isReg() || Reg != Op0.getReg())
3826 return std::nullopt;
3827
3828 // Don't consider ADDIW as a candidate because the caller may not be aware
3829 // of its sign extension behaviour.
3830 if (MI.getOpcode() == RISCV::ADDI && MI.getOperand(1).isReg() &&
3831 MI.getOperand(2).isImm())
3832 return RegImmPair{MI.getOperand(1).getReg(), MI.getOperand(2).getImm()};
3833
3834 return std::nullopt;
3835}
3836
3837// MIR printer helper function to annotate Operands with a comment.
3839 const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx,
3840 const TargetRegisterInfo *TRI) const {
3841 // Print a generic comment for this operand if there is one.
3842 std::string GenericComment =
3844 if (!GenericComment.empty())
3845 return GenericComment;
3846
3847 // If not, we must have an immediate operand.
3848 if (!Op.isImm())
3849 return std::string();
3850
3851 const MCInstrDesc &Desc = MI.getDesc();
3852 if (OpIdx >= Desc.getNumOperands())
3853 return std::string();
3854
3855 std::string Comment;
3856 raw_string_ostream OS(Comment);
3857
3858 const MCOperandInfo &OpInfo = Desc.operands()[OpIdx];
3859
3860 // Print the full VType operand of vsetvli/vsetivli instructions, and the SEW
3861 // operand of vector codegen pseudos.
3862 switch (OpInfo.OperandType) {
3865 unsigned Imm = Op.getImm();
3866 RISCVVType::printVType(Imm, OS);
3867 break;
3868 }
3870 unsigned Imm = Op.getImm();
3872 break;
3873 }
3875 unsigned Imm = Op.getImm();
3876 OS << "w" << Imm;
3877 break;
3878 }
3881 unsigned Log2SEW = Op.getImm();
3882 unsigned SEW = Log2SEW ? 1 << Log2SEW : 8;
3883 assert(RISCVVType::isValidSEW(SEW) && "Unexpected SEW");
3884 OS << "e" << SEW;
3885 break;
3886 }
3888 unsigned Policy = Op.getImm();
3890 "Invalid Policy Value");
3891 OS << (Policy & RISCVVType::TAIL_AGNOSTIC ? "ta" : "tu") << ", "
3892 << (Policy & RISCVVType::MASK_AGNOSTIC ? "ma" : "mu");
3893 break;
3894 }
3895
3896 return Comment;
3897}
3898
3899// clang-format off
3900#define CASE_RVV_OPCODE_UNMASK_LMUL(OP, LMUL) \
3901 RISCV::Pseudo##OP##_##LMUL
3902
3903#define CASE_RVV_OPCODE_MASK_LMUL(OP, LMUL) \
3904 RISCV::Pseudo##OP##_##LMUL##_MASK
3905
3906#define CASE_RVV_OPCODE_LMUL(OP, LMUL) \
3907 CASE_RVV_OPCODE_UNMASK_LMUL(OP, LMUL): \
3908 case CASE_RVV_OPCODE_MASK_LMUL(OP, LMUL)
3909
3910#define CASE_RVV_OPCODE_UNMASK_WIDEN(OP) \
3911 CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF8): \
3912 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF4): \
3913 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF2): \
3914 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M1): \
3915 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M2): \
3916 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M4)
3917
3918#define CASE_RVV_OPCODE_UNMASK(OP) \
3919 CASE_RVV_OPCODE_UNMASK_WIDEN(OP): \
3920 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M8)
3921
3922#define CASE_RVV_OPCODE_MASK_WIDEN(OP) \
3923 CASE_RVV_OPCODE_MASK_LMUL(OP, MF8): \
3924 case CASE_RVV_OPCODE_MASK_LMUL(OP, MF4): \
3925 case CASE_RVV_OPCODE_MASK_LMUL(OP, MF2): \
3926 case CASE_RVV_OPCODE_MASK_LMUL(OP, M1): \
3927 case CASE_RVV_OPCODE_MASK_LMUL(OP, M2): \
3928 case CASE_RVV_OPCODE_MASK_LMUL(OP, M4)
3929
3930#define CASE_RVV_OPCODE_MASK(OP) \
3931 CASE_RVV_OPCODE_MASK_WIDEN(OP): \
3932 case CASE_RVV_OPCODE_MASK_LMUL(OP, M8)
3933
3934#define CASE_RVV_OPCODE_WIDEN(OP) \
3935 CASE_RVV_OPCODE_UNMASK_WIDEN(OP): \
3936 case CASE_RVV_OPCODE_MASK_WIDEN(OP)
3937
3938#define CASE_RVV_OPCODE(OP) \
3939 CASE_RVV_OPCODE_UNMASK(OP): \
3940 case CASE_RVV_OPCODE_MASK(OP)
3941// clang-format on
3942
3943// clang-format off
3944#define CASE_VMA_OPCODE_COMMON(OP, TYPE, LMUL) \
3945 RISCV::PseudoV##OP##_##TYPE##_##LMUL
3946
3947#define CASE_VMA_OPCODE_LMULS(OP, TYPE) \
3948 CASE_VMA_OPCODE_COMMON(OP, TYPE, MF8): \
3949 case CASE_VMA_OPCODE_COMMON(OP, TYPE, MF4): \
3950 case CASE_VMA_OPCODE_COMMON(OP, TYPE, MF2): \
3951 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M1): \
3952 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M2): \
3953 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M4): \
3954 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M8)
3955
3956// VFMA instructions are SEW specific.
3957#define CASE_VFMA_OPCODE_COMMON(OP, TYPE, LMUL, SEW) \
3958 RISCV::PseudoV##OP##_##TYPE##_##LMUL##_##SEW
3959
3960#define CASE_VFMA_OPCODE_LMULS_M1(OP, TYPE, SEW) \
3961 CASE_VFMA_OPCODE_COMMON(OP, TYPE, M1, SEW): \
3962 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M2, SEW): \
3963 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M4, SEW): \
3964 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M8, SEW)
3965
3966#define CASE_VFMA_OPCODE_LMULS_MF2(OP, TYPE, SEW) \
3967 CASE_VFMA_OPCODE_COMMON(OP, TYPE, MF2, SEW): \
3968 case CASE_VFMA_OPCODE_LMULS_M1(OP, TYPE, SEW)
3969
3970#define CASE_VFMA_OPCODE_LMULS_MF4(OP, TYPE, SEW) \
3971 CASE_VFMA_OPCODE_COMMON(OP, TYPE, MF4, SEW): \
3972 case CASE_VFMA_OPCODE_LMULS_MF2(OP, TYPE, SEW)
3973
3974#define CASE_VFMA_OPCODE_VV(OP) \
3975 CASE_VFMA_OPCODE_LMULS_MF4(OP, VV, E16): \
3976 case CASE_VFMA_OPCODE_LMULS_MF4(OP##_ALT, VV, E16): \
3977 case CASE_VFMA_OPCODE_LMULS_MF2(OP, VV, E32): \
3978 case CASE_VFMA_OPCODE_LMULS_M1(OP, VV, E64)
3979
3980#define CASE_VFMA_SPLATS(OP) \
3981 CASE_VFMA_OPCODE_LMULS_MF4(OP, VFPR16, E16): \
3982 case CASE_VFMA_OPCODE_LMULS_MF4(OP##_ALT, VFPR16, E16): \
3983 case CASE_VFMA_OPCODE_LMULS_MF2(OP, VFPR32, E32): \
3984 case CASE_VFMA_OPCODE_LMULS_M1(OP, VFPR64, E64)
3985// clang-format on
3986
3988 unsigned &SrcOpIdx1,
3989 unsigned &SrcOpIdx2) const {
3990 const MCInstrDesc &Desc = MI.getDesc();
3991 if (!Desc.isCommutable())
3992 return false;
3993
3994 switch (MI.getOpcode()) {
3995 case RISCV::TH_MVEQZ:
3996 case RISCV::TH_MVNEZ:
3997 // We can't commute operands if operand 2 (i.e., rs1 in
3998 // mveqz/mvnez rd,rs1,rs2) is the zero-register (as it is
3999 // not valid as the in/out-operand 1).
4000 if (MI.getOperand(2).getReg() == RISCV::X0)
4001 return false;
4002 // Operands 1 and 2 are commutable, if we switch the opcode.
4003 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 2);
4004 case RISCV::QC_SELECTIEQ:
4005 case RISCV::QC_SELECTINE:
4006 case RISCV::QC_SELECTIIEQ:
4007 case RISCV::QC_SELECTIINE:
4008 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 2);
4009 case RISCV::QC_MVEQ:
4010 case RISCV::QC_MVNE:
4011 case RISCV::QC_MVLT:
4012 case RISCV::QC_MVGE:
4013 case RISCV::QC_MVLTU:
4014 case RISCV::QC_MVGEU:
4015 case RISCV::QC_MVEQI:
4016 case RISCV::QC_MVNEI:
4017 case RISCV::QC_MVLTI:
4018 case RISCV::QC_MVGEI:
4019 case RISCV::QC_MVLTUI:
4020 case RISCV::QC_MVGEUI:
4021 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 4);
4022 case RISCV::TH_MULA:
4023 case RISCV::TH_MULAW:
4024 case RISCV::TH_MULAH:
4025 case RISCV::TH_MULS:
4026 case RISCV::TH_MULSW:
4027 case RISCV::TH_MULSH:
4028 // Operands 2 and 3 are commutable.
4029 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 2, 3);
4030 case RISCV::PseudoCCMOVGPRNoX0:
4031 case RISCV::PseudoCCMOVGPR:
4032 // Operands 4 and 5 are commutable.
4033 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 4, 5);
4034 case CASE_RVV_OPCODE(VADD_VV):
4035 case CASE_RVV_OPCODE(VAND_VV):
4036 case CASE_RVV_OPCODE(VOR_VV):
4037 case CASE_RVV_OPCODE(VXOR_VV):
4038 case CASE_RVV_OPCODE_MASK(VMSEQ_VV):
4039 case CASE_RVV_OPCODE_MASK(VMSNE_VV):
4040 case CASE_RVV_OPCODE(VMIN_VV):
4041 case CASE_RVV_OPCODE(VMINU_VV):
4042 case CASE_RVV_OPCODE(VMAX_VV):
4043 case CASE_RVV_OPCODE(VMAXU_VV):
4044 case CASE_RVV_OPCODE(VMUL_VV):
4045 case CASE_RVV_OPCODE(VMULH_VV):
4046 case CASE_RVV_OPCODE(VMULHU_VV):
4047 case CASE_RVV_OPCODE_WIDEN(VWADD_VV):
4048 case CASE_RVV_OPCODE_WIDEN(VWADDU_VV):
4049 case CASE_RVV_OPCODE_WIDEN(VWMUL_VV):
4050 case CASE_RVV_OPCODE_WIDEN(VWMULU_VV):
4051 case CASE_RVV_OPCODE_WIDEN(VWMACC_VV):
4052 case CASE_RVV_OPCODE_WIDEN(VWMACCU_VV):
4053 case CASE_RVV_OPCODE_UNMASK(VADC_VVM):
4054 case CASE_RVV_OPCODE(VSADD_VV):
4055 case CASE_RVV_OPCODE(VSADDU_VV):
4056 case CASE_RVV_OPCODE(VAADD_VV):
4057 case CASE_RVV_OPCODE(VAADDU_VV):
4058 case CASE_RVV_OPCODE(VSMUL_VV):
4059 // Operands 2 and 3 are commutable.
4060 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 2, 3);
4061 case CASE_VFMA_SPLATS(FMADD):
4062 case CASE_VFMA_SPLATS(FMSUB):
4063 case CASE_VFMA_SPLATS(FMACC):
4064 case CASE_VFMA_SPLATS(FMSAC):
4067 case CASE_VFMA_SPLATS(FNMACC):
4068 case CASE_VFMA_SPLATS(FNMSAC):
4069 case CASE_VFMA_OPCODE_VV(FMACC):
4070 case CASE_VFMA_OPCODE_VV(FMSAC):
4071 case CASE_VFMA_OPCODE_VV(FNMACC):
4072 case CASE_VFMA_OPCODE_VV(FNMSAC):
4073 case CASE_VMA_OPCODE_LMULS(MADD, VX):
4074 case CASE_VMA_OPCODE_LMULS(NMSUB, VX):
4075 case CASE_VMA_OPCODE_LMULS(MACC, VX):
4076 case CASE_VMA_OPCODE_LMULS(NMSAC, VX):
4077 case CASE_VMA_OPCODE_LMULS(MACC, VV):
4078 case CASE_VMA_OPCODE_LMULS(NMSAC, VV): {
4079 // If the tail policy is undisturbed we can't commute.
4080 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags));
4081 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4082 1) == 0)
4083 return false;
4084
4085 // For these instructions we can only swap operand 1 and operand 3 by
4086 // changing the opcode.
4087 unsigned CommutableOpIdx1 = 1;
4088 unsigned CommutableOpIdx2 = 3;
4089 if (!fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, CommutableOpIdx1,
4090 CommutableOpIdx2))
4091 return false;
4092 return true;
4093 }
4094 case CASE_VFMA_OPCODE_VV(FMADD):
4098 case CASE_VMA_OPCODE_LMULS(MADD, VV):
4099 case CASE_VMA_OPCODE_LMULS(NMSUB, VV): {
4100 // If the tail policy is undisturbed we can't commute.
4101 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags));
4102 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4103 1) == 0)
4104 return false;
4105
4106 // For these instructions we have more freedom. We can commute with the
4107 // other multiplicand or with the addend/subtrahend/minuend.
4108
4109 // Any fixed operand must be from source 1, 2 or 3.
4110 if (SrcOpIdx1 != CommuteAnyOperandIndex && SrcOpIdx1 > 3)
4111 return false;
4112 if (SrcOpIdx2 != CommuteAnyOperandIndex && SrcOpIdx2 > 3)
4113 return false;
4114
4115 // It both ops are fixed one must be the tied source.
4116 if (SrcOpIdx1 != CommuteAnyOperandIndex &&
4117 SrcOpIdx2 != CommuteAnyOperandIndex && SrcOpIdx1 != 1 && SrcOpIdx2 != 1)
4118 return false;
4119
4120 // Look for two different register operands assumed to be commutable
4121 // regardless of the FMA opcode. The FMA opcode is adjusted later if
4122 // needed.
4123 if (SrcOpIdx1 == CommuteAnyOperandIndex ||
4124 SrcOpIdx2 == CommuteAnyOperandIndex) {
4125 // At least one of operands to be commuted is not specified and
4126 // this method is free to choose appropriate commutable operands.
4127 unsigned CommutableOpIdx1 = SrcOpIdx1;
4128 if (SrcOpIdx1 == SrcOpIdx2) {
4129 // Both of operands are not fixed. Set one of commutable
4130 // operands to the tied source.
4131 CommutableOpIdx1 = 1;
4132 } else if (SrcOpIdx1 == CommuteAnyOperandIndex) {
4133 // Only one of the operands is not fixed.
4134 CommutableOpIdx1 = SrcOpIdx2;
4135 }
4136
4137 // CommutableOpIdx1 is well defined now. Let's choose another commutable
4138 // operand and assign its index to CommutableOpIdx2.
4139 unsigned CommutableOpIdx2;
4140 if (CommutableOpIdx1 != 1) {
4141 // If we haven't already used the tied source, we must use it now.
4142 CommutableOpIdx2 = 1;
4143 } else {
4144 Register Op1Reg = MI.getOperand(CommutableOpIdx1).getReg();
4145
4146 // The commuted operands should have different registers.
4147 // Otherwise, the commute transformation does not change anything and
4148 // is useless. We use this as a hint to make our decision.
4149 if (Op1Reg != MI.getOperand(2).getReg())
4150 CommutableOpIdx2 = 2;
4151 else
4152 CommutableOpIdx2 = 3;
4153 }
4154
4155 // Assign the found pair of commutable indices to SrcOpIdx1 and
4156 // SrcOpIdx2 to return those values.
4157 if (!fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, CommutableOpIdx1,
4158 CommutableOpIdx2))
4159 return false;
4160 }
4161
4162 return true;
4163 }
4164 }
4165
4166 return TargetInstrInfo::findCommutedOpIndices(MI, SrcOpIdx1, SrcOpIdx2);
4167}
4168
4169// clang-format off
4170#define CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, LMUL) \
4171 case RISCV::PseudoV##OLDOP##_##TYPE##_##LMUL: \
4172 Opc = RISCV::PseudoV##NEWOP##_##TYPE##_##LMUL; \
4173 break;
4174
4175#define CASE_VMA_CHANGE_OPCODE_LMULS(OLDOP, NEWOP, TYPE) \
4176 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF8) \
4177 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF4) \
4178 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF2) \
4179 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M1) \
4180 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M2) \
4181 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M4) \
4182 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M8)
4183
4184// VFMA depends on SEW.
4185#define CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, LMUL, SEW) \
4186 case RISCV::PseudoV##OLDOP##_##TYPE##_##LMUL##_##SEW: \
4187 Opc = RISCV::PseudoV##NEWOP##_##TYPE##_##LMUL##_##SEW; \
4188 break;
4189
4190#define CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, TYPE, SEW) \
4191 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M1, SEW) \
4192 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M2, SEW) \
4193 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M4, SEW) \
4194 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M8, SEW)
4195
4196#define CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, TYPE, SEW) \
4197 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF2, SEW) \
4198 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, TYPE, SEW)
4199
4200#define CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, TYPE, SEW) \
4201 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF4, SEW) \
4202 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, TYPE, SEW)
4203
4204#define CASE_VFMA_CHANGE_OPCODE_VV(OLDOP, NEWOP) \
4205 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, VV, E16) \
4206 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP##_ALT, NEWOP##_ALT, VV, E16) \
4207 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, VV, E32) \
4208 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, VV, E64)
4209
4210#define CASE_VFMA_CHANGE_OPCODE_SPLATS(OLDOP, NEWOP) \
4211 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, VFPR16, E16) \
4212 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP##_ALT, NEWOP##_ALT, VFPR16, E16) \
4213 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, VFPR32, E32) \
4214 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, VFPR64, E64)
4215// clang-format on
4216
4218 bool NewMI,
4219 unsigned OpIdx1,
4220 unsigned OpIdx2) const {
4221 auto cloneIfNew = [NewMI](MachineInstr &MI) -> MachineInstr & {
4222 if (NewMI)
4223 return *MI.getParent()->getParent()->CloneMachineInstr(&MI);
4224 return MI;
4225 };
4226
4227 switch (MI.getOpcode()) {
4228 case RISCV::TH_MVEQZ:
4229 case RISCV::TH_MVNEZ: {
4230 auto &WorkingMI = cloneIfNew(MI);
4231 WorkingMI.setDesc(get(MI.getOpcode() == RISCV::TH_MVEQZ ? RISCV::TH_MVNEZ
4232 : RISCV::TH_MVEQZ));
4233 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, false, OpIdx1,
4234 OpIdx2);
4235 }
4236 case RISCV::QC_SELECTIEQ:
4237 case RISCV::QC_SELECTINE:
4238 case RISCV::QC_SELECTIIEQ:
4239 case RISCV::QC_SELECTIINE:
4240 return TargetInstrInfo::commuteInstructionImpl(MI, NewMI, OpIdx1, OpIdx2);
4241 case RISCV::QC_MVEQ:
4242 case RISCV::QC_MVNE:
4243 case RISCV::QC_MVLT:
4244 case RISCV::QC_MVGE:
4245 case RISCV::QC_MVLTU:
4246 case RISCV::QC_MVGEU:
4247 case RISCV::QC_MVEQI:
4248 case RISCV::QC_MVNEI:
4249 case RISCV::QC_MVLTI:
4250 case RISCV::QC_MVGEI:
4251 case RISCV::QC_MVLTUI:
4252 case RISCV::QC_MVGEUI: {
4253 auto &WorkingMI = cloneIfNew(MI);
4254 WorkingMI.setDesc(get(getInverseXqcicmOpcode(MI.getOpcode())));
4255 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, false, OpIdx1,
4256 OpIdx2);
4257 }
4258 case RISCV::PseudoCCMOVGPRNoX0:
4259 case RISCV::PseudoCCMOVGPR: {
4260 // CCMOV can be commuted by inverting the condition.
4261 auto CC = static_cast<RISCVCC::CondCode>(MI.getOperand(3).getImm());
4263 auto &WorkingMI = cloneIfNew(MI);
4264 WorkingMI.getOperand(3).setImm(CC);
4265 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, /*NewMI*/ false,
4266 OpIdx1, OpIdx2);
4267 }
4268 case CASE_VFMA_SPLATS(FMACC):
4269 case CASE_VFMA_SPLATS(FMADD):
4270 case CASE_VFMA_SPLATS(FMSAC):
4271 case CASE_VFMA_SPLATS(FMSUB):
4272 case CASE_VFMA_SPLATS(FNMACC):
4274 case CASE_VFMA_SPLATS(FNMSAC):
4276 case CASE_VFMA_OPCODE_VV(FMACC):
4277 case CASE_VFMA_OPCODE_VV(FMSAC):
4278 case CASE_VFMA_OPCODE_VV(FNMACC):
4279 case CASE_VFMA_OPCODE_VV(FNMSAC):
4280 case CASE_VMA_OPCODE_LMULS(MADD, VX):
4281 case CASE_VMA_OPCODE_LMULS(NMSUB, VX):
4282 case CASE_VMA_OPCODE_LMULS(MACC, VX):
4283 case CASE_VMA_OPCODE_LMULS(NMSAC, VX):
4284 case CASE_VMA_OPCODE_LMULS(MACC, VV):
4285 case CASE_VMA_OPCODE_LMULS(NMSAC, VV): {
4286 // It only make sense to toggle these between clobbering the
4287 // addend/subtrahend/minuend one of the multiplicands.
4288 assert((OpIdx1 == 1 || OpIdx2 == 1) && "Unexpected opcode index");
4289 assert((OpIdx1 == 3 || OpIdx2 == 3) && "Unexpected opcode index");
4290 unsigned Opc;
4291 switch (MI.getOpcode()) {
4292 default:
4293 llvm_unreachable("Unexpected opcode");
4294 CASE_VFMA_CHANGE_OPCODE_SPLATS(FMACC, FMADD)
4295 CASE_VFMA_CHANGE_OPCODE_SPLATS(FMADD, FMACC)
4302 CASE_VFMA_CHANGE_OPCODE_VV(FMACC, FMADD)
4306 CASE_VMA_CHANGE_OPCODE_LMULS(MACC, MADD, VX)
4307 CASE_VMA_CHANGE_OPCODE_LMULS(MADD, MACC, VX)
4308 CASE_VMA_CHANGE_OPCODE_LMULS(NMSAC, NMSUB, VX)
4309 CASE_VMA_CHANGE_OPCODE_LMULS(NMSUB, NMSAC, VX)
4310 CASE_VMA_CHANGE_OPCODE_LMULS(MACC, MADD, VV)
4311 CASE_VMA_CHANGE_OPCODE_LMULS(NMSAC, NMSUB, VV)
4312 }
4313
4314 auto &WorkingMI = cloneIfNew(MI);
4315 WorkingMI.setDesc(get(Opc));
4316 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, /*NewMI=*/false,
4317 OpIdx1, OpIdx2);
4318 }
4319 case CASE_VFMA_OPCODE_VV(FMADD):
4323 case CASE_VMA_OPCODE_LMULS(MADD, VV):
4324 case CASE_VMA_OPCODE_LMULS(NMSUB, VV): {
4325 assert((OpIdx1 == 1 || OpIdx2 == 1) && "Unexpected opcode index");
4326 // If one of the operands, is the addend we need to change opcode.
4327 // Otherwise we're just swapping 2 of the multiplicands.
4328 if (OpIdx1 == 3 || OpIdx2 == 3) {
4329 unsigned Opc;
4330 switch (MI.getOpcode()) {
4331 default:
4332 llvm_unreachable("Unexpected opcode");
4333 CASE_VFMA_CHANGE_OPCODE_VV(FMADD, FMACC)
4337 CASE_VMA_CHANGE_OPCODE_LMULS(MADD, MACC, VV)
4338 CASE_VMA_CHANGE_OPCODE_LMULS(NMSUB, NMSAC, VV)
4339 }
4340
4341 auto &WorkingMI = cloneIfNew(MI);
4342 WorkingMI.setDesc(get(Opc));
4343 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, /*NewMI=*/false,
4344 OpIdx1, OpIdx2);
4345 }
4346 // Let the default code handle it.
4347 break;
4348 }
4349 }
4350
4351 return TargetInstrInfo::commuteInstructionImpl(MI, NewMI, OpIdx1, OpIdx2);
4352}
4353
4354#undef CASE_VMA_CHANGE_OPCODE_COMMON
4355#undef CASE_VMA_CHANGE_OPCODE_LMULS
4356#undef CASE_VFMA_CHANGE_OPCODE_COMMON
4357#undef CASE_VFMA_CHANGE_OPCODE_LMULS_M1
4358#undef CASE_VFMA_CHANGE_OPCODE_LMULS_MF2
4359#undef CASE_VFMA_CHANGE_OPCODE_LMULS_MF4
4360#undef CASE_VFMA_CHANGE_OPCODE_VV
4361#undef CASE_VFMA_CHANGE_OPCODE_SPLATS
4362
4363#undef CASE_RVV_OPCODE_UNMASK_LMUL
4364#undef CASE_RVV_OPCODE_MASK_LMUL
4365#undef CASE_RVV_OPCODE_LMUL
4366#undef CASE_RVV_OPCODE_UNMASK_WIDEN
4367#undef CASE_RVV_OPCODE_UNMASK
4368#undef CASE_RVV_OPCODE_MASK_WIDEN
4369#undef CASE_RVV_OPCODE_MASK
4370#undef CASE_RVV_OPCODE_WIDEN
4371#undef CASE_RVV_OPCODE
4372
4373#undef CASE_VMA_OPCODE_COMMON
4374#undef CASE_VMA_OPCODE_LMULS
4375#undef CASE_VFMA_OPCODE_COMMON
4376#undef CASE_VFMA_OPCODE_LMULS_M1
4377#undef CASE_VFMA_OPCODE_LMULS_MF2
4378#undef CASE_VFMA_OPCODE_LMULS_MF4
4379#undef CASE_VFMA_OPCODE_VV
4380#undef CASE_VFMA_SPLATS
4381
4383 switch (MI.getOpcode()) {
4384 default:
4385 break;
4386 case RISCV::ADD:
4387 case RISCV::OR:
4388 case RISCV::XOR:
4389 // Normalize (so we hit the next if clause).
4390 // add/[x]or rd, zero, rs => add/[x]or rd, rs, zero
4391 if (MI.getOperand(1).getReg() == RISCV::X0)
4392 commuteInstruction(MI);
4393 // add/[x]or rd, rs, zero => addi rd, rs, 0
4394 if (MI.getOperand(2).getReg() == RISCV::X0) {
4395 MI.getOperand(2).ChangeToImmediate(0);
4396 MI.setDesc(get(RISCV::ADDI));
4397 return true;
4398 }
4399 // xor rd, rs, rs => addi rd, zero, 0
4400 if (MI.getOpcode() == RISCV::XOR &&
4401 MI.getOperand(1).getReg() == MI.getOperand(2).getReg()) {
4402 MI.getOperand(1).setReg(RISCV::X0);
4403 MI.getOperand(2).ChangeToImmediate(0);
4404 MI.setDesc(get(RISCV::ADDI));
4405 return true;
4406 }
4407 break;
4408 case RISCV::ORI:
4409 case RISCV::XORI:
4410 // [x]ori rd, zero, N => addi rd, zero, N
4411 if (MI.getOperand(1).getReg() == RISCV::X0) {
4412 MI.setDesc(get(RISCV::ADDI));
4413 return true;
4414 }
4415 break;
4416 case RISCV::SUB:
4417 // sub rd, rs, zero => addi rd, rs, 0
4418 if (MI.getOperand(2).getReg() == RISCV::X0) {
4419 MI.getOperand(2).ChangeToImmediate(0);
4420 MI.setDesc(get(RISCV::ADDI));
4421 return true;
4422 }
4423 break;
4424 case RISCV::SUBW:
4425 // subw rd, rs, zero => addiw rd, rs, 0
4426 if (MI.getOperand(2).getReg() == RISCV::X0) {
4427 MI.getOperand(2).ChangeToImmediate(0);
4428 MI.setDesc(get(RISCV::ADDIW));
4429 return true;
4430 }
4431 break;
4432 case RISCV::ADDW:
4433 // Normalize (so we hit the next if clause).
4434 // addw rd, zero, rs => addw rd, rs, zero
4435 if (MI.getOperand(1).getReg() == RISCV::X0)
4436 commuteInstruction(MI);
4437 // addw rd, rs, zero => addiw rd, rs, 0
4438 if (MI.getOperand(2).getReg() == RISCV::X0) {
4439 MI.getOperand(2).ChangeToImmediate(0);
4440 MI.setDesc(get(RISCV::ADDIW));
4441 return true;
4442 }
4443 break;
4444 case RISCV::SH1ADD:
4445 case RISCV::SH1ADD_UW:
4446 case RISCV::SH2ADD:
4447 case RISCV::SH2ADD_UW:
4448 case RISCV::SH3ADD:
4449 case RISCV::SH3ADD_UW:
4450 // shNadd[.uw] rd, zero, rs => addi rd, rs, 0
4451 if (MI.getOperand(1).getReg() == RISCV::X0) {
4452 MI.removeOperand(1);
4453 MI.addOperand(MachineOperand::CreateImm(0));
4454 MI.setDesc(get(RISCV::ADDI));
4455 return true;
4456 }
4457 // shNadd[.uw] rd, rs, zero => slli[.uw] rd, rs, N
4458 if (MI.getOperand(2).getReg() == RISCV::X0) {
4459 MI.removeOperand(2);
4460 unsigned Opc = MI.getOpcode();
4461 if (Opc == RISCV::SH1ADD_UW || Opc == RISCV::SH2ADD_UW ||
4462 Opc == RISCV::SH3ADD_UW) {
4464 MI.setDesc(get(RISCV::SLLI_UW));
4465 return true;
4466 }
4468 MI.setDesc(get(RISCV::SLLI));
4469 return true;
4470 }
4471 break;
4472 case RISCV::AND:
4473 case RISCV::MUL:
4474 case RISCV::MULH:
4475 case RISCV::MULHSU:
4476 case RISCV::MULHU:
4477 case RISCV::MULW:
4478 // and rd, zero, rs => addi rd, zero, 0
4479 // mul* rd, zero, rs => addi rd, zero, 0
4480 // and rd, rs, zero => addi rd, zero, 0
4481 // mul* rd, rs, zero => addi rd, zero, 0
4482 if (MI.getOperand(1).getReg() == RISCV::X0 ||
4483 MI.getOperand(2).getReg() == RISCV::X0) {
4484 MI.getOperand(1).setReg(RISCV::X0);
4485 MI.getOperand(2).ChangeToImmediate(0);
4486 MI.setDesc(get(RISCV::ADDI));
4487 return true;
4488 }
4489 break;
4490 case RISCV::ANDI:
4491 // andi rd, zero, C => addi rd, zero, 0
4492 if (MI.getOperand(1).getReg() == RISCV::X0) {
4493 MI.getOperand(2).setImm(0);
4494 MI.setDesc(get(RISCV::ADDI));
4495 return true;
4496 }
4497 break;
4498 case RISCV::SLL:
4499 case RISCV::SRL:
4500 case RISCV::SRA:
4501 // shift rd, zero, rs => addi rd, zero, 0
4502 if (MI.getOperand(1).getReg() == RISCV::X0) {
4503 MI.getOperand(2).ChangeToImmediate(0);
4504 MI.setDesc(get(RISCV::ADDI));
4505 return true;
4506 }
4507 // shift rd, rs, zero => addi rd, rs, 0
4508 if (MI.getOperand(2).getReg() == RISCV::X0) {
4509 MI.getOperand(2).ChangeToImmediate(0);
4510 MI.setDesc(get(RISCV::ADDI));
4511 return true;
4512 }
4513 break;
4514 case RISCV::SLLW:
4515 case RISCV::SRLW:
4516 case RISCV::SRAW:
4517 // shiftw rd, zero, rs => addi rd, zero, 0
4518 if (MI.getOperand(1).getReg() == RISCV::X0) {
4519 MI.getOperand(2).ChangeToImmediate(0);
4520 MI.setDesc(get(RISCV::ADDI));
4521 return true;
4522 }
4523 break;
4524 case RISCV::SLLI:
4525 case RISCV::SRLI:
4526 case RISCV::SRAI:
4527 case RISCV::SLLIW:
4528 case RISCV::SRLIW:
4529 case RISCV::SRAIW:
4530 case RISCV::SLLI_UW:
4531 // shiftimm rd, zero, N => addi rd, zero, 0
4532 if (MI.getOperand(1).getReg() == RISCV::X0) {
4533 MI.getOperand(2).setImm(0);
4534 MI.setDesc(get(RISCV::ADDI));
4535 return true;
4536 }
4537 break;
4538 case RISCV::SLTU:
4539 case RISCV::ADD_UW:
4540 // sltu rd, zero, zero => addi rd, zero, 0
4541 // add.uw rd, zero, zero => addi rd, zero, 0
4542 if (MI.getOperand(1).getReg() == RISCV::X0 &&
4543 MI.getOperand(2).getReg() == RISCV::X0) {
4544 MI.getOperand(2).ChangeToImmediate(0);
4545 MI.setDesc(get(RISCV::ADDI));
4546 return true;
4547 }
4548 // add.uw rd, zero, rs => addi rd, rs, 0
4549 if (MI.getOpcode() == RISCV::ADD_UW &&
4550 MI.getOperand(1).getReg() == RISCV::X0) {
4551 MI.removeOperand(1);
4552 MI.addOperand(MachineOperand::CreateImm(0));
4553 MI.setDesc(get(RISCV::ADDI));
4554 }
4555 break;
4556 case RISCV::SLTIU:
4557 // sltiu rd, zero, NZC => addi rd, zero, 1
4558 // sltiu rd, zero, 0 => addi rd, zero, 0
4559 if (MI.getOperand(1).getReg() == RISCV::X0) {
4560 MI.getOperand(2).setImm(MI.getOperand(2).getImm() != 0);
4561 MI.setDesc(get(RISCV::ADDI));
4562 return true;
4563 }
4564 break;
4565 case RISCV::SEXT_H:
4566 case RISCV::SEXT_B:
4567 case RISCV::ZEXT_H_RV32:
4568 case RISCV::ZEXT_H_RV64:
4569 // sext.[hb] rd, zero => addi rd, zero, 0
4570 // zext.h rd, zero => addi rd, zero, 0
4571 if (MI.getOperand(1).getReg() == RISCV::X0) {
4572 MI.addOperand(MachineOperand::CreateImm(0));
4573 MI.setDesc(get(RISCV::ADDI));
4574 return true;
4575 }
4576 break;
4577 case RISCV::MIN:
4578 case RISCV::MINU:
4579 case RISCV::MAX:
4580 case RISCV::MAXU:
4581 // min|max rd, rs, rs => addi rd, rs, 0
4582 if (MI.getOperand(1).getReg() == MI.getOperand(2).getReg()) {
4583 MI.getOperand(2).ChangeToImmediate(0);
4584 MI.setDesc(get(RISCV::ADDI));
4585 return true;
4586 }
4587 break;
4588 case RISCV::BEQ:
4589 case RISCV::BNE:
4590 // b{eq,ne} zero, rs, imm => b{eq,ne} rs, zero, imm
4591 if (MI.getOperand(0).getReg() == RISCV::X0) {
4592 MachineOperand MO0 = MI.getOperand(0);
4593 MI.removeOperand(0);
4594 MI.insert(MI.operands_begin() + 1, {MO0});
4595 }
4596 break;
4597 case RISCV::BLTU:
4598 // bltu zero, rs, imm => bne rs, zero, imm
4599 if (MI.getOperand(0).getReg() == RISCV::X0) {
4600 MachineOperand MO0 = MI.getOperand(0);
4601 MI.removeOperand(0);
4602 MI.insert(MI.operands_begin() + 1, {MO0});
4603 MI.setDesc(get(RISCV::BNE));
4604 }
4605 break;
4606 case RISCV::BGEU:
4607 // bgeu zero, rs, imm => beq rs, zero, imm
4608 if (MI.getOperand(0).getReg() == RISCV::X0) {
4609 MachineOperand MO0 = MI.getOperand(0);
4610 MI.removeOperand(0);
4611 MI.insert(MI.operands_begin() + 1, {MO0});
4612 MI.setDesc(get(RISCV::BEQ));
4613 }
4614 break;
4615 }
4616 return false;
4617}
4618
4619// clang-format off
4620#define CASE_WIDEOP_OPCODE_COMMON(OP, LMUL) \
4621 RISCV::PseudoV##OP##_##LMUL##_TIED
4622
4623#define CASE_WIDEOP_OPCODE_LMULS(OP) \
4624 CASE_WIDEOP_OPCODE_COMMON(OP, MF8): \
4625 case CASE_WIDEOP_OPCODE_COMMON(OP, MF4): \
4626 case CASE_WIDEOP_OPCODE_COMMON(OP, MF2): \
4627 case CASE_WIDEOP_OPCODE_COMMON(OP, M1): \
4628 case CASE_WIDEOP_OPCODE_COMMON(OP, M2): \
4629 case CASE_WIDEOP_OPCODE_COMMON(OP, M4)
4630
4631#define CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, LMUL) \
4632 case RISCV::PseudoV##OP##_##LMUL##_TIED: \
4633 NewOpc = RISCV::PseudoV##OP##_##LMUL; \
4634 break;
4635
4636#define CASE_WIDEOP_CHANGE_OPCODE_LMULS(OP) \
4637 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF8) \
4638 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4) \
4639 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2) \
4640 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1) \
4641 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2) \
4642 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4)
4643
4644// FP Widening Ops may by SEW aware. Create SEW aware cases for these cases.
4645#define CASE_FP_WIDEOP_OPCODE_COMMON(OP, LMUL, SEW) \
4646 RISCV::PseudoV##OP##_##LMUL##_##SEW##_TIED
4647
4648#define CASE_FP_WIDEOP_OPCODE_LMULS(OP) \
4649 CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF4, E16): \
4650 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E16): \
4651 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E32): \
4652 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E16): \
4653 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E32): \
4654 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E16): \
4655 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E32): \
4656 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E16): \
4657 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E32) \
4658
4659#define CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, LMUL, SEW) \
4660 case RISCV::PseudoV##OP##_##LMUL##_##SEW##_TIED: \
4661 NewOpc = RISCV::PseudoV##OP##_##LMUL##_##SEW; \
4662 break;
4663
4664#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS(OP) \
4665 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4, E16) \
4666 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E16) \
4667 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E32) \
4668 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E16) \
4669 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E32) \
4670 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E16) \
4671 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E32) \
4672 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E16) \
4673 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E32) \
4674
4675#define CASE_FP_WIDEOP_OPCODE_LMULS_ALT(OP) \
4676 CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF4, E16): \
4677 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E16): \
4678 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E16): \
4679 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E16): \
4680 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E16)
4681
4682#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS_ALT(OP) \
4683 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4, E16) \
4684 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E16) \
4685 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E16) \
4686 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E16) \
4687 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E16)
4688// clang-format on
4689
4691 LiveVariables *LV,
4692 LiveIntervals *LIS) const {
4694 switch (MI.getOpcode()) {
4695 default:
4696 return nullptr;
4697 case CASE_FP_WIDEOP_OPCODE_LMULS_ALT(FWADD_ALT_WV):
4698 case CASE_FP_WIDEOP_OPCODE_LMULS_ALT(FWSUB_ALT_WV):
4699 case CASE_FP_WIDEOP_OPCODE_LMULS(FWADD_WV):
4700 case CASE_FP_WIDEOP_OPCODE_LMULS(FWSUB_WV): {
4701 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags) &&
4702 MI.getNumExplicitOperands() == 7 &&
4703 "Expect 7 explicit operands rd, rs2, rs1, rm, vl, sew, policy");
4704 // If the tail policy is undisturbed we can't convert.
4705 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4706 1) == 0)
4707 return nullptr;
4708 // clang-format off
4709 unsigned NewOpc;
4710 switch (MI.getOpcode()) {
4711 default:
4712 llvm_unreachable("Unexpected opcode");
4717 }
4718 // clang-format on
4719
4720 MachineBasicBlock &MBB = *MI.getParent();
4721 MIB = BuildMI(MBB, MI, MI.getDebugLoc(), get(NewOpc))
4722 .add(MI.getOperand(0))
4723 .addReg(MI.getOperand(0).getReg(), RegState::Undef)
4724 .add(MI.getOperand(1))
4725 .add(MI.getOperand(2))
4726 .add(MI.getOperand(3))
4727 .add(MI.getOperand(4))
4728 .add(MI.getOperand(5))
4729 .add(MI.getOperand(6));
4730 break;
4731 }
4732 case CASE_WIDEOP_OPCODE_LMULS(WADD_WV):
4733 case CASE_WIDEOP_OPCODE_LMULS(WADDU_WV):
4734 case CASE_WIDEOP_OPCODE_LMULS(WSUB_WV):
4735 case CASE_WIDEOP_OPCODE_LMULS(WSUBU_WV): {
4736 // If the tail policy is undisturbed we can't convert.
4737 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags) &&
4738 MI.getNumExplicitOperands() == 6);
4739 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4740 1) == 0)
4741 return nullptr;
4742
4743 // clang-format off
4744 unsigned NewOpc;
4745 switch (MI.getOpcode()) {
4746 default:
4747 llvm_unreachable("Unexpected opcode");
4752 }
4753 // clang-format on
4754
4755 MachineBasicBlock &MBB = *MI.getParent();
4756 MIB = BuildMI(MBB, MI, MI.getDebugLoc(), get(NewOpc))
4757 .add(MI.getOperand(0))
4758 .addReg(MI.getOperand(0).getReg(), RegState::Undef)
4759 .add(MI.getOperand(1))
4760 .add(MI.getOperand(2))
4761 .add(MI.getOperand(3))
4762 .add(MI.getOperand(4))
4763 .add(MI.getOperand(5));
4764 break;
4765 }
4766 }
4767 MIB.copyImplicitOps(MI);
4768
4769 if (LV) {
4770 unsigned NumOps = MI.getNumOperands();
4771 for (unsigned I = 1; I < NumOps; ++I) {
4772 MachineOperand &Op = MI.getOperand(I);
4773 if (Op.isReg() && Op.isKill())
4774 LV->replaceKillInstruction(Op.getReg(), MI, *MIB);
4775 }
4776 }
4777
4778 if (LIS) {
4779 SlotIndex Idx = LIS->ReplaceMachineInstrInMaps(MI, *MIB);
4780
4781 if (MI.getOperand(0).isEarlyClobber()) {
4782 // Use operand 1 was tied to early-clobber def operand 0, so its live
4783 // interval could have ended at an early-clobber slot. Now they are not
4784 // tied we need to update it to the normal register slot.
4785 LiveInterval &LI = LIS->getInterval(MI.getOperand(1).getReg());
4787 if (S->end == Idx.getRegSlot(true))
4788 S->end = Idx.getRegSlot();
4789 }
4790 }
4791
4792 return MIB;
4793}
4794
4795#undef CASE_WIDEOP_OPCODE_COMMON
4796#undef CASE_WIDEOP_OPCODE_LMULS
4797#undef CASE_WIDEOP_CHANGE_OPCODE_COMMON
4798#undef CASE_WIDEOP_CHANGE_OPCODE_LMULS
4799#undef CASE_FP_WIDEOP_OPCODE_COMMON
4800#undef CASE_FP_WIDEOP_OPCODE_LMULS
4801#undef CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON
4802#undef CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS
4803
4806 Register DestReg, uint32_t Amount,
4807 MachineInstr::MIFlag Flag) const {
4809 if (llvm::has_single_bit<uint32_t>(Amount)) {
4810 uint32_t ShiftAmount = Log2_32(Amount);
4811 if (ShiftAmount == 0)
4812 return;
4813 BuildMI(MBB, II, DL, get(RISCV::SLLI), DestReg)
4814 .addReg(DestReg, RegState::Kill)
4815 .addImm(ShiftAmount)
4816 .setMIFlag(Flag);
4817 } else if (int ShXAmount, ShiftAmount;
4818 STI.hasShlAdd(3) &&
4819 (ShXAmount = isShifted359(Amount, ShiftAmount)) != 0) {
4820 // We can use Zba SHXADD+SLLI instructions for multiply in some cases.
4821 unsigned Opc;
4822 switch (ShXAmount) {
4823 case 1:
4824 Opc = RISCV::SH1ADD;
4825 break;
4826 case 2:
4827 Opc = RISCV::SH2ADD;
4828 break;
4829 case 3:
4830 Opc = RISCV::SH3ADD;
4831 break;
4832 default:
4833 llvm_unreachable("unexpected result of isShifted359");
4834 }
4835 if (ShiftAmount)
4836 BuildMI(MBB, II, DL, get(RISCV::SLLI), DestReg)
4837 .addReg(DestReg, RegState::Kill)
4838 .addImm(ShiftAmount)
4839 .setMIFlag(Flag);
4840 BuildMI(MBB, II, DL, get(Opc), DestReg)
4841 .addReg(DestReg, RegState::Kill)
4842 .addReg(DestReg)
4843 .setMIFlag(Flag);
4844 } else if (llvm::has_single_bit<uint32_t>(Amount - 1)) {
4845 Register ScaledRegister = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4846 uint32_t ShiftAmount = Log2_32(Amount - 1);
4847 BuildMI(MBB, II, DL, get(RISCV::SLLI), ScaledRegister)
4848 .addReg(DestReg)
4849 .addImm(ShiftAmount)
4850 .setMIFlag(Flag);
4851 BuildMI(MBB, II, DL, get(RISCV::ADD), DestReg)
4852 .addReg(ScaledRegister, RegState::Kill)
4853 .addReg(DestReg, RegState::Kill)
4854 .setMIFlag(Flag);
4855 } else if (llvm::has_single_bit<uint32_t>(Amount + 1)) {
4856 Register ScaledRegister = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4857 uint32_t ShiftAmount = Log2_32(Amount + 1);
4858 BuildMI(MBB, II, DL, get(RISCV::SLLI), ScaledRegister)
4859 .addReg(DestReg)
4860 .addImm(ShiftAmount)
4861 .setMIFlag(Flag);
4862 BuildMI(MBB, II, DL, get(RISCV::SUB), DestReg)
4863 .addReg(ScaledRegister, RegState::Kill)
4864 .addReg(DestReg, RegState::Kill)
4865 .setMIFlag(Flag);
4866 } else if (STI.hasStdExtZmmul()) {
4867 Register N = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4868 movImm(MBB, II, DL, N, Amount, Flag);
4869 BuildMI(MBB, II, DL, get(RISCV::MUL), DestReg)
4870 .addReg(DestReg, RegState::Kill)
4872 .setMIFlag(Flag);
4873 } else {
4874 Register Acc;
4875 uint32_t PrevShiftAmount = 0;
4876 for (uint32_t ShiftAmount = 0; Amount >> ShiftAmount; ShiftAmount++) {
4877 if (Amount & (1U << ShiftAmount)) {
4878 if (ShiftAmount)
4879 BuildMI(MBB, II, DL, get(RISCV::SLLI), DestReg)
4880 .addReg(DestReg, RegState::Kill)
4881 .addImm(ShiftAmount - PrevShiftAmount)
4882 .setMIFlag(Flag);
4883 if (Amount >> (ShiftAmount + 1)) {
4884 // If we don't have an accmulator yet, create it and copy DestReg.
4885 if (!Acc) {
4886 Acc = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4887 BuildMI(MBB, II, DL, get(TargetOpcode::COPY), Acc)
4888 .addReg(DestReg)
4889 .setMIFlag(Flag);
4890 } else {
4891 BuildMI(MBB, II, DL, get(RISCV::ADD), Acc)
4892 .addReg(Acc, RegState::Kill)
4893 .addReg(DestReg)
4894 .setMIFlag(Flag);
4895 }
4896 }
4897 PrevShiftAmount = ShiftAmount;
4898 }
4899 }
4900 assert(Acc && "Expected valid accumulator");
4901 BuildMI(MBB, II, DL, get(RISCV::ADD), DestReg)
4902 .addReg(DestReg, RegState::Kill)
4903 .addReg(Acc, RegState::Kill)
4904 .setMIFlag(Flag);
4905 }
4906}
4907
4910 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
4911 {{MONontemporalBit0, "riscv-nontemporal-domain-bit-0"},
4912 {MONontemporalBit1, "riscv-nontemporal-domain-bit-1"}};
4913 return ArrayRef(TargetFlags);
4914}
4915
4917 return OptLevel >= CodeGenOptLevel::Aggressive
4918 ? STI.getTailDupAggressiveThreshold()
4919 : 2;
4920}
4921
4923 // RVV lacks any support for immediate addressing for stack addresses, so be
4924 // conservative.
4925 unsigned Opcode = MI.getOpcode();
4926 if (!RISCVVPseudosTable::getPseudoInfo(Opcode) &&
4928 return false;
4929 return true;
4930}
4931
4932/// Return true if \p MI is a copy that will be lowered to one or more vmvNr.vs.
4934 const MachineInstr &MI) {
4935 return MI.isCopy() && MI.getOperand(0).getReg().isPhysical() &&
4937 TRI->getMinimalPhysRegClass(MI.getOperand(0).getReg()));
4938}
4939
4940std::optional<std::pair<unsigned, unsigned>>
4942 switch (Opcode) {
4943 default:
4944 return std::nullopt;
4945 case RISCV::PseudoVSPILL2_M1:
4946 case RISCV::PseudoVRELOAD2_M1:
4947 return std::make_pair(2u, 1u);
4948 case RISCV::PseudoVSPILL2_M2:
4949 case RISCV::PseudoVRELOAD2_M2:
4950 return std::make_pair(2u, 2u);
4951 case RISCV::PseudoVSPILL2_M4:
4952 case RISCV::PseudoVRELOAD2_M4:
4953 return std::make_pair(2u, 4u);
4954 case RISCV::PseudoVSPILL3_M1:
4955 case RISCV::PseudoVRELOAD3_M1:
4956 return std::make_pair(3u, 1u);
4957 case RISCV::PseudoVSPILL3_M2:
4958 case RISCV::PseudoVRELOAD3_M2:
4959 return std::make_pair(3u, 2u);
4960 case RISCV::PseudoVSPILL4_M1:
4961 case RISCV::PseudoVRELOAD4_M1:
4962 return std::make_pair(4u, 1u);
4963 case RISCV::PseudoVSPILL4_M2:
4964 case RISCV::PseudoVRELOAD4_M2:
4965 return std::make_pair(4u, 2u);
4966 case RISCV::PseudoVSPILL5_M1:
4967 case RISCV::PseudoVRELOAD5_M1:
4968 return std::make_pair(5u, 1u);
4969 case RISCV::PseudoVSPILL6_M1:
4970 case RISCV::PseudoVRELOAD6_M1:
4971 return std::make_pair(6u, 1u);
4972 case RISCV::PseudoVSPILL7_M1:
4973 case RISCV::PseudoVRELOAD7_M1:
4974 return std::make_pair(7u, 1u);
4975 case RISCV::PseudoVSPILL8_M1:
4976 case RISCV::PseudoVRELOAD8_M1:
4977 return std::make_pair(8u, 1u);
4978 }
4979}
4980
4981bool RISCV::hasEqualFRM(const MachineInstr &MI1, const MachineInstr &MI2) {
4982 int16_t MI1FrmOpIdx =
4983 RISCV::getNamedOperandIdx(MI1.getOpcode(), RISCV::OpName::frm);
4984 int16_t MI2FrmOpIdx =
4985 RISCV::getNamedOperandIdx(MI2.getOpcode(), RISCV::OpName::frm);
4986 if (MI1FrmOpIdx < 0 || MI2FrmOpIdx < 0)
4987 return false;
4988 MachineOperand FrmOp1 = MI1.getOperand(MI1FrmOpIdx);
4989 MachineOperand FrmOp2 = MI2.getOperand(MI2FrmOpIdx);
4990 return FrmOp1.getImm() == FrmOp2.getImm();
4991}
4992
4993std::optional<unsigned>
4994RISCV::getVectorLowDemandedScalarBits(unsigned Opcode, unsigned Log2SEW) {
4995 switch (Opcode) {
4996 default:
4997 return std::nullopt;
4998
4999 // 11.6. Vector Single-Width Shift Instructions
5000 case RISCV::VSLL_VX:
5001 case RISCV::VSRL_VX:
5002 case RISCV::VSRA_VX:
5003 // 12.4. Vector Single-Width Scaling Shift Instructions
5004 case RISCV::VSSRL_VX:
5005 case RISCV::VSSRA_VX:
5006 // Zvbb
5007 case RISCV::VROL_VX:
5008 case RISCV::VROR_VX:
5009 // Only the low lg2(SEW) bits of the shift-amount value are used.
5010 return Log2SEW;
5011
5012 // 11.7 Vector Narrowing Integer Right Shift Instructions
5013 case RISCV::VNSRL_WX:
5014 case RISCV::VNSRA_WX:
5015 // 12.5. Vector Narrowing Fixed-Point Clip Instructions
5016 case RISCV::VNCLIPU_WX:
5017 case RISCV::VNCLIP_WX:
5018 // Zvbb
5019 case RISCV::VWSLL_VX:
5020 // Only the low lg2(2*SEW) bits of the shift-amount value are used.
5021 return Log2SEW + 1;
5022
5023 // 11.1. Vector Single-Width Integer Add and Subtract
5024 case RISCV::VADD_VX:
5025 case RISCV::VSUB_VX:
5026 case RISCV::VRSUB_VX:
5027 // 11.2. Vector Widening Integer Add/Subtract
5028 case RISCV::VWADDU_VX:
5029 case RISCV::VWSUBU_VX:
5030 case RISCV::VWADD_VX:
5031 case RISCV::VWSUB_VX:
5032 case RISCV::VWADDU_WX:
5033 case RISCV::VWSUBU_WX:
5034 case RISCV::VWADD_WX:
5035 case RISCV::VWSUB_WX:
5036 // 11.4. Vector Integer Add-with-Carry / Subtract-with-Borrow Instructions
5037 case RISCV::VADC_VXM:
5038 case RISCV::VADC_VIM:
5039 case RISCV::VMADC_VXM:
5040 case RISCV::VMADC_VIM:
5041 case RISCV::VMADC_VX:
5042 case RISCV::VSBC_VXM:
5043 case RISCV::VMSBC_VXM:
5044 case RISCV::VMSBC_VX:
5045 // 11.5 Vector Bitwise Logical Instructions
5046 case RISCV::VAND_VX:
5047 case RISCV::VOR_VX:
5048 case RISCV::VXOR_VX:
5049 // 11.8. Vector Integer Compare Instructions
5050 case RISCV::VMSEQ_VX:
5051 case RISCV::VMSNE_VX:
5052 case RISCV::VMSLTU_VX:
5053 case RISCV::VMSLT_VX:
5054 case RISCV::VMSLEU_VX:
5055 case RISCV::VMSLE_VX:
5056 case RISCV::VMSGTU_VX:
5057 case RISCV::VMSGT_VX:
5058 // 11.9. Vector Integer Min/Max Instructions
5059 case RISCV::VMINU_VX:
5060 case RISCV::VMIN_VX:
5061 case RISCV::VMAXU_VX:
5062 case RISCV::VMAX_VX:
5063 // 11.10. Vector Single-Width Integer Multiply Instructions
5064 case RISCV::VMUL_VX:
5065 case RISCV::VMULH_VX:
5066 case RISCV::VMULHU_VX:
5067 case RISCV::VMULHSU_VX:
5068 // 11.11. Vector Integer Divide Instructions
5069 case RISCV::VDIVU_VX:
5070 case RISCV::VDIV_VX:
5071 case RISCV::VREMU_VX:
5072 case RISCV::VREM_VX:
5073 // 11.12. Vector Widening Integer Multiply Instructions
5074 case RISCV::VWMUL_VX:
5075 case RISCV::VWMULU_VX:
5076 case RISCV::VWMULSU_VX:
5077 // 11.13. Vector Single-Width Integer Multiply-Add Instructions
5078 case RISCV::VMACC_VX:
5079 case RISCV::VNMSAC_VX:
5080 case RISCV::VMADD_VX:
5081 case RISCV::VNMSUB_VX:
5082 // 11.14. Vector Widening Integer Multiply-Add Instructions
5083 case RISCV::VWMACCU_VX:
5084 case RISCV::VWMACC_VX:
5085 case RISCV::VWMACCSU_VX:
5086 case RISCV::VWMACCUS_VX:
5087 // 11.15. Vector Integer Merge Instructions
5088 case RISCV::VMERGE_VXM:
5089 // 11.16. Vector Integer Move Instructions
5090 case RISCV::VMV_V_X:
5091 // 12.1. Vector Single-Width Saturating Add and Subtract
5092 case RISCV::VSADDU_VX:
5093 case RISCV::VSADD_VX:
5094 case RISCV::VSSUBU_VX:
5095 case RISCV::VSSUB_VX:
5096 // 12.2. Vector Single-Width Averaging Add and Subtract
5097 case RISCV::VAADDU_VX:
5098 case RISCV::VAADD_VX:
5099 case RISCV::VASUBU_VX:
5100 case RISCV::VASUB_VX:
5101 // 12.3. Vector Single-Width Fractional Multiply with Rounding and Saturation
5102 case RISCV::VSMUL_VX:
5103 // 16.1. Integer Scalar Move Instructions
5104 case RISCV::VMV_S_X:
5105 // Zvbb
5106 case RISCV::VANDN_VX:
5107 return 1U << Log2SEW;
5108 }
5109}
5110
5111unsigned RISCV::getRVVMCOpcode(unsigned RVVPseudoOpcode) {
5113 RISCVVPseudosTable::getPseudoInfo(RVVPseudoOpcode);
5114 if (!RVV)
5115 return 0;
5116 return RVV->BaseInstr;
5117}
5118
5119unsigned RISCV::getDestLog2EEW(const MCInstrDesc &Desc, unsigned Log2SEW) {
5120 unsigned DestEEW =
5122 // EEW = 1
5123 if (DestEEW == 0)
5124 return 0;
5125 // EEW = SEW * n
5126 unsigned Scaled = Log2SEW + (DestEEW - 1);
5127 assert(Scaled >= 3 && Scaled <= 6);
5128 return Scaled;
5129}
5130
5131static std::optional<int64_t> getEffectiveImm(const MachineOperand &MO) {
5132 assert(MO.isImm() || MO.getReg().isVirtual());
5133 if (MO.isImm())
5134 return MO.getImm();
5135 const MachineInstr *Def =
5136 MO.getParent()->getMF()->getRegInfo().getVRegDef(MO.getReg());
5137 int64_t Imm;
5138 if (isLoadImm(Def, Imm))
5139 return Imm;
5140 return std::nullopt;
5141}
5142
5143/// Given two VL operands, do we know that LHS <= RHS? Must be used in SSA form.
5145 assert((LHS.isImm() || LHS.getParent()->getMF()->getRegInfo().isSSA()) &&
5146 (RHS.isImm() || RHS.getParent()->getMF()->getRegInfo().isSSA()));
5147 if (LHS.isReg() && RHS.isReg() && LHS.getReg().isVirtual() &&
5148 LHS.getReg() == RHS.getReg())
5149 return true;
5150 if (RHS.isImm() && RHS.getImm() == RISCV::VLMaxSentinel)
5151 return true;
5152 if (LHS.isImm() && LHS.getImm() == 0)
5153 return true;
5154 if (LHS.isImm() && LHS.getImm() == RISCV::VLMaxSentinel)
5155 return false;
5156 std::optional<int64_t> LHSImm = getEffectiveImm(LHS),
5157 RHSImm = getEffectiveImm(RHS);
5158 if (!LHSImm || !RHSImm)
5159 return false;
5160 return LHSImm <= RHSImm;
5161}
5162
5163namespace {
5164class RISCVPipelinerLoopInfo : public TargetInstrInfo::PipelinerLoopInfo {
5165 const MachineInstr *LHS;
5166 const MachineInstr *RHS;
5168
5169public:
5170 RISCVPipelinerLoopInfo(const MachineInstr *LHS, const MachineInstr *RHS,
5172 : LHS(LHS), RHS(RHS), Cond(Cond.begin(), Cond.end()) {}
5173
5174 bool shouldIgnoreForPipelining(const MachineInstr *MI) const override {
5175 // Make the instructions for loop control be placed in stage 0.
5176 // The predecessors of LHS/RHS are considered by the caller.
5177 if (LHS && MI == LHS)
5178 return true;
5179 if (RHS && MI == RHS)
5180 return true;
5181 return false;
5182 }
5183
5184 std::optional<bool> createTripCountGreaterCondition(
5185 int TC, MachineBasicBlock &MBB,
5186 SmallVectorImpl<MachineOperand> &CondParam) override {
5187 // A branch instruction will be inserted as "if (Cond) goto epilogue".
5188 // Cond is normalized for such use.
5189 // The predecessors of the branch are assumed to have already been inserted.
5190 CondParam = Cond;
5191 return {};
5192 }
5193
5194 void setPreheader(MachineBasicBlock *NewPreheader) override {}
5195
5196 void adjustTripCount(int TripCountAdjust) override {}
5197};
5198} // namespace
5199
5200std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
5202 MachineBasicBlock *TBB = nullptr, *FBB = nullptr;
5204 if (analyzeBranch(*LoopBB, TBB, FBB, Cond, /*AllowModify=*/false))
5205 return nullptr;
5206
5207 // Infinite loops are not supported
5208 if (TBB == LoopBB && FBB == LoopBB)
5209 return nullptr;
5210
5211 // Must be conditional branch
5212 if (FBB == nullptr)
5213 return nullptr;
5214
5215 assert((TBB == LoopBB || FBB == LoopBB) &&
5216 "The Loop must be a single-basic-block loop");
5217
5218 // Normalization for createTripCountGreaterCondition()
5219 if (TBB == LoopBB)
5221
5222 const MachineRegisterInfo &MRI = LoopBB->getParent()->getRegInfo();
5223 auto FindRegDef = [&MRI](MachineOperand &Op) -> const MachineInstr * {
5224 if (!Op.isReg())
5225 return nullptr;
5226 Register Reg = Op.getReg();
5227 if (!Reg.isVirtual())
5228 return nullptr;
5229 return MRI.getVRegDef(Reg);
5230 };
5231
5232 const MachineInstr *LHS = FindRegDef(Cond[1]);
5233 const MachineInstr *RHS = FindRegDef(Cond[2]);
5234 if (LHS && LHS->isPHI())
5235 return nullptr;
5236 if (RHS && RHS->isPHI())
5237 return nullptr;
5238
5239 return std::make_unique<RISCVPipelinerLoopInfo>(LHS, RHS, Cond);
5240}
5241
5242// FIXME: We should remove this if we have a default generic scheduling model.
5244 unsigned RVVMCOpcode = RISCV::getRVVMCOpcode(Opc);
5245 Opc = RVVMCOpcode ? RVVMCOpcode : Opc;
5246 switch (Opc) {
5247 default:
5248 return false;
5249 // Integer div/rem.
5250 case RISCV::DIV:
5251 case RISCV::DIVW:
5252 case RISCV::DIVU:
5253 case RISCV::DIVUW:
5254 case RISCV::REM:
5255 case RISCV::REMW:
5256 case RISCV::REMU:
5257 case RISCV::REMUW:
5258 // Floating-point div/sqrt.
5259 case RISCV::FDIV_H:
5260 case RISCV::FDIV_S:
5261 case RISCV::FDIV_D:
5262 case RISCV::FDIV_H_INX:
5263 case RISCV::FDIV_S_INX:
5264 case RISCV::FDIV_D_INX:
5265 case RISCV::FDIV_D_IN32X:
5266 case RISCV::FSQRT_H:
5267 case RISCV::FSQRT_S:
5268 case RISCV::FSQRT_D:
5269 case RISCV::FSQRT_H_INX:
5270 case RISCV::FSQRT_S_INX:
5271 case RISCV::FSQRT_D_INX:
5272 case RISCV::FSQRT_D_IN32X:
5273 // Vector integer div/rem
5274 case RISCV::VDIV_VV:
5275 case RISCV::VDIV_VX:
5276 case RISCV::VDIVU_VV:
5277 case RISCV::VDIVU_VX:
5278 case RISCV::VREM_VV:
5279 case RISCV::VREM_VX:
5280 case RISCV::VREMU_VV:
5281 case RISCV::VREMU_VX:
5282 // Vector floating-point div/sqrt.
5283 case RISCV::VFDIV_VV:
5284 case RISCV::VFDIV_VF:
5285 case RISCV::VFRDIV_VF:
5286 case RISCV::VFSQRT_V:
5287 case RISCV::VFRSQRT7_V:
5288 return true;
5289 }
5290}
5291
5292bool RISCVInstrInfo::isVRegCopy(const MachineInstr *MI, unsigned LMul) const {
5293 if (MI->getOpcode() != TargetOpcode::COPY)
5294 return false;
5295 const MachineRegisterInfo &MRI = MI->getMF()->getRegInfo();
5296 const TargetRegisterInfo *TRI = MRI.getTargetRegisterInfo();
5297
5298 Register DstReg = MI->getOperand(0).getReg();
5299 const TargetRegisterClass *RC = DstReg.isVirtual()
5300 ? MRI.getRegClass(DstReg)
5301 : TRI->getMinimalPhysRegClass(DstReg);
5302
5304 return false;
5305
5306 if (!LMul)
5307 return true;
5308
5309 // TODO: Perhaps we could distinguish segment register classes (e.g. VRN3M2)
5310 // in the future.
5311 auto [RCLMul, RCFractional] =
5313 return (!RCFractional && LMul == RCLMul) || (RCFractional && LMul == 1);
5314}
unsigned SubReg
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerDefault
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
SmallVector< int16_t, MAX_SRC_OPERANDS_NUM > OperandIndices
@ Scaled
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
basic Basic Alias true
#define clEnumValN(ENUMVAL, FLAGNAME, DESC)
const HexagonInstrInfo * TII
#define _
IRTranslator LLVM IR MI
Module.h This file contains the declarations for the Module class.
const size_t AbstractManglingParser< Derived, Alloc >::NumOps
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
Register Reg
Register const TargetRegisterInfo * TRI
Promote Memory to Register
Definition Mem2Reg.cpp:110
This file provides utility analysis objects describing memory locations.
MachineInstr unsigned OpIdx
uint64_t IntrinsicInst * II
static bool cannotInsertTailCall(const MachineBasicBlock &MBB)
#define CASE_VFMA_CHANGE_OPCODE_SPLATS(OLDOP, NEWOP)
#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS_ALT(OP)
#define CASE_FP_WIDEOP_OPCODE_LMULS(OP)
#define CASE_OPERAND_SIMM(NUM)
static std::optional< unsigned > getLMULForRVVWholeLoadStore(unsigned Opcode)
#define CASE_VFMA_CHANGE_OPCODE_VV(OLDOP, NEWOP)
static bool analyzeCandidate(outliner::Candidate &C)
static unsigned getFPFusedMultiplyOpcode(unsigned RootOpc, unsigned Pattern)
std::optional< unsigned > getFoldedOpcode(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, const RISCVSubtarget &ST)
#define RVV_OPC_LMUL_CASE(OPC, INV)
#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS(OP)
static void combineFPFusedMultiply(MachineInstr &Root, MachineInstr &Prev, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs)
static unsigned getAddendOperandIdx(unsigned Pattern)
#define CASE_RVV_OPCODE_UNMASK(OP)
#define CASE_WIDEOP_CHANGE_OPCODE_LMULS(OP)
static cl::opt< bool > PreferWholeRegisterMove("riscv-prefer-whole-register-move", cl::init(false), cl::Hidden, cl::desc("Prefer whole register move for vector registers."))
#define CASE_VFMA_SPLATS(OP)
unsigned getPredicatedOpcode(unsigned Opcode)
#define CASE_FP_WIDEOP_OPCODE_LMULS_ALT(OP)
#define CASE_WIDEOP_OPCODE_LMULS(OP)
static bool isMIReadsReg(const MachineInstr &MI, const TargetRegisterInfo *TRI, MCRegister RegNo)
#define OPCODE_LMUL_MASK_CASE(OPC)
static bool isFSUB(unsigned Opc)
#define CASE_VMA_CHANGE_OPCODE_LMULS(OLDOP, NEWOP, TYPE)
#define CASE_RVV_OPCODE(OP)
static std::optional< int64_t > getEffectiveImm(const MachineOperand &MO)
#define CASE_VFMA_OPCODE_VV(OP)
MachineOutlinerConstructionID
#define CASE_RVV_OPCODE_WIDEN(OP)
static unsigned getLoadPredicatedOpcode(unsigned Opcode)
static unsigned getSHXADDUWShiftAmount(unsigned Opc)
#define CASE_VMA_OPCODE_LMULS(OP, TYPE)
static bool isConvertibleToVMV_V_V(const RISCVSubtarget &STI, const MachineBasicBlock &MBB, MachineBasicBlock::const_iterator MBBI, MachineBasicBlock::const_iterator &DefMBBI, RISCVVType::VLMUL LMul)
static bool isFMUL(unsigned Opc)
static unsigned getInverseXqcicmOpcode(unsigned Opcode)
static bool getFPPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce)
#define OPCODE_LMUL_CASE(OPC)
#define CASE_OPERAND_UIMM(NUM)
static bool canCombineShiftIntoShXAdd(const MachineBasicBlock &MBB, const MachineOperand &MO, unsigned OuterShiftAmt)
Utility routine that checks if.
static bool isCandidatePatchable(const MachineBasicBlock &MBB)
static bool isFADD(unsigned Opc)
static void genShXAddAddShift(MachineInstr &Root, unsigned AddOpIdx, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
static bool isLoadImm(const MachineInstr *MI, int64_t &Imm)
static bool isMIModifiesReg(const MachineInstr &MI, const TargetRegisterInfo *TRI, MCRegister RegNo)
static bool canCombineFPFusedMultiply(const MachineInstr &Root, const MachineOperand &MO, bool DoRegPressureReduce)
static bool getSHXADDPatterns(const MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getFPFusedMultiplyPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce)
static cl::opt< MachineTraceStrategy > ForceMachineCombinerStrategy("riscv-force-machine-combiner-strategy", cl::Hidden, cl::desc("Force machine combiner to use a specific strategy for machine " "trace metrics evaluation."), cl::init(MachineTraceStrategy::TS_NumStrategies), cl::values(clEnumValN(MachineTraceStrategy::TS_Local, "local", "Local strategy."), clEnumValN(MachineTraceStrategy::TS_MinInstrCount, "min-instr", "MinInstrCount strategy.")))
static unsigned getSHXADDShiftAmount(unsigned Opc)
#define CASE_RVV_OPCODE_MASK(OP)
#define RVV_OPC_LMUL_MASK_CASE(OPC, INV)
static MachineInstr * canFoldAsPredicatedOp(Register Reg, const MachineRegisterInfo &MRI, const TargetInstrInfo *TII, const RISCVSubtarget &STI)
Identify instructions that can be folded into a CCMOV instruction, and return the defining instructio...
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool memOpsHaveSameBasePtr(const MachineInstr &MI1, ArrayRef< const MachineOperand * > BaseOps1, const MachineInstr &MI2, ArrayRef< const MachineOperand * > BaseOps2)
This file contains some templates that are useful if you are working with the STL at all.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
Definition Value.cpp:487
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
Definition Statistic.h:171
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
static cl::opt< unsigned > CacheLineSize("cache-line-size", cl::init(0), cl::Hidden, cl::desc("Use this to override the target cache line size when " "specified by the user."))
Value * RHS
Value * LHS
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
const T & front() const
front - Get the first element.
Definition ArrayRef.h:145
bool empty() const
empty - Check if the array is empty.
Definition ArrayRef.h:137
static LLVM_ABI DILocation * getMergedLocation(DILocation *LocA, DILocation *LocB)
Attempts to merge LocA and LocB into a single location; see DebugLoc::getMergedLocation for more deta...
bool isBigEndian() const
Definition DataLayout.h:215
A debug info location.
Definition DebugLoc.h:123
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
Definition DenseMap.h:241
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Definition Function.h:703
LiveInterval - This class represents the liveness of a register, or stack slot.
LiveInterval & getInterval(Register Reg)
SlotIndex ReplaceMachineInstrInMaps(MachineInstr &MI, MachineInstr &NewMI)
const Segment * getSegmentContaining(SlotIndex Idx) const
Return the segment that contains the specified index, or null if there is none.
LLVM_ABI void replaceKillInstruction(Register Reg, MachineInstr &OldMI, MachineInstr &NewMI)
replaceKillInstruction - Update register kill info by replacing a kill instruction with a new one.
bool hasValue() const
static LocationSize precise(uint64_t Value)
TypeSize getValue() const
MCInstBuilder & addReg(MCRegister Reg)
Add a new register operand.
MCInstBuilder & addImm(int64_t Val)
Add a new integer immediate operand.
Instances of this class represent a single low-level machine instruction.
Definition MCInst.h:188
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
bool isConditionalBranch() const
Return true if this is a branch which may fall through to the next instruction or may transfer contro...
This holds information about one operand of a machine instruction, indicating the register class for ...
Definition MCInstrDesc.h:86
Wrapper class representing physical registers. Should be passed by value.
Definition MCRegister.h:41
const FeatureBitset & getFeatureBits() const
Set of metadata that should be preserved when using BuildMI().
MachineInstrBundleIterator< const MachineInstr > const_iterator
MachineInstrBundleIterator< MachineInstr, true > reverse_iterator
Instructions::const_iterator const_instr_iterator
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
MachineInstrBundleIterator< MachineInstr > iterator
MachineInstrBundleIterator< const MachineInstr, true > const_reverse_iterator
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
void setStackID(int ObjectIdx, uint8_t ID)
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & cloneMemRefs(const MachineInstr &OtherMI) const
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & copyImplicitOps(const MachineInstr &OtherMI) const
Copy all the implicit operands from OtherMI onto this one.
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool isReturn(QueryType Type=AnyInBundle) const
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool modifiesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr modifies (fully define or partially define) the specified register.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool hasOneMemOperand() const
Return true if this instruction has exactly one MachineMemOperand.
mmo_iterator memoperands_begin() const
Access to memory operands of the instruction.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI void clearKillInfo()
Clears kill flags on all operands.
A description of a memory reference used in the backend.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
MachineOperand class - Representation of each machine instruction operand.
int64_t getImm() const
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
@ MO_Immediate
Immediate operand.
@ MO_Register
Register operand.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI MachineInstr * getVRegDef(Register Reg) const
getVRegDef - Return the machine instr that defines the specified virtual register or null if none is ...
A Module instance is used to store all the information related to an LLVM module.
Definition Module.h:67
MI-level patchpoint operands.
Definition StackMaps.h:77
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
Definition StackMaps.h:105
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool IsKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
MachineInstr * convertToThreeAddress(MachineInstr &MI, LiveVariables *LV, LiveIntervals *LIS) const override
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
std::optional< std::unique_ptr< outliner::OutlinedFunction > > getOutliningCandidateInfo(const MachineModuleInfo &MMI, std::vector< outliner::Candidate > &RepeatedSequenceLocs, unsigned MinRepeats) const override
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg) const override
void movImm(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, Register DstReg, uint64_t Val, MachineInstr::MIFlag Flag=MachineInstr::NoFlags, bool DstRenamable=false, bool DstIsDead=false) const
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
void mulImm(MachineFunction &MF, MachineBasicBlock &MBB, MachineBasicBlock::iterator II, const DebugLoc &DL, Register DestReg, uint32_t Amt, MachineInstr::MIFlag Flag) const
Generate code to multiply the value in DestReg by Amt - handles all the common optimizations for this...
static bool isPairableLdStInstOpc(unsigned Opc)
Return true if pairing the given load or store may be paired with another.
RISCVInstrInfo(const RISCVSubtarget &STI)
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DstReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, unsigned SubReg=0, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
bool isFunctionSafeToOutlineFrom(MachineFunction &MF, bool OutlineFromLinkOnceODRs) const override
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &dl, int *BytesAdded=nullptr) const override
bool hasReassociableSibling(const MachineInstr &Inst, bool &Commuted) const override
static bool isLdStSafeToPair(const MachineInstr &LdSt, const TargetRegisterInfo *TRI)
void copyPhysRegVector(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, MCRegister DstReg, MCRegister SrcReg, bool KillSrc, const TargetRegisterClass *RegClass) const
bool isReMaterializableImpl(const MachineInstr &MI) const override
MachineInstr * optimizeSelect(MachineInstr &MI, SmallPtrSetImpl< MachineInstr * > &SeenMIs, bool) const override
bool isVRegCopy(const MachineInstr *MI, unsigned LMul=0) const
Return true if MI is a COPY to a vector register of a specific LMul, or any kind of vector registers ...
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
bool isAsCheapAsAMove(const MachineInstr &MI) const override
bool verifyInstruction(const MachineInstr &MI, StringRef &ErrInfo) const override
bool getMemOperandWithOffsetWidth(const MachineInstr &LdSt, const MachineOperand *&BaseOp, int64_t &Offset, LocationSize &Width, const TargetRegisterInfo *TRI) const
unsigned getTailDuplicateSize(CodeGenOptLevel OptLevel) const override
void getReassociateOperandIndices(const MachineInstr &Root, unsigned Pattern, std::array< unsigned, 5 > &OperandIndices) const override
const RISCVSubtarget & STI
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
std::optional< unsigned > getInverseOpcode(unsigned Opcode) const override
bool simplifyInstruction(MachineInstr &MI) const override
ArrayRef< std::pair< unsigned, const char * > > getSerializableDirectMachineOperandTargetFlags() const override
outliner::InstrType getOutliningTypeImpl(const MachineModuleInfo &MMI, MachineBasicBlock::iterator &MBBI, unsigned Flags) const override
MachineTraceStrategy getMachineCombinerTraceStrategy() const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
std::optional< RegImmPair > isAddImmediate(const MachineInstr &MI, Register Reg) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
ArrayRef< std::pair< MachineMemOperand::Flags, const char * > > getSerializableMachineMemOperandTargetFlags() const override
MCInst getNop() const override
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
bool isMBBSafeToOutlineFrom(MachineBasicBlock &MBB, unsigned &Flags) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
void buildOutlinedFrame(MachineBasicBlock &MBB, MachineFunction &MF, const outliner::OutlinedFunction &OF) const override
void finalizeInsInstrs(MachineInstr &Root, unsigned &Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs) const override
std::pair< unsigned, unsigned > decomposeMachineOperandsTargetFlags(unsigned TF) const override
MachineInstr * commuteInstructionImpl(MachineInstr &MI, bool NewMI, unsigned OpIdx1, unsigned OpIdx2) const override
bool hasReassociableOperands(const MachineInstr &Inst, const MachineBasicBlock *MBB) const override
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
std::string createMIROperandComment(const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx, const TargetRegisterInfo *TRI) const override
bool shouldOutlineFromFunctionByDefault(MachineFunction &MF) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, Register DstReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, unsigned &SrcOpIdx2) const override
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify) const override
MachineBasicBlock::iterator insertOutlinedCall(Module &M, MachineBasicBlock &MBB, MachineBasicBlock::iterator &It, MachineFunction &MF, outliner::Candidate &C) const override
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
static RISCVCC::CondCode getCondFromBranchOpc(unsigned Opc)
bool isAssociativeAndCommutative(const MachineInstr &Inst, bool Invert) const override
CombinerObjective getCombinerObjective(unsigned Pattern) const override
bool isHighLatencyDef(int Opc) const override
static bool evaluateCondBranch(RISCVCC::CondCode CC, int64_t C0, int64_t C1)
Return the result of the evaluation of C0 CC C1, where CC is a RISCVCC::CondCode.
bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const override
bool optimizeCondBranch(MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
bool analyzeSelect(const MachineInstr &MI, SmallVectorImpl< MachineOperand > &Cond, unsigned &TrueOp, unsigned &FalseOp, bool &Optimizable) const override
static bool isFromLoadImm(const MachineRegisterInfo &MRI, const MachineOperand &Op, int64_t &Imm)
Return true if the operand is a load immediate instruction and sets Imm to the immediate value.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
RISCVMachineFunctionInfo - This class is derived from MachineFunctionInfo and contains private RISCV-...
const RISCVRegisterInfo * getRegisterInfo() const override
Wrapper class representing virtual and physical registers.
Definition Register.h:20
constexpr bool isValid() const
Definition Register.h:112
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
Definition Register.h:79
SlotIndex - An opaque wrapper around machine indexes.
Definition SlotIndexes.h:66
SlotIndex getRegSlot(bool EC=false) const
Returns the register use/def slot in the current instruction for a normal or early-clobber def.
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
Definition StackMaps.h:36
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
Definition StackMaps.h:51
MI-level Statepoint operands.
Definition StackMaps.h:159
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
Definition StackMaps.h:208
StringRef - Represent a constant reference to a string, i.e.
Definition StringRef.h:55
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, unsigned &SrcOpIdx2) const
Returns true iff the routine could find two commutable operands in the given machine instruction.
virtual bool hasReassociableOperands(const MachineInstr &Inst, const MachineBasicBlock *MBB) const
Return true when \P Inst has reassociable operands in the same \P MBB.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isReMaterializableImpl(const MachineInstr &MI) const
For instructions with opcodes for which the M_REMATERIALIZABLE flag is set, this hook lets the target...
virtual bool isMBBSafeToOutlineFrom(MachineBasicBlock &MBB, unsigned &Flags) const
Optional target hook that returns true if MBB is safe to outline from, and returns any target-specifi...
virtual void getReassociateOperandIndices(const MachineInstr &Root, unsigned Pattern, std::array< unsigned, 5 > &OperandIndices) const
The returned array encodes the operand index for each parameter because the operands may be commuted;...
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual MachineInstr * commuteInstructionImpl(MachineInstr &MI, bool NewMI, unsigned OpIdx1, unsigned OpIdx2) const
This method commutes the operands of the given machine instruction MI.
virtual bool hasReassociableSibling(const MachineInstr &Inst, bool &Commuted) const
Return true when \P Inst has reassociable sibling.
virtual std::string createMIROperandComment(const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx, const TargetRegisterInfo *TRI) const
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
const uint8_t TSFlags
Configurable target specific flags.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition TypeSize.h:343
static constexpr TypeSize getZero()
Definition TypeSize.h:349
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Definition TypeSize.h:346
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
CondCode getInverseBranchCondition(CondCode)
unsigned getBrCond(CondCode CC, unsigned SelectOpc=0)
static bool isValidRoundingMode(unsigned Mode)
static unsigned getVecPolicyOpNum(const MCInstrDesc &Desc)
static bool usesMaskPolicy(uint64_t TSFlags)
static bool hasRoundModeOp(uint64_t TSFlags)
static unsigned getVLOpNum(const MCInstrDesc &Desc)
static bool hasVLOp(uint64_t TSFlags)
static MCRegister getTailExpandUseRegNo(const FeatureBitset &FeatureBits)
static int getFRMOpNum(const MCInstrDesc &Desc)
static bool hasVecPolicyOp(uint64_t TSFlags)
static bool usesVXRM(uint64_t TSFlags)
static bool isRVVWideningReduction(uint64_t TSFlags)
static unsigned getSEWOpNum(const MCInstrDesc &Desc)
static bool hasSEWOp(uint64_t TSFlags)
static bool isFirstDefTiedToFirstUse(const MCInstrDesc &Desc)
InstSeq generateInstSeq(int64_t Val, const MCSubtargetInfo &STI)
SmallVector< Inst, 8 > InstSeq
Definition RISCVMatInt.h:43
@ OPERAND_SIMM10_LSB0000_NONZERO
static unsigned getNF(uint8_t TSFlags)
static RISCVVType::VLMUL getLMul(uint8_t TSFlags)
static bool isTailAgnostic(unsigned VType)
LLVM_ABI void printXSfmmVType(unsigned VType, raw_ostream &OS)
LLVM_ABI std::pair< unsigned, bool > decodeVLMUL(VLMUL VLMul)
static bool isValidSEW(unsigned SEW)
LLVM_ABI void printVType(unsigned VType, raw_ostream &OS)
static bool isValidXSfmmVType(unsigned VTypeI)
static unsigned getSEW(unsigned VType)
static VLMUL getVLMUL(unsigned VType)
bool hasEqualFRM(const MachineInstr &MI1, const MachineInstr &MI2)
bool isVLKnownLE(const MachineOperand &LHS, const MachineOperand &RHS)
Given two VL operands, do we know that LHS <= RHS?
unsigned getRVVMCOpcode(unsigned RVVPseudoOpcode)
unsigned getDestLog2EEW(const MCInstrDesc &Desc, unsigned Log2SEW)
std::optional< unsigned > getVectorLowDemandedScalarBits(unsigned Opcode, unsigned Log2SEW)
std::optional< std::pair< unsigned, unsigned > > isRVVSpillForZvlsseg(unsigned Opcode)
static constexpr unsigned RVVBitsPerBlock
bool isRVVSpill(const MachineInstr &MI)
static constexpr unsigned RVVBytesPerBlock
static constexpr int64_t VLMaxSentinel
bool isVectorCopy(const TargetRegisterInfo *TRI, const MachineInstr &MI)
Return true if MI is a copy that will be lowered to one or more vmvNr.vs.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Dead
Unused definition.
@ Define
Register definition.
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
ValuesClass values(OptsTy... Options)
Helper to build a ValuesClass by forwarding a variable number of arguments as an initializer list to ...
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
This is an optimization pass for GlobalISel generic memory operations.
Definition Types.h:26
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
Definition STLExtras.h:316
@ Offset
Definition DWP.cpp:532
@ SHXADD_ADD_SLLI_OP2
@ SHXADD_ADD_SLLI_OP1
MachineTraceStrategy
Strategies for selecting traces.
@ TS_MinInstrCount
Select the trace through a block that has the fewest instructions.
@ TS_Local
Select the trace that contains only the current basic block.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1737
static const MachineMemOperand::Flags MONontemporalBit1
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
Definition MathExtras.h:165
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
Definition STLExtras.h:2530
bool isValidAtomicOrdering(Int I)
static const MachineMemOperand::Flags MONontemporalBit0
unsigned getDeadRegState(bool B)
Op::Description Desc
constexpr bool has_single_bit(T Value) noexcept
Definition bit.h:147
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1744
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition MathExtras.h:331
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
Definition Error.cpp:167
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
constexpr bool isUInt(uint64_t x)
Checks if an unsigned integer fits into the given bit width.
Definition MathExtras.h:189
CodeGenOptLevel
Code generation optimization level.
Definition CodeGen.h:82
int isShifted359(T Value, int &Shift)
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
unsigned getKillRegState(bool B)
unsigned getRenamableRegState(bool B)
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
constexpr bool isShiftedInt(int64_t x)
Checks if a signed integer is an N bit number shifted left by S.
Definition MathExtras.h:182
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
Definition STLExtras.h:2168
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
Definition MathExtras.h:572
LLVM_ABI const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=MaxLookupSearchDepth)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
constexpr bool isShiftedUInt(uint64_t x)
Checks if a unsigned integer is an N bit number shifted left by S.
Definition MathExtras.h:198
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
#define N
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
This represents a simple continuous liveness interval for a value.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
static bool isRVVRegClass(const TargetRegisterClass *RC)
Used to describe a register and immediate addition.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.