LLVM 23.0.0git
RISCVInstrInfo.cpp
Go to the documentation of this file.
1//===-- RISCVInstrInfo.cpp - RISC-V Instruction Information -----*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file contains the RISC-V implementation of the TargetInstrInfo class.
10//
11//===----------------------------------------------------------------------===//
12
13#include "RISCVInstrInfo.h"
16#include "RISCV.h"
18#include "RISCVSubtarget.h"
19#include "llvm/ADT/STLExtras.h"
21#include "llvm/ADT/Statistic.h"
33#include "llvm/IR/Module.h"
34#include "llvm/MC/MCDwarf.h"
38
39using namespace llvm;
40
41#define GEN_CHECK_COMPRESS_INSTR
42#include "RISCVGenCompressInstEmitter.inc"
43
44#define GET_INSTRINFO_CTOR_DTOR
45#define GET_INSTRINFO_NAMED_OPS
46#include "RISCVGenInstrInfo.inc"
47
48#define DEBUG_TYPE "riscv-instr-info"
49STATISTIC(NumVRegSpilled,
50 "Number of registers within vector register groups spilled");
51STATISTIC(NumVRegReloaded,
52 "Number of registers within vector register groups reloaded");
53
55 "riscv-prefer-whole-register-move", cl::init(false), cl::Hidden,
56 cl::desc("Prefer whole register move for vector registers."));
57
59 "riscv-force-machine-combiner-strategy", cl::Hidden,
60 cl::desc("Force machine combiner to use a specific strategy for machine "
61 "trace metrics evaluation."),
64 "Local strategy."),
66 "MinInstrCount strategy.")));
67
69
70using namespace RISCV;
71
72#define GET_RISCVVPseudosTable_IMPL
73#include "RISCVGenSearchableTables.inc"
74
75} // namespace llvm::RISCVVPseudosTable
76
77namespace llvm::RISCV {
78
79#define GET_RISCVMaskedPseudosTable_IMPL
80#include "RISCVGenSearchableTables.inc"
81
82} // end namespace llvm::RISCV
83
85 : RISCVGenInstrInfo(STI, RegInfo, RISCV::ADJCALLSTACKDOWN,
86 RISCV::ADJCALLSTACKUP),
87 RegInfo(STI.getHwMode()), STI(STI) {}
88
89#define GET_INSTRINFO_HELPERS
90#include "RISCVGenInstrInfo.inc"
91
93 if (STI.hasStdExtZca())
94 return MCInstBuilder(RISCV::C_NOP);
95 return MCInstBuilder(RISCV::ADDI)
96 .addReg(RISCV::X0)
97 .addReg(RISCV::X0)
98 .addImm(0);
99}
100
102 int &FrameIndex) const {
103 TypeSize Dummy = TypeSize::getZero();
104 return isLoadFromStackSlot(MI, FrameIndex, Dummy);
105}
106
107static std::optional<unsigned> getLMULForRVVWholeLoadStore(unsigned Opcode) {
108 switch (Opcode) {
109 default:
110 return std::nullopt;
111 case RISCV::VS1R_V:
112 case RISCV::VL1RE8_V:
113 case RISCV::VL1RE16_V:
114 case RISCV::VL1RE32_V:
115 case RISCV::VL1RE64_V:
116 return 1;
117 case RISCV::VS2R_V:
118 case RISCV::VL2RE8_V:
119 case RISCV::VL2RE16_V:
120 case RISCV::VL2RE32_V:
121 case RISCV::VL2RE64_V:
122 return 2;
123 case RISCV::VS4R_V:
124 case RISCV::VL4RE8_V:
125 case RISCV::VL4RE16_V:
126 case RISCV::VL4RE32_V:
127 case RISCV::VL4RE64_V:
128 return 4;
129 case RISCV::VS8R_V:
130 case RISCV::VL8RE8_V:
131 case RISCV::VL8RE16_V:
132 case RISCV::VL8RE32_V:
133 case RISCV::VL8RE64_V:
134 return 8;
135 }
136}
137
139 int &FrameIndex,
140 TypeSize &MemBytes) const {
141 switch (MI.getOpcode()) {
142 default:
143 return 0;
144 case RISCV::LB:
145 case RISCV::LBU:
146 MemBytes = TypeSize::getFixed(1);
147 break;
148 case RISCV::LH:
149 case RISCV::LH_INX:
150 case RISCV::LHU:
151 case RISCV::FLH:
152 MemBytes = TypeSize::getFixed(2);
153 break;
154 case RISCV::LW:
155 case RISCV::LW_INX:
156 case RISCV::FLW:
157 case RISCV::LWU:
158 MemBytes = TypeSize::getFixed(4);
159 break;
160 case RISCV::LD:
161 case RISCV::LD_RV32:
162 case RISCV::FLD:
163 MemBytes = TypeSize::getFixed(8);
164 break;
165 case RISCV::VL1RE8_V:
166 case RISCV::VL2RE8_V:
167 case RISCV::VL4RE8_V:
168 case RISCV::VL8RE8_V:
169 if (!MI.getOperand(1).isFI())
170 return Register();
171 FrameIndex = MI.getOperand(1).getIndex();
172 unsigned LMUL = *getLMULForRVVWholeLoadStore(MI.getOpcode());
174 return MI.getOperand(0).getReg();
175 }
176
177 if (MI.getOperand(1).isFI() && MI.getOperand(2).isImm() &&
178 MI.getOperand(2).getImm() == 0) {
179 FrameIndex = MI.getOperand(1).getIndex();
180 return MI.getOperand(0).getReg();
181 }
182
183 return 0;
184}
185
187 int &FrameIndex) const {
188 TypeSize Dummy = TypeSize::getZero();
189 return isStoreToStackSlot(MI, FrameIndex, Dummy);
190}
191
193 int &FrameIndex,
194 TypeSize &MemBytes) const {
195 switch (MI.getOpcode()) {
196 default:
197 return 0;
198 case RISCV::SB:
199 MemBytes = TypeSize::getFixed(1);
200 break;
201 case RISCV::SH:
202 case RISCV::SH_INX:
203 case RISCV::FSH:
204 MemBytes = TypeSize::getFixed(2);
205 break;
206 case RISCV::SW:
207 case RISCV::SW_INX:
208 case RISCV::FSW:
209 MemBytes = TypeSize::getFixed(4);
210 break;
211 case RISCV::SD:
212 case RISCV::SD_RV32:
213 case RISCV::FSD:
214 MemBytes = TypeSize::getFixed(8);
215 break;
216 case RISCV::VS1R_V:
217 case RISCV::VS2R_V:
218 case RISCV::VS4R_V:
219 case RISCV::VS8R_V:
220 if (!MI.getOperand(1).isFI())
221 return Register();
222 FrameIndex = MI.getOperand(1).getIndex();
223 unsigned LMUL = *getLMULForRVVWholeLoadStore(MI.getOpcode());
225 return MI.getOperand(0).getReg();
226 }
227
228 if (MI.getOperand(1).isFI() && MI.getOperand(2).isImm() &&
229 MI.getOperand(2).getImm() == 0) {
230 FrameIndex = MI.getOperand(1).getIndex();
231 return MI.getOperand(0).getReg();
232 }
233
234 return 0;
235}
236
238 const MachineInstr &MI) const {
239 switch (RISCV::getRVVMCOpcode(MI.getOpcode())) {
240 case RISCV::VMV_V_X:
241 case RISCV::VFMV_V_F:
242 case RISCV::VMV_V_I:
243 case RISCV::VMV_S_X:
244 case RISCV::VFMV_S_F:
245 case RISCV::VID_V:
246 return MI.getOperand(1).isUndef();
247 default:
249 }
250}
251
252static bool forwardCopyWillClobberTuple(unsigned DstReg, unsigned SrcReg,
253 unsigned NumRegs) {
254 return DstReg > SrcReg && (DstReg - SrcReg) < NumRegs;
255}
256
258 const MachineBasicBlock &MBB,
261 RISCVVType::VLMUL LMul) {
263 return false;
264
265 assert(MBBI->getOpcode() == TargetOpcode::COPY &&
266 "Unexpected COPY instruction.");
267 Register SrcReg = MBBI->getOperand(1).getReg();
269
270 bool FoundDef = false;
271 bool FirstVSetVLI = false;
272 unsigned FirstSEW = 0;
273 while (MBBI != MBB.begin()) {
274 --MBBI;
275 if (MBBI->isMetaInstruction())
276 continue;
277
278 if (RISCVInstrInfo::isVectorConfigInstr(*MBBI)) {
279 // There is a vsetvli between COPY and source define instruction.
280 // vy = def_vop ... (producing instruction)
281 // ...
282 // vsetvli
283 // ...
284 // vx = COPY vy
285 if (!FoundDef) {
286 if (!FirstVSetVLI) {
287 FirstVSetVLI = true;
288 unsigned FirstVType = MBBI->getOperand(2).getImm();
289 RISCVVType::VLMUL FirstLMul = RISCVVType::getVLMUL(FirstVType);
290 FirstSEW = RISCVVType::getSEW(FirstVType);
291 // The first encountered vsetvli must have the same lmul as the
292 // register class of COPY.
293 if (FirstLMul != LMul)
294 return false;
295 }
296 // Only permit `vsetvli x0, x0, vtype` between COPY and the source
297 // define instruction.
298 if (!RISCVInstrInfo::isVLPreservingConfig(*MBBI))
299 return false;
300 continue;
301 }
302
303 // MBBI is the first vsetvli before the producing instruction.
304 unsigned VType = MBBI->getOperand(2).getImm();
305 // If there is a vsetvli between COPY and the producing instruction.
306 if (FirstVSetVLI) {
307 // If SEW is different, return false.
308 if (RISCVVType::getSEW(VType) != FirstSEW)
309 return false;
310 }
311
312 // If the vsetvli is tail undisturbed, keep the whole register move.
313 if (!RISCVVType::isTailAgnostic(VType))
314 return false;
315
316 // The checking is conservative. We only have register classes for
317 // LMUL = 1/2/4/8. We should be able to convert vmv1r.v to vmv.v.v
318 // for fractional LMUL operations. However, we could not use the vsetvli
319 // lmul for widening operations. The result of widening operation is
320 // 2 x LMUL.
321 return LMul == RISCVVType::getVLMUL(VType);
322 } else if (MBBI->isInlineAsm() || MBBI->isCall()) {
323 return false;
324 } else if (MBBI->getNumDefs()) {
325 // Check all the instructions which will change VL.
326 // For example, vleff has implicit def VL.
327 if (MBBI->modifiesRegister(RISCV::VL, /*TRI=*/nullptr))
328 return false;
329
330 // Only converting whole register copies to vmv.v.v when the defining
331 // value appears in the explicit operands.
332 for (const MachineOperand &MO : MBBI->explicit_operands()) {
333 if (!MO.isReg() || !MO.isDef())
334 continue;
335 if (!FoundDef && TRI->regsOverlap(MO.getReg(), SrcReg)) {
336 // We only permit the source of COPY has the same LMUL as the defined
337 // operand.
338 // There are cases we need to keep the whole register copy if the LMUL
339 // is different.
340 // For example,
341 // $x0 = PseudoVSETIVLI 4, 73 // vsetivli zero, 4, e16,m2,ta,m
342 // $v28m4 = PseudoVWADD_VV_M2 $v26m2, $v8m2
343 // # The COPY may be created by vlmul_trunc intrinsic.
344 // $v26m2 = COPY renamable $v28m2, implicit killed $v28m4
345 //
346 // After widening, the valid value will be 4 x e32 elements. If we
347 // convert the COPY to vmv.v.v, it will only copy 4 x e16 elements.
348 // FIXME: The COPY of subregister of Zvlsseg register will not be able
349 // to convert to vmv.v.[v|i] under the constraint.
350 if (MO.getReg() != SrcReg)
351 return false;
352
353 // In widening reduction instructions with LMUL_1 input vector case,
354 // only checking the LMUL is insufficient due to reduction result is
355 // always LMUL_1.
356 // For example,
357 // $x11 = PseudoVSETIVLI 1, 64 // vsetivli a1, 1, e8, m1, ta, mu
358 // $v8m1 = PseudoVWREDSUM_VS_M1 $v26, $v27
359 // $v26 = COPY killed renamable $v8
360 // After widening, The valid value will be 1 x e16 elements. If we
361 // convert the COPY to vmv.v.v, it will only copy 1 x e8 elements.
362 uint64_t TSFlags = MBBI->getDesc().TSFlags;
364 return false;
365
366 // If the producing instruction does not depend on vsetvli, do not
367 // convert COPY to vmv.v.v. For example, VL1R_V or PseudoVRELOAD.
368 if (!RISCVII::hasSEWOp(TSFlags) || !RISCVII::hasVLOp(TSFlags))
369 return false;
370
371 // Found the definition.
372 FoundDef = true;
373 DefMBBI = MBBI;
374 break;
375 }
376 }
377 }
378 }
379
380 return false;
381}
382
385 const DebugLoc &DL, MCRegister DstReg, MCRegister SrcReg, bool KillSrc,
386 const TargetRegisterClass *RegClass) const {
387 const RISCVRegisterInfo *TRI = STI.getRegisterInfo();
389 unsigned NF = RISCVRI::getNF(RegClass->TSFlags);
390
391 uint16_t SrcEncoding = TRI->getEncodingValue(SrcReg);
392 uint16_t DstEncoding = TRI->getEncodingValue(DstReg);
393 auto [LMulVal, Fractional] = RISCVVType::decodeVLMUL(LMul);
394 assert(!Fractional && "It is impossible be fractional lmul here.");
395 unsigned NumRegs = NF * LMulVal;
396 bool ReversedCopy =
397 forwardCopyWillClobberTuple(DstEncoding, SrcEncoding, NumRegs);
398 if (ReversedCopy) {
399 // If the src and dest overlap when copying a tuple, we need to copy the
400 // registers in reverse.
401 SrcEncoding += NumRegs - 1;
402 DstEncoding += NumRegs - 1;
403 }
404
405 unsigned I = 0;
406 auto GetCopyInfo = [&](uint16_t SrcEncoding, uint16_t DstEncoding)
407 -> std::tuple<RISCVVType::VLMUL, const TargetRegisterClass &, unsigned,
408 unsigned, unsigned> {
409 if (ReversedCopy) {
410 // For reversed copying, if there are enough aligned registers(8/4/2), we
411 // can do a larger copy(LMUL8/4/2).
412 // Besides, we have already known that DstEncoding is larger than
413 // SrcEncoding in forwardCopyWillClobberTuple, so the difference between
414 // DstEncoding and SrcEncoding should be >= LMUL value we try to use to
415 // avoid clobbering.
416 uint16_t Diff = DstEncoding - SrcEncoding;
417 if (I + 8 <= NumRegs && Diff >= 8 && SrcEncoding % 8 == 7 &&
418 DstEncoding % 8 == 7)
419 return {RISCVVType::LMUL_8, RISCV::VRM8RegClass, RISCV::VMV8R_V,
420 RISCV::PseudoVMV_V_V_M8, RISCV::PseudoVMV_V_I_M8};
421 if (I + 4 <= NumRegs && Diff >= 4 && SrcEncoding % 4 == 3 &&
422 DstEncoding % 4 == 3)
423 return {RISCVVType::LMUL_4, RISCV::VRM4RegClass, RISCV::VMV4R_V,
424 RISCV::PseudoVMV_V_V_M4, RISCV::PseudoVMV_V_I_M4};
425 if (I + 2 <= NumRegs && Diff >= 2 && SrcEncoding % 2 == 1 &&
426 DstEncoding % 2 == 1)
427 return {RISCVVType::LMUL_2, RISCV::VRM2RegClass, RISCV::VMV2R_V,
428 RISCV::PseudoVMV_V_V_M2, RISCV::PseudoVMV_V_I_M2};
429 // Or we should do LMUL1 copying.
430 return {RISCVVType::LMUL_1, RISCV::VRRegClass, RISCV::VMV1R_V,
431 RISCV::PseudoVMV_V_V_M1, RISCV::PseudoVMV_V_I_M1};
432 }
433
434 // For forward copying, if source register encoding and destination register
435 // encoding are aligned to 8/4/2, we can do a LMUL8/4/2 copying.
436 if (I + 8 <= NumRegs && SrcEncoding % 8 == 0 && DstEncoding % 8 == 0)
437 return {RISCVVType::LMUL_8, RISCV::VRM8RegClass, RISCV::VMV8R_V,
438 RISCV::PseudoVMV_V_V_M8, RISCV::PseudoVMV_V_I_M8};
439 if (I + 4 <= NumRegs && SrcEncoding % 4 == 0 && DstEncoding % 4 == 0)
440 return {RISCVVType::LMUL_4, RISCV::VRM4RegClass, RISCV::VMV4R_V,
441 RISCV::PseudoVMV_V_V_M4, RISCV::PseudoVMV_V_I_M4};
442 if (I + 2 <= NumRegs && SrcEncoding % 2 == 0 && DstEncoding % 2 == 0)
443 return {RISCVVType::LMUL_2, RISCV::VRM2RegClass, RISCV::VMV2R_V,
444 RISCV::PseudoVMV_V_V_M2, RISCV::PseudoVMV_V_I_M2};
445 // Or we should do LMUL1 copying.
446 return {RISCVVType::LMUL_1, RISCV::VRRegClass, RISCV::VMV1R_V,
447 RISCV::PseudoVMV_V_V_M1, RISCV::PseudoVMV_V_I_M1};
448 };
449
450 while (I != NumRegs) {
451 // For non-segment copying, we only do this once as the registers are always
452 // aligned.
453 // For segment copying, we may do this several times. If the registers are
454 // aligned to larger LMUL, we can eliminate some copyings.
455 auto [LMulCopied, RegClass, Opc, VVOpc, VIOpc] =
456 GetCopyInfo(SrcEncoding, DstEncoding);
457 auto [NumCopied, _] = RISCVVType::decodeVLMUL(LMulCopied);
458
460 if (LMul == LMulCopied &&
461 isConvertibleToVMV_V_V(STI, MBB, MBBI, DefMBBI, LMul)) {
462 Opc = VVOpc;
463 if (DefMBBI->getOpcode() == VIOpc)
464 Opc = VIOpc;
465 }
466
467 // Emit actual copying.
468 // For reversed copying, the encoding should be decreased.
469 MCRegister ActualSrcReg = TRI->findVRegWithEncoding(
470 RegClass, ReversedCopy ? (SrcEncoding - NumCopied + 1) : SrcEncoding);
471 MCRegister ActualDstReg = TRI->findVRegWithEncoding(
472 RegClass, ReversedCopy ? (DstEncoding - NumCopied + 1) : DstEncoding);
473
474 auto MIB = BuildMI(MBB, MBBI, DL, get(Opc), ActualDstReg);
475 bool UseVMV_V_I = RISCV::getRVVMCOpcode(Opc) == RISCV::VMV_V_I;
476 bool UseVMV = UseVMV_V_I || RISCV::getRVVMCOpcode(Opc) == RISCV::VMV_V_V;
477 if (UseVMV)
478 MIB.addReg(ActualDstReg, RegState::Undef);
479 if (UseVMV_V_I)
480 MIB = MIB.add(DefMBBI->getOperand(2));
481 else
482 MIB = MIB.addReg(ActualSrcReg, getKillRegState(KillSrc));
483 if (UseVMV) {
484 const MCInstrDesc &Desc = DefMBBI->getDesc();
485 MIB.add(DefMBBI->getOperand(RISCVII::getVLOpNum(Desc))); // AVL
486 unsigned Log2SEW =
487 DefMBBI->getOperand(RISCVII::getSEWOpNum(Desc)).getImm();
488 MIB.addImm(Log2SEW ? Log2SEW : 3); // SEW
489 MIB.addImm(0); // tu, mu
490 MIB.addReg(RISCV::VL, RegState::Implicit);
491 MIB.addReg(RISCV::VTYPE, RegState::Implicit);
492 }
493 // Add an implicit read of the original source to silence the verifier
494 // in the cases where some of the smaller VRs we're copying from might be
495 // undef, caused by the fact that the original, larger source VR might not
496 // be fully initialized at the time this COPY happens.
497 MIB.addReg(SrcReg, RegState::Implicit);
498
499 // If we are copying reversely, we should decrease the encoding.
500 SrcEncoding += (ReversedCopy ? -NumCopied : NumCopied);
501 DstEncoding += (ReversedCopy ? -NumCopied : NumCopied);
502 I += NumCopied;
503 }
504}
505
508 const DebugLoc &DL, Register DstReg,
509 Register SrcReg, bool KillSrc,
510 bool RenamableDest, bool RenamableSrc) const {
511 const TargetRegisterInfo *TRI = STI.getRegisterInfo();
512 unsigned KillFlag = getKillRegState(KillSrc);
513
514 if (RISCV::GPRRegClass.contains(DstReg, SrcReg)) {
515 BuildMI(MBB, MBBI, DL, get(RISCV::ADDI), DstReg)
516 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc))
517 .addImm(0);
518 return;
519 }
520
521 if (RISCV::GPRF16RegClass.contains(DstReg, SrcReg)) {
522 BuildMI(MBB, MBBI, DL, get(RISCV::PseudoMV_FPR16INX), DstReg)
523 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc));
524 return;
525 }
526
527 if (RISCV::GPRF32RegClass.contains(DstReg, SrcReg)) {
528 BuildMI(MBB, MBBI, DL, get(RISCV::PseudoMV_FPR32INX), DstReg)
529 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc));
530 return;
531 }
532
533 if (RISCV::GPRPairRegClass.contains(DstReg, SrcReg)) {
534 if (STI.isRV32() && STI.hasStdExtZdinx()) {
535 // On RV32_Zdinx, FMV.D will move a pair of registers to another pair of
536 // registers, in one instruction.
537 BuildMI(MBB, MBBI, DL, get(RISCV::FSGNJ_D_IN32X), DstReg)
538 .addReg(SrcReg, getRenamableRegState(RenamableSrc))
539 .addReg(SrcReg, KillFlag | getRenamableRegState(RenamableSrc));
540 return;
541 }
542
543 MCRegister EvenReg = TRI->getSubReg(SrcReg, RISCV::sub_gpr_even);
544 MCRegister OddReg = TRI->getSubReg(SrcReg, RISCV::sub_gpr_odd);
545 // We need to correct the odd register of X0_Pair.
546 if (OddReg == RISCV::DUMMY_REG_PAIR_WITH_X0)
547 OddReg = RISCV::X0;
548 assert(DstReg != RISCV::X0_Pair && "Cannot write to X0_Pair");
549
550 // Emit an ADDI for both parts of GPRPair.
551 BuildMI(MBB, MBBI, DL, get(RISCV::ADDI),
552 TRI->getSubReg(DstReg, RISCV::sub_gpr_even))
553 .addReg(EvenReg, KillFlag)
554 .addImm(0);
555 BuildMI(MBB, MBBI, DL, get(RISCV::ADDI),
556 TRI->getSubReg(DstReg, RISCV::sub_gpr_odd))
557 .addReg(OddReg, KillFlag)
558 .addImm(0);
559 return;
560 }
561
562 // Handle copy from csr
563 if (RISCV::VCSRRegClass.contains(SrcReg) &&
564 RISCV::GPRRegClass.contains(DstReg)) {
565 BuildMI(MBB, MBBI, DL, get(RISCV::CSRRS), DstReg)
566 .addImm(RISCVSysReg::lookupSysRegByName(TRI->getName(SrcReg))->Encoding)
567 .addReg(RISCV::X0);
568 return;
569 }
570
571 if (RISCV::FPR16RegClass.contains(DstReg, SrcReg)) {
572 unsigned Opc;
573 if (STI.hasStdExtZfh()) {
574 Opc = RISCV::FSGNJ_H;
575 } else {
576 assert(STI.hasStdExtF() &&
577 (STI.hasStdExtZfhmin() || STI.hasStdExtZfbfmin()) &&
578 "Unexpected extensions");
579 // Zfhmin/Zfbfmin doesn't have FSGNJ_H, replace FSGNJ_H with FSGNJ_S.
580 DstReg = TRI->getMatchingSuperReg(DstReg, RISCV::sub_16,
581 &RISCV::FPR32RegClass);
582 SrcReg = TRI->getMatchingSuperReg(SrcReg, RISCV::sub_16,
583 &RISCV::FPR32RegClass);
584 Opc = RISCV::FSGNJ_S;
585 }
586 BuildMI(MBB, MBBI, DL, get(Opc), DstReg)
587 .addReg(SrcReg, KillFlag)
588 .addReg(SrcReg, KillFlag);
589 return;
590 }
591
592 if (RISCV::FPR32RegClass.contains(DstReg, SrcReg)) {
593 BuildMI(MBB, MBBI, DL, get(RISCV::FSGNJ_S), DstReg)
594 .addReg(SrcReg, KillFlag)
595 .addReg(SrcReg, KillFlag);
596 return;
597 }
598
599 if (RISCV::FPR64RegClass.contains(DstReg, SrcReg)) {
600 BuildMI(MBB, MBBI, DL, get(RISCV::FSGNJ_D), DstReg)
601 .addReg(SrcReg, KillFlag)
602 .addReg(SrcReg, KillFlag);
603 return;
604 }
605
606 if (RISCV::FPR32RegClass.contains(DstReg) &&
607 RISCV::GPRRegClass.contains(SrcReg)) {
608 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_W_X), DstReg)
609 .addReg(SrcReg, KillFlag);
610 return;
611 }
612
613 if (RISCV::GPRRegClass.contains(DstReg) &&
614 RISCV::FPR32RegClass.contains(SrcReg)) {
615 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_X_W), DstReg)
616 .addReg(SrcReg, KillFlag);
617 return;
618 }
619
620 if (RISCV::FPR64RegClass.contains(DstReg) &&
621 RISCV::GPRRegClass.contains(SrcReg)) {
622 assert(STI.getXLen() == 64 && "Unexpected GPR size");
623 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_D_X), DstReg)
624 .addReg(SrcReg, KillFlag);
625 return;
626 }
627
628 if (RISCV::GPRRegClass.contains(DstReg) &&
629 RISCV::FPR64RegClass.contains(SrcReg)) {
630 assert(STI.getXLen() == 64 && "Unexpected GPR size");
631 BuildMI(MBB, MBBI, DL, get(RISCV::FMV_X_D), DstReg)
632 .addReg(SrcReg, KillFlag);
633 return;
634 }
635
636 // VR->VR copies.
637 const TargetRegisterClass *RegClass =
638 TRI->getCommonMinimalPhysRegClass(SrcReg, DstReg);
639 if (RISCVRegisterInfo::isRVVRegClass(RegClass)) {
640 copyPhysRegVector(MBB, MBBI, DL, DstReg, SrcReg, KillSrc, RegClass);
641 return;
642 }
643
644 llvm_unreachable("Impossible reg-to-reg copy");
645}
646
649 Register SrcReg, bool IsKill, int FI,
650 const TargetRegisterClass *RC,
651 Register VReg,
652 MachineInstr::MIFlag Flags) const {
653 MachineFunction *MF = MBB.getParent();
654 MachineFrameInfo &MFI = MF->getFrameInfo();
655 Align Alignment = MFI.getObjectAlign(FI);
656
657 unsigned Opcode;
658 if (RISCV::GPRRegClass.hasSubClassEq(RC)) {
659 Opcode = RegInfo.getRegSizeInBits(RISCV::GPRRegClass) == 32 ? RISCV::SW
660 : RISCV::SD;
661 } else if (RISCV::GPRF16RegClass.hasSubClassEq(RC)) {
662 Opcode = RISCV::SH_INX;
663 } else if (RISCV::GPRF32RegClass.hasSubClassEq(RC)) {
664 Opcode = RISCV::SW_INX;
665 } else if (RISCV::GPRPairRegClass.hasSubClassEq(RC)) {
666 if (!STI.is64Bit() && STI.hasStdExtZilsd() &&
667 Alignment >= STI.getZilsdAlign()) {
668 Opcode = RISCV::SD_RV32;
669 } else {
670 Opcode = RISCV::PseudoRV32ZdinxSD;
671 }
672 } else if (RISCV::FPR16RegClass.hasSubClassEq(RC)) {
673 Opcode = RISCV::FSH;
674 } else if (RISCV::FPR32RegClass.hasSubClassEq(RC)) {
675 Opcode = RISCV::FSW;
676 } else if (RISCV::FPR64RegClass.hasSubClassEq(RC)) {
677 Opcode = RISCV::FSD;
678 } else if (RISCV::VRRegClass.hasSubClassEq(RC)) {
679 Opcode = RISCV::VS1R_V;
680 } else if (RISCV::VRM2RegClass.hasSubClassEq(RC)) {
681 Opcode = RISCV::VS2R_V;
682 } else if (RISCV::VRM4RegClass.hasSubClassEq(RC)) {
683 Opcode = RISCV::VS4R_V;
684 } else if (RISCV::VRM8RegClass.hasSubClassEq(RC)) {
685 Opcode = RISCV::VS8R_V;
686 } else if (RISCV::VRN2M1RegClass.hasSubClassEq(RC))
687 Opcode = RISCV::PseudoVSPILL2_M1;
688 else if (RISCV::VRN2M2RegClass.hasSubClassEq(RC))
689 Opcode = RISCV::PseudoVSPILL2_M2;
690 else if (RISCV::VRN2M4RegClass.hasSubClassEq(RC))
691 Opcode = RISCV::PseudoVSPILL2_M4;
692 else if (RISCV::VRN3M1RegClass.hasSubClassEq(RC))
693 Opcode = RISCV::PseudoVSPILL3_M1;
694 else if (RISCV::VRN3M2RegClass.hasSubClassEq(RC))
695 Opcode = RISCV::PseudoVSPILL3_M2;
696 else if (RISCV::VRN4M1RegClass.hasSubClassEq(RC))
697 Opcode = RISCV::PseudoVSPILL4_M1;
698 else if (RISCV::VRN4M2RegClass.hasSubClassEq(RC))
699 Opcode = RISCV::PseudoVSPILL4_M2;
700 else if (RISCV::VRN5M1RegClass.hasSubClassEq(RC))
701 Opcode = RISCV::PseudoVSPILL5_M1;
702 else if (RISCV::VRN6M1RegClass.hasSubClassEq(RC))
703 Opcode = RISCV::PseudoVSPILL6_M1;
704 else if (RISCV::VRN7M1RegClass.hasSubClassEq(RC))
705 Opcode = RISCV::PseudoVSPILL7_M1;
706 else if (RISCV::VRN8M1RegClass.hasSubClassEq(RC))
707 Opcode = RISCV::PseudoVSPILL8_M1;
708 else
709 llvm_unreachable("Can't store this register to stack slot");
710
714 TypeSize::getScalable(MFI.getObjectSize(FI)), Alignment);
715
717 BuildMI(MBB, I, DebugLoc(), get(Opcode))
718 .addReg(SrcReg, getKillRegState(IsKill))
719 .addFrameIndex(FI)
720 .addMemOperand(MMO)
721 .setMIFlag(Flags);
722 NumVRegSpilled += RegInfo.getRegSizeInBits(*RC) / RISCV::RVVBitsPerBlock;
723 } else {
726 MFI.getObjectSize(FI), Alignment);
727
728 BuildMI(MBB, I, DebugLoc(), get(Opcode))
729 .addReg(SrcReg, getKillRegState(IsKill))
730 .addFrameIndex(FI)
731 .addImm(0)
732 .addMemOperand(MMO)
733 .setMIFlag(Flags);
734 }
735}
736
739 Register DstReg, int FI,
740 const TargetRegisterClass *RC,
741 Register VReg, unsigned SubReg,
742 MachineInstr::MIFlag Flags) const {
743 MachineFunction *MF = MBB.getParent();
744 MachineFrameInfo &MFI = MF->getFrameInfo();
745 Align Alignment = MFI.getObjectAlign(FI);
746 DebugLoc DL =
747 Flags & MachineInstr::FrameDestroy ? MBB.findDebugLoc(I) : DebugLoc();
748
749 unsigned Opcode;
750 if (RISCV::GPRRegClass.hasSubClassEq(RC)) {
751 Opcode = RegInfo.getRegSizeInBits(RISCV::GPRRegClass) == 32 ? RISCV::LW
752 : RISCV::LD;
753 } else if (RISCV::GPRF16RegClass.hasSubClassEq(RC)) {
754 Opcode = RISCV::LH_INX;
755 } else if (RISCV::GPRF32RegClass.hasSubClassEq(RC)) {
756 Opcode = RISCV::LW_INX;
757 } else if (RISCV::GPRPairRegClass.hasSubClassEq(RC)) {
758 if (!STI.is64Bit() && STI.hasStdExtZilsd() &&
759 Alignment >= STI.getZilsdAlign()) {
760 Opcode = RISCV::LD_RV32;
761 } else {
762 Opcode = RISCV::PseudoRV32ZdinxLD;
763 }
764 } else if (RISCV::FPR16RegClass.hasSubClassEq(RC)) {
765 Opcode = RISCV::FLH;
766 } else if (RISCV::FPR32RegClass.hasSubClassEq(RC)) {
767 Opcode = RISCV::FLW;
768 } else if (RISCV::FPR64RegClass.hasSubClassEq(RC)) {
769 Opcode = RISCV::FLD;
770 } else if (RISCV::VRRegClass.hasSubClassEq(RC)) {
771 Opcode = RISCV::VL1RE8_V;
772 } else if (RISCV::VRM2RegClass.hasSubClassEq(RC)) {
773 Opcode = RISCV::VL2RE8_V;
774 } else if (RISCV::VRM4RegClass.hasSubClassEq(RC)) {
775 Opcode = RISCV::VL4RE8_V;
776 } else if (RISCV::VRM8RegClass.hasSubClassEq(RC)) {
777 Opcode = RISCV::VL8RE8_V;
778 } else if (RISCV::VRN2M1RegClass.hasSubClassEq(RC))
779 Opcode = RISCV::PseudoVRELOAD2_M1;
780 else if (RISCV::VRN2M2RegClass.hasSubClassEq(RC))
781 Opcode = RISCV::PseudoVRELOAD2_M2;
782 else if (RISCV::VRN2M4RegClass.hasSubClassEq(RC))
783 Opcode = RISCV::PseudoVRELOAD2_M4;
784 else if (RISCV::VRN3M1RegClass.hasSubClassEq(RC))
785 Opcode = RISCV::PseudoVRELOAD3_M1;
786 else if (RISCV::VRN3M2RegClass.hasSubClassEq(RC))
787 Opcode = RISCV::PseudoVRELOAD3_M2;
788 else if (RISCV::VRN4M1RegClass.hasSubClassEq(RC))
789 Opcode = RISCV::PseudoVRELOAD4_M1;
790 else if (RISCV::VRN4M2RegClass.hasSubClassEq(RC))
791 Opcode = RISCV::PseudoVRELOAD4_M2;
792 else if (RISCV::VRN5M1RegClass.hasSubClassEq(RC))
793 Opcode = RISCV::PseudoVRELOAD5_M1;
794 else if (RISCV::VRN6M1RegClass.hasSubClassEq(RC))
795 Opcode = RISCV::PseudoVRELOAD6_M1;
796 else if (RISCV::VRN7M1RegClass.hasSubClassEq(RC))
797 Opcode = RISCV::PseudoVRELOAD7_M1;
798 else if (RISCV::VRN8M1RegClass.hasSubClassEq(RC))
799 Opcode = RISCV::PseudoVRELOAD8_M1;
800 else
801 llvm_unreachable("Can't load this register from stack slot");
802
806 TypeSize::getScalable(MFI.getObjectSize(FI)), Alignment);
807
809 BuildMI(MBB, I, DL, get(Opcode), DstReg)
810 .addFrameIndex(FI)
811 .addMemOperand(MMO)
812 .setMIFlag(Flags);
813 NumVRegReloaded += RegInfo.getRegSizeInBits(*RC) / RISCV::RVVBitsPerBlock;
814 } else {
817 MFI.getObjectSize(FI), Alignment);
818
819 BuildMI(MBB, I, DL, get(Opcode), DstReg)
820 .addFrameIndex(FI)
821 .addImm(0)
822 .addMemOperand(MMO)
823 .setMIFlag(Flags);
824 }
825}
826std::optional<unsigned> getFoldedOpcode(MachineFunction &MF, MachineInstr &MI,
828 const RISCVSubtarget &ST) {
829
830 // The below optimizations narrow the load so they are only valid for little
831 // endian.
832 // TODO: Support big endian by adding an offset into the frame object?
833 if (MF.getDataLayout().isBigEndian())
834 return std::nullopt;
835
836 // Fold load from stack followed by sext.b/sext.h/sext.w/zext.b/zext.h/zext.w.
837 if (Ops.size() != 1 || Ops[0] != 1)
838 return std::nullopt;
839
840 switch (MI.getOpcode()) {
841 default:
842 if (RISCVInstrInfo::isSEXT_W(MI))
843 return RISCV::LW;
844 if (RISCVInstrInfo::isZEXT_W(MI))
845 return RISCV::LWU;
846 if (RISCVInstrInfo::isZEXT_B(MI))
847 return RISCV::LBU;
848 break;
849 case RISCV::SEXT_H:
850 return RISCV::LH;
851 case RISCV::SEXT_B:
852 return RISCV::LB;
853 case RISCV::ZEXT_H_RV32:
854 case RISCV::ZEXT_H_RV64:
855 return RISCV::LHU;
856 }
857
858 switch (RISCV::getRVVMCOpcode(MI.getOpcode())) {
859 default:
860 return std::nullopt;
861 case RISCV::VMV_X_S: {
862 unsigned Log2SEW =
863 MI.getOperand(RISCVII::getSEWOpNum(MI.getDesc())).getImm();
864 if (ST.getXLen() < (1U << Log2SEW))
865 return std::nullopt;
866 switch (Log2SEW) {
867 case 3:
868 return RISCV::LB;
869 case 4:
870 return RISCV::LH;
871 case 5:
872 return RISCV::LW;
873 case 6:
874 return RISCV::LD;
875 default:
876 llvm_unreachable("Unexpected SEW");
877 }
878 }
879 case RISCV::VFMV_F_S: {
880 unsigned Log2SEW =
881 MI.getOperand(RISCVII::getSEWOpNum(MI.getDesc())).getImm();
882 switch (Log2SEW) {
883 case 4:
884 return RISCV::FLH;
885 case 5:
886 return RISCV::FLW;
887 case 6:
888 return RISCV::FLD;
889 default:
890 llvm_unreachable("Unexpected SEW");
891 }
892 }
893 }
894}
895
896// This is the version used during InlineSpiller::spillAroundUses
899 MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS,
900 VirtRegMap *VRM) const {
901
902 std::optional<unsigned> LoadOpc = getFoldedOpcode(MF, MI, Ops, STI);
903 if (!LoadOpc)
904 return nullptr;
905 Register DstReg = MI.getOperand(0).getReg();
906 return BuildMI(*MI.getParent(), InsertPt, MI.getDebugLoc(), get(*LoadOpc),
907 DstReg)
908 .addFrameIndex(FrameIndex)
909 .addImm(0);
910}
911
912static unsigned getLoadPredicatedOpcode(unsigned Opcode) {
913 switch (Opcode) {
914 case RISCV::LB:
915 return RISCV::PseudoCCLB;
916 case RISCV::LBU:
917 return RISCV::PseudoCCLBU;
918 case RISCV::LH:
919 return RISCV::PseudoCCLH;
920 case RISCV::LHU:
921 return RISCV::PseudoCCLHU;
922 case RISCV::LW:
923 return RISCV::PseudoCCLW;
924 case RISCV::LWU:
925 return RISCV::PseudoCCLWU;
926 case RISCV::LD:
927 return RISCV::PseudoCCLD;
928 case RISCV::QC_E_LB:
929 return RISCV::PseudoCCQC_E_LB;
930 case RISCV::QC_E_LBU:
931 return RISCV::PseudoCCQC_E_LBU;
932 case RISCV::QC_E_LH:
933 return RISCV::PseudoCCQC_E_LH;
934 case RISCV::QC_E_LHU:
935 return RISCV::PseudoCCQC_E_LHU;
936 case RISCV::QC_E_LW:
937 return RISCV::PseudoCCQC_E_LW;
938 default:
939 return 0;
940 }
941}
942
946 LiveIntervals *LIS) const {
947 // For now, only handle RISCV::PseudoCCMOVGPR.
948 if (MI.getOpcode() != RISCV::PseudoCCMOVGPR)
949 return nullptr;
950
951 unsigned PredOpc = getLoadPredicatedOpcode(LoadMI.getOpcode());
952
953 if (!STI.hasShortForwardBranchILoad() || !PredOpc)
954 return nullptr;
955
957 if (Ops.size() != 1 || (Ops[0] != 4 && Ops[0] != 5))
958 return nullptr;
959
960 bool Invert = Ops[0] == 5;
961 const MachineOperand &FalseReg = MI.getOperand(!Invert ? 5 : 4);
962 Register DestReg = MI.getOperand(0).getReg();
963 const TargetRegisterClass *PreviousClass = MRI.getRegClass(FalseReg.getReg());
964 if (!MRI.constrainRegClass(DestReg, PreviousClass))
965 return nullptr;
966
967 // Create a new predicated version of DefMI.
968 MachineInstrBuilder NewMI = BuildMI(*MI.getParent(), InsertPt,
969 MI.getDebugLoc(), get(PredOpc), DestReg)
970 .add({MI.getOperand(1), MI.getOperand(2)});
971
972 // Add condition code, inverting if necessary.
973 auto CC = static_cast<RISCVCC::CondCode>(MI.getOperand(3).getImm());
974 if (!Invert)
976 NewMI.addImm(CC);
977
978 // Copy the false register.
979 NewMI.add(FalseReg);
980
981 // Copy all the DefMI operands.
982 const MCInstrDesc &DefDesc = LoadMI.getDesc();
983 for (unsigned i = 1, e = DefDesc.getNumOperands(); i != e; ++i)
984 NewMI.add(LoadMI.getOperand(i));
985
986 NewMI.cloneMemRefs(LoadMI);
987 return NewMI;
988}
989
992 const DebugLoc &DL, Register DstReg, uint64_t Val,
993 MachineInstr::MIFlag Flag, bool DstRenamable,
994 bool DstIsDead) const {
995 Register SrcReg = RISCV::X0;
996
997 // For RV32, allow a sign or unsigned 32 bit value.
998 if (!STI.is64Bit() && !isInt<32>(Val)) {
999 // If have a uimm32 it will still fit in a register so we can allow it.
1000 if (!isUInt<32>(Val))
1001 report_fatal_error("Should only materialize 32-bit constants for RV32");
1002
1003 // Sign extend for generateInstSeq.
1004 Val = SignExtend64<32>(Val);
1005 }
1006
1008 assert(!Seq.empty());
1009
1010 bool SrcRenamable = false;
1011 unsigned Num = 0;
1012
1013 for (const RISCVMatInt::Inst &Inst : Seq) {
1014 bool LastItem = ++Num == Seq.size();
1015 unsigned DstRegState = getDeadRegState(DstIsDead && LastItem) |
1016 getRenamableRegState(DstRenamable);
1017 unsigned SrcRegState = getKillRegState(SrcReg != RISCV::X0) |
1018 getRenamableRegState(SrcRenamable);
1019 switch (Inst.getOpndKind()) {
1020 case RISCVMatInt::Imm:
1021 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1022 .addReg(DstReg, RegState::Define | DstRegState)
1023 .addImm(Inst.getImm())
1024 .setMIFlag(Flag);
1025 break;
1026 case RISCVMatInt::RegX0:
1027 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1028 .addReg(DstReg, RegState::Define | DstRegState)
1029 .addReg(SrcReg, SrcRegState)
1030 .addReg(RISCV::X0)
1031 .setMIFlag(Flag);
1032 break;
1034 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1035 .addReg(DstReg, RegState::Define | DstRegState)
1036 .addReg(SrcReg, SrcRegState)
1037 .addReg(SrcReg, SrcRegState)
1038 .setMIFlag(Flag);
1039 break;
1041 BuildMI(MBB, MBBI, DL, get(Inst.getOpcode()))
1042 .addReg(DstReg, RegState::Define | DstRegState)
1043 .addReg(SrcReg, SrcRegState)
1044 .addImm(Inst.getImm())
1045 .setMIFlag(Flag);
1046 break;
1047 }
1048
1049 // Only the first instruction has X0 as its source.
1050 SrcReg = DstReg;
1051 SrcRenamable = DstRenamable;
1052 }
1053}
1054
1056 switch (Opc) {
1057 default:
1058 return RISCVCC::COND_INVALID;
1059 case RISCV::BEQ:
1060 case RISCV::BEQI:
1061 case RISCV::CV_BEQIMM:
1062 case RISCV::QC_BEQI:
1063 case RISCV::QC_E_BEQI:
1064 case RISCV::NDS_BBC:
1065 case RISCV::NDS_BEQC:
1066 return RISCVCC::COND_EQ;
1067 case RISCV::BNE:
1068 case RISCV::BNEI:
1069 case RISCV::QC_BNEI:
1070 case RISCV::QC_E_BNEI:
1071 case RISCV::CV_BNEIMM:
1072 case RISCV::NDS_BBS:
1073 case RISCV::NDS_BNEC:
1074 return RISCVCC::COND_NE;
1075 case RISCV::BLT:
1076 case RISCV::QC_BLTI:
1077 case RISCV::QC_E_BLTI:
1078 return RISCVCC::COND_LT;
1079 case RISCV::BGE:
1080 case RISCV::QC_BGEI:
1081 case RISCV::QC_E_BGEI:
1082 return RISCVCC::COND_GE;
1083 case RISCV::BLTU:
1084 case RISCV::QC_BLTUI:
1085 case RISCV::QC_E_BLTUI:
1086 return RISCVCC::COND_LTU;
1087 case RISCV::BGEU:
1088 case RISCV::QC_BGEUI:
1089 case RISCV::QC_E_BGEUI:
1090 return RISCVCC::COND_GEU;
1091 }
1092}
1093
1095 int64_t C1) {
1096 switch (CC) {
1097 default:
1098 llvm_unreachable("Unexpected CC");
1099 case RISCVCC::COND_EQ:
1100 return C0 == C1;
1101 case RISCVCC::COND_NE:
1102 return C0 != C1;
1103 case RISCVCC::COND_LT:
1104 return C0 < C1;
1105 case RISCVCC::COND_GE:
1106 return C0 >= C1;
1107 case RISCVCC::COND_LTU:
1108 return (uint64_t)C0 < (uint64_t)C1;
1109 case RISCVCC::COND_GEU:
1110 return (uint64_t)C0 >= (uint64_t)C1;
1111 }
1112}
1113
1114// The contents of values added to Cond are not examined outside of
1115// RISCVInstrInfo, giving us flexibility in what to push to it. For RISCV, we
1116// push BranchOpcode, Reg1, Reg2.
1119 // Block ends with fall-through condbranch.
1120 assert(LastInst.getDesc().isConditionalBranch() &&
1121 "Unknown conditional branch");
1122 Target = LastInst.getOperand(2).getMBB();
1123 Cond.push_back(MachineOperand::CreateImm(LastInst.getOpcode()));
1124 Cond.push_back(LastInst.getOperand(0));
1125 Cond.push_back(LastInst.getOperand(1));
1126}
1127
1128static unsigned getInverseXqcicmOpcode(unsigned Opcode) {
1129 switch (Opcode) {
1130 default:
1131 llvm_unreachable("Unexpected Opcode");
1132 case RISCV::QC_MVEQ:
1133 return RISCV::QC_MVNE;
1134 case RISCV::QC_MVNE:
1135 return RISCV::QC_MVEQ;
1136 case RISCV::QC_MVLT:
1137 return RISCV::QC_MVGE;
1138 case RISCV::QC_MVGE:
1139 return RISCV::QC_MVLT;
1140 case RISCV::QC_MVLTU:
1141 return RISCV::QC_MVGEU;
1142 case RISCV::QC_MVGEU:
1143 return RISCV::QC_MVLTU;
1144 case RISCV::QC_MVEQI:
1145 return RISCV::QC_MVNEI;
1146 case RISCV::QC_MVNEI:
1147 return RISCV::QC_MVEQI;
1148 case RISCV::QC_MVLTI:
1149 return RISCV::QC_MVGEI;
1150 case RISCV::QC_MVGEI:
1151 return RISCV::QC_MVLTI;
1152 case RISCV::QC_MVLTUI:
1153 return RISCV::QC_MVGEUI;
1154 case RISCV::QC_MVGEUI:
1155 return RISCV::QC_MVLTUI;
1156 }
1157}
1158
1159unsigned RISCVCC::getBrCond(RISCVCC::CondCode CC, unsigned SelectOpc) {
1160 switch (SelectOpc) {
1161 default:
1162 switch (CC) {
1163 default:
1164 llvm_unreachable("Unexpected condition code!");
1165 case RISCVCC::COND_EQ:
1166 return RISCV::BEQ;
1167 case RISCVCC::COND_NE:
1168 return RISCV::BNE;
1169 case RISCVCC::COND_LT:
1170 return RISCV::BLT;
1171 case RISCVCC::COND_GE:
1172 return RISCV::BGE;
1173 case RISCVCC::COND_LTU:
1174 return RISCV::BLTU;
1175 case RISCVCC::COND_GEU:
1176 return RISCV::BGEU;
1177 }
1178 break;
1179 case RISCV::Select_GPR_Using_CC_Imm5_Zibi:
1180 switch (CC) {
1181 default:
1182 llvm_unreachable("Unexpected condition code!");
1183 case RISCVCC::COND_EQ:
1184 return RISCV::BEQI;
1185 case RISCVCC::COND_NE:
1186 return RISCV::BNEI;
1187 }
1188 break;
1189 case RISCV::Select_GPR_Using_CC_SImm5_CV:
1190 switch (CC) {
1191 default:
1192 llvm_unreachable("Unexpected condition code!");
1193 case RISCVCC::COND_EQ:
1194 return RISCV::CV_BEQIMM;
1195 case RISCVCC::COND_NE:
1196 return RISCV::CV_BNEIMM;
1197 }
1198 break;
1199 case RISCV::Select_GPRNoX0_Using_CC_SImm5NonZero_QC:
1200 switch (CC) {
1201 default:
1202 llvm_unreachable("Unexpected condition code!");
1203 case RISCVCC::COND_EQ:
1204 return RISCV::QC_BEQI;
1205 case RISCVCC::COND_NE:
1206 return RISCV::QC_BNEI;
1207 case RISCVCC::COND_LT:
1208 return RISCV::QC_BLTI;
1209 case RISCVCC::COND_GE:
1210 return RISCV::QC_BGEI;
1211 }
1212 break;
1213 case RISCV::Select_GPRNoX0_Using_CC_UImm5NonZero_QC:
1214 switch (CC) {
1215 default:
1216 llvm_unreachable("Unexpected condition code!");
1217 case RISCVCC::COND_LTU:
1218 return RISCV::QC_BLTUI;
1219 case RISCVCC::COND_GEU:
1220 return RISCV::QC_BGEUI;
1221 }
1222 break;
1223 case RISCV::Select_GPRNoX0_Using_CC_SImm16NonZero_QC:
1224 switch (CC) {
1225 default:
1226 llvm_unreachable("Unexpected condition code!");
1227 case RISCVCC::COND_EQ:
1228 return RISCV::QC_E_BEQI;
1229 case RISCVCC::COND_NE:
1230 return RISCV::QC_E_BNEI;
1231 case RISCVCC::COND_LT:
1232 return RISCV::QC_E_BLTI;
1233 case RISCVCC::COND_GE:
1234 return RISCV::QC_E_BGEI;
1235 }
1236 break;
1237 case RISCV::Select_GPRNoX0_Using_CC_UImm16NonZero_QC:
1238 switch (CC) {
1239 default:
1240 llvm_unreachable("Unexpected condition code!");
1241 case RISCVCC::COND_LTU:
1242 return RISCV::QC_E_BLTUI;
1243 case RISCVCC::COND_GEU:
1244 return RISCV::QC_E_BGEUI;
1245 }
1246 break;
1247 case RISCV::Select_GPR_Using_CC_UImmLog2XLen_NDS:
1248 switch (CC) {
1249 default:
1250 llvm_unreachable("Unexpected condition code!");
1251 case RISCVCC::COND_EQ:
1252 return RISCV::NDS_BBC;
1253 case RISCVCC::COND_NE:
1254 return RISCV::NDS_BBS;
1255 }
1256 break;
1257 case RISCV::Select_GPR_Using_CC_UImm7_NDS:
1258 switch (CC) {
1259 default:
1260 llvm_unreachable("Unexpected condition code!");
1261 case RISCVCC::COND_EQ:
1262 return RISCV::NDS_BEQC;
1263 case RISCVCC::COND_NE:
1264 return RISCV::NDS_BNEC;
1265 }
1266 break;
1267 }
1268}
1269
1271 switch (CC) {
1272 default:
1273 llvm_unreachable("Unrecognized conditional branch");
1274 case RISCVCC::COND_EQ:
1275 return RISCVCC::COND_NE;
1276 case RISCVCC::COND_NE:
1277 return RISCVCC::COND_EQ;
1278 case RISCVCC::COND_LT:
1279 return RISCVCC::COND_GE;
1280 case RISCVCC::COND_GE:
1281 return RISCVCC::COND_LT;
1282 case RISCVCC::COND_LTU:
1283 return RISCVCC::COND_GEU;
1284 case RISCVCC::COND_GEU:
1285 return RISCVCC::COND_LTU;
1286 }
1287}
1288
1291 MachineBasicBlock *&FBB,
1293 bool AllowModify) const {
1294 TBB = FBB = nullptr;
1295 Cond.clear();
1296
1297 // If the block has no terminators, it just falls into the block after it.
1298 MachineBasicBlock::iterator I = MBB.getLastNonDebugInstr();
1299 if (I == MBB.end() || !isUnpredicatedTerminator(*I))
1300 return false;
1301
1302 // Count the number of terminators and find the first unconditional or
1303 // indirect branch.
1304 MachineBasicBlock::iterator FirstUncondOrIndirectBr = MBB.end();
1305 int NumTerminators = 0;
1306 for (auto J = I.getReverse(); J != MBB.rend() && isUnpredicatedTerminator(*J);
1307 J++) {
1308 NumTerminators++;
1309 if (J->getDesc().isUnconditionalBranch() ||
1310 J->getDesc().isIndirectBranch()) {
1311 FirstUncondOrIndirectBr = J.getReverse();
1312 }
1313 }
1314
1315 // If AllowModify is true, we can erase any terminators after
1316 // FirstUncondOrIndirectBR.
1317 if (AllowModify && FirstUncondOrIndirectBr != MBB.end()) {
1318 while (std::next(FirstUncondOrIndirectBr) != MBB.end()) {
1319 std::next(FirstUncondOrIndirectBr)->eraseFromParent();
1320 NumTerminators--;
1321 }
1322 I = FirstUncondOrIndirectBr;
1323 }
1324
1325 // We can't handle blocks that end in an indirect branch.
1326 if (I->getDesc().isIndirectBranch())
1327 return true;
1328
1329 // We can't handle Generic branch opcodes from Global ISel.
1330 if (I->isPreISelOpcode())
1331 return true;
1332
1333 // We can't handle blocks with more than 2 terminators.
1334 if (NumTerminators > 2)
1335 return true;
1336
1337 // Handle a single unconditional branch.
1338 if (NumTerminators == 1 && I->getDesc().isUnconditionalBranch()) {
1340 return false;
1341 }
1342
1343 // Handle a single conditional branch.
1344 if (NumTerminators == 1 && I->getDesc().isConditionalBranch()) {
1346 return false;
1347 }
1348
1349 // Handle a conditional branch followed by an unconditional branch.
1350 if (NumTerminators == 2 && std::prev(I)->getDesc().isConditionalBranch() &&
1351 I->getDesc().isUnconditionalBranch()) {
1352 parseCondBranch(*std::prev(I), TBB, Cond);
1353 FBB = getBranchDestBlock(*I);
1354 return false;
1355 }
1356
1357 // Otherwise, we can't handle this.
1358 return true;
1359}
1360
1362 int *BytesRemoved) const {
1363 if (BytesRemoved)
1364 *BytesRemoved = 0;
1365 MachineBasicBlock::iterator I = MBB.getLastNonDebugInstr();
1366 if (I == MBB.end())
1367 return 0;
1368
1369 if (!I->getDesc().isUnconditionalBranch() &&
1370 !I->getDesc().isConditionalBranch())
1371 return 0;
1372
1373 // Remove the branch.
1374 if (BytesRemoved)
1375 *BytesRemoved += getInstSizeInBytes(*I);
1376 I->eraseFromParent();
1377
1378 I = MBB.end();
1379
1380 if (I == MBB.begin())
1381 return 1;
1382 --I;
1383 if (!I->getDesc().isConditionalBranch())
1384 return 1;
1385
1386 // Remove the branch.
1387 if (BytesRemoved)
1388 *BytesRemoved += getInstSizeInBytes(*I);
1389 I->eraseFromParent();
1390 return 2;
1391}
1392
1393// Inserts a branch into the end of the specific MachineBasicBlock, returning
1394// the number of instructions inserted.
1397 ArrayRef<MachineOperand> Cond, const DebugLoc &DL, int *BytesAdded) const {
1398 if (BytesAdded)
1399 *BytesAdded = 0;
1400
1401 // Shouldn't be a fall through.
1402 assert(TBB && "insertBranch must not be told to insert a fallthrough");
1403 assert((Cond.size() == 3 || Cond.size() == 0) &&
1404 "RISC-V branch conditions have two components!");
1405
1406 // Unconditional branch.
1407 if (Cond.empty()) {
1408 MachineInstr &MI = *BuildMI(&MBB, DL, get(RISCV::PseudoBR)).addMBB(TBB);
1409 if (BytesAdded)
1410 *BytesAdded += getInstSizeInBytes(MI);
1411 return 1;
1412 }
1413
1414 // Either a one or two-way conditional branch.
1415 MachineInstr &CondMI = *BuildMI(&MBB, DL, get(Cond[0].getImm()))
1416 .add(Cond[1])
1417 .add(Cond[2])
1418 .addMBB(TBB);
1419 if (BytesAdded)
1420 *BytesAdded += getInstSizeInBytes(CondMI);
1421
1422 // One-way conditional branch.
1423 if (!FBB)
1424 return 1;
1425
1426 // Two-way conditional branch.
1427 MachineInstr &MI = *BuildMI(&MBB, DL, get(RISCV::PseudoBR)).addMBB(FBB);
1428 if (BytesAdded)
1429 *BytesAdded += getInstSizeInBytes(MI);
1430 return 2;
1431}
1432
1434 MachineBasicBlock &DestBB,
1435 MachineBasicBlock &RestoreBB,
1436 const DebugLoc &DL, int64_t BrOffset,
1437 RegScavenger *RS) const {
1438 assert(RS && "RegScavenger required for long branching");
1439 assert(MBB.empty() &&
1440 "new block should be inserted for expanding unconditional branch");
1441 assert(MBB.pred_size() == 1);
1442 assert(RestoreBB.empty() &&
1443 "restore block should be inserted for restoring clobbered registers");
1444
1445 MachineFunction *MF = MBB.getParent();
1449
1450 if (!isInt<32>(BrOffset))
1452 "Branch offsets outside of the signed 32-bit range not supported");
1453
1454 // FIXME: A virtual register must be used initially, as the register
1455 // scavenger won't work with empty blocks (SIInstrInfo::insertIndirectBranch
1456 // uses the same workaround).
1457 Register ScratchReg = MRI.createVirtualRegister(&RISCV::GPRJALRRegClass);
1458 auto II = MBB.end();
1459 // We may also update the jump target to RestoreBB later.
1460 MachineInstr &MI = *BuildMI(MBB, II, DL, get(RISCV::PseudoJump))
1461 .addReg(ScratchReg, RegState::Define | RegState::Dead)
1462 .addMBB(&DestBB, RISCVII::MO_CALL);
1463
1464 RS->enterBasicBlockEnd(MBB);
1465 const TargetRegisterClass *RC = &RISCV::GPRRegClass;
1466 if (STI.hasStdExtZicfilp())
1467 RC = &RISCV::GPRX7RegClass;
1468 Register TmpGPR =
1469 RS->scavengeRegisterBackwards(*RC, MI.getIterator(),
1470 /*RestoreAfter=*/false, /*SpAdj=*/0,
1471 /*AllowSpill=*/false);
1472 if (TmpGPR.isValid())
1473 RS->setRegUsed(TmpGPR);
1474 else {
1475 // The case when there is no scavenged register needs special handling.
1476
1477 // Pick s11(or s1 for rve) because it doesn't make a difference.
1478 TmpGPR = STI.hasStdExtE() ? RISCV::X9 : RISCV::X27;
1479 // Force t2 if Zicfilp is on
1480 if (STI.hasStdExtZicfilp())
1481 TmpGPR = RISCV::X7;
1482
1483 int FrameIndex = RVFI->getBranchRelaxationScratchFrameIndex();
1484 if (FrameIndex == -1)
1485 report_fatal_error("underestimated function size");
1486
1487 storeRegToStackSlot(MBB, MI, TmpGPR, /*IsKill=*/true, FrameIndex,
1488 &RISCV::GPRRegClass, Register());
1489 TRI->eliminateFrameIndex(std::prev(MI.getIterator()),
1490 /*SpAdj=*/0, /*FIOperandNum=*/1);
1491
1492 MI.getOperand(1).setMBB(&RestoreBB);
1493
1494 loadRegFromStackSlot(RestoreBB, RestoreBB.end(), TmpGPR, FrameIndex,
1495 &RISCV::GPRRegClass, Register());
1496 TRI->eliminateFrameIndex(RestoreBB.back(),
1497 /*SpAdj=*/0, /*FIOperandNum=*/1);
1498 }
1499
1500 MRI.replaceRegWith(ScratchReg, TmpGPR);
1501 MRI.clearVirtRegs();
1502}
1503
1506 assert((Cond.size() == 3) && "Invalid branch condition!");
1507 switch (Cond[0].getImm()) {
1508 default:
1509 llvm_unreachable("Unknown conditional branch!");
1510 case RISCV::BEQ:
1511 Cond[0].setImm(RISCV::BNE);
1512 break;
1513 case RISCV::BEQI:
1514 Cond[0].setImm(RISCV::BNEI);
1515 break;
1516 case RISCV::BNE:
1517 Cond[0].setImm(RISCV::BEQ);
1518 break;
1519 case RISCV::BNEI:
1520 Cond[0].setImm(RISCV::BEQI);
1521 break;
1522 case RISCV::BLT:
1523 Cond[0].setImm(RISCV::BGE);
1524 break;
1525 case RISCV::BGE:
1526 Cond[0].setImm(RISCV::BLT);
1527 break;
1528 case RISCV::BLTU:
1529 Cond[0].setImm(RISCV::BGEU);
1530 break;
1531 case RISCV::BGEU:
1532 Cond[0].setImm(RISCV::BLTU);
1533 break;
1534 case RISCV::CV_BEQIMM:
1535 Cond[0].setImm(RISCV::CV_BNEIMM);
1536 break;
1537 case RISCV::CV_BNEIMM:
1538 Cond[0].setImm(RISCV::CV_BEQIMM);
1539 break;
1540 case RISCV::QC_BEQI:
1541 Cond[0].setImm(RISCV::QC_BNEI);
1542 break;
1543 case RISCV::QC_BNEI:
1544 Cond[0].setImm(RISCV::QC_BEQI);
1545 break;
1546 case RISCV::QC_BGEI:
1547 Cond[0].setImm(RISCV::QC_BLTI);
1548 break;
1549 case RISCV::QC_BLTI:
1550 Cond[0].setImm(RISCV::QC_BGEI);
1551 break;
1552 case RISCV::QC_BGEUI:
1553 Cond[0].setImm(RISCV::QC_BLTUI);
1554 break;
1555 case RISCV::QC_BLTUI:
1556 Cond[0].setImm(RISCV::QC_BGEUI);
1557 break;
1558 case RISCV::QC_E_BEQI:
1559 Cond[0].setImm(RISCV::QC_E_BNEI);
1560 break;
1561 case RISCV::QC_E_BNEI:
1562 Cond[0].setImm(RISCV::QC_E_BEQI);
1563 break;
1564 case RISCV::QC_E_BGEI:
1565 Cond[0].setImm(RISCV::QC_E_BLTI);
1566 break;
1567 case RISCV::QC_E_BLTI:
1568 Cond[0].setImm(RISCV::QC_E_BGEI);
1569 break;
1570 case RISCV::QC_E_BGEUI:
1571 Cond[0].setImm(RISCV::QC_E_BLTUI);
1572 break;
1573 case RISCV::QC_E_BLTUI:
1574 Cond[0].setImm(RISCV::QC_E_BGEUI);
1575 break;
1576 case RISCV::NDS_BBC:
1577 Cond[0].setImm(RISCV::NDS_BBS);
1578 break;
1579 case RISCV::NDS_BBS:
1580 Cond[0].setImm(RISCV::NDS_BBC);
1581 break;
1582 case RISCV::NDS_BEQC:
1583 Cond[0].setImm(RISCV::NDS_BNEC);
1584 break;
1585 case RISCV::NDS_BNEC:
1586 Cond[0].setImm(RISCV::NDS_BEQC);
1587 break;
1588 }
1589
1590 return false;
1591}
1592
1593// Return true if the instruction is a load immediate instruction (i.e.
1594// ADDI x0, imm).
1595static bool isLoadImm(const MachineInstr *MI, int64_t &Imm) {
1596 if (MI->getOpcode() == RISCV::ADDI && MI->getOperand(1).isReg() &&
1597 MI->getOperand(1).getReg() == RISCV::X0) {
1598 Imm = MI->getOperand(2).getImm();
1599 return true;
1600 }
1601 return false;
1602}
1603
1605 const MachineOperand &Op, int64_t &Imm) {
1606 // Either a load from immediate instruction or X0.
1607 if (!Op.isReg())
1608 return false;
1609
1610 Register Reg = Op.getReg();
1611 if (Reg == RISCV::X0) {
1612 Imm = 0;
1613 return true;
1614 }
1615 return Reg.isVirtual() && isLoadImm(MRI.getVRegDef(Reg), Imm);
1616}
1617
1619 bool IsSigned = false;
1620 bool IsEquality = false;
1621 switch (MI.getOpcode()) {
1622 default:
1623 return false;
1624 case RISCV::BEQ:
1625 case RISCV::BNE:
1626 IsEquality = true;
1627 break;
1628 case RISCV::BGE:
1629 case RISCV::BLT:
1630 IsSigned = true;
1631 break;
1632 case RISCV::BGEU:
1633 case RISCV::BLTU:
1634 break;
1635 }
1636
1637 MachineBasicBlock *MBB = MI.getParent();
1638 MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo();
1639
1640 const MachineOperand &LHS = MI.getOperand(0);
1641 const MachineOperand &RHS = MI.getOperand(1);
1642 MachineBasicBlock *TBB = MI.getOperand(2).getMBB();
1643
1644 RISCVCC::CondCode CC = getCondFromBranchOpc(MI.getOpcode());
1646
1647 // Canonicalize conditional branches which can be constant folded into
1648 // beqz or bnez. We can't modify the CFG here.
1649 int64_t C0, C1;
1650 if (isFromLoadImm(MRI, LHS, C0) && isFromLoadImm(MRI, RHS, C1)) {
1651 unsigned NewOpc = evaluateCondBranch(CC, C0, C1) ? RISCV::BEQ : RISCV::BNE;
1652 // Build the new branch and remove the old one.
1653 BuildMI(*MBB, MI, MI.getDebugLoc(), get(NewOpc))
1654 .addReg(RISCV::X0)
1655 .addReg(RISCV::X0)
1656 .addMBB(TBB);
1657 MI.eraseFromParent();
1658 return true;
1659 }
1660
1661 if (IsEquality)
1662 return false;
1663
1664 // For two constants C0 and C1 from
1665 // ```
1666 // li Y, C0
1667 // li Z, C1
1668 // ```
1669 // 1. if C1 = C0 + 1
1670 // we can turn:
1671 // (a) blt Y, X -> bge X, Z
1672 // (b) bge Y, X -> blt X, Z
1673 //
1674 // 2. if C1 = C0 - 1
1675 // we can turn:
1676 // (a) blt X, Y -> bge Z, X
1677 // (b) bge X, Y -> blt Z, X
1678 //
1679 // To make sure this optimization is really beneficial, we only
1680 // optimize for cases where Y had only one use (i.e. only used by the branch).
1681 // Try to find the register for constant Z; return
1682 // invalid register otherwise.
1683 auto searchConst = [&](int64_t C1) -> Register {
1685 auto DefC1 = std::find_if(++II, E, [&](const MachineInstr &I) -> bool {
1686 int64_t Imm;
1687 return isLoadImm(&I, Imm) && Imm == C1 &&
1688 I.getOperand(0).getReg().isVirtual();
1689 });
1690 if (DefC1 != E)
1691 return DefC1->getOperand(0).getReg();
1692
1693 return Register();
1694 };
1695
1696 unsigned NewOpc = RISCVCC::getBrCond(getInverseBranchCondition(CC));
1697
1698 // Might be case 1.
1699 // Don't change 0 to 1 since we can use x0.
1700 // For unsigned cases changing -1U to 0 would be incorrect.
1701 // The incorrect case for signed would be INT_MAX, but isFromLoadImm can't
1702 // return that.
1703 if (isFromLoadImm(MRI, LHS, C0) && C0 != 0 && LHS.getReg().isVirtual() &&
1704 MRI.hasOneUse(LHS.getReg()) && (IsSigned || C0 != -1)) {
1705 assert(isInt<12>(C0) && "Unexpected immediate");
1706 if (Register RegZ = searchConst(C0 + 1)) {
1707 BuildMI(*MBB, MI, MI.getDebugLoc(), get(NewOpc))
1708 .add(RHS)
1709 .addReg(RegZ)
1710 .addMBB(TBB);
1711 // We might extend the live range of Z, clear its kill flag to
1712 // account for this.
1713 MRI.clearKillFlags(RegZ);
1714 MI.eraseFromParent();
1715 return true;
1716 }
1717 }
1718
1719 // Might be case 2.
1720 // For signed cases we don't want to change 0 since we can use x0.
1721 // For unsigned cases changing 0 to -1U would be incorrect.
1722 // The incorrect case for signed would be INT_MIN, but isFromLoadImm can't
1723 // return that.
1724 if (isFromLoadImm(MRI, RHS, C0) && C0 != 0 && RHS.getReg().isVirtual() &&
1725 MRI.hasOneUse(RHS.getReg())) {
1726 assert(isInt<12>(C0) && "Unexpected immediate");
1727 if (Register RegZ = searchConst(C0 - 1)) {
1728 BuildMI(*MBB, MI, MI.getDebugLoc(), get(NewOpc))
1729 .addReg(RegZ)
1730 .add(LHS)
1731 .addMBB(TBB);
1732 // We might extend the live range of Z, clear its kill flag to
1733 // account for this.
1734 MRI.clearKillFlags(RegZ);
1735 MI.eraseFromParent();
1736 return true;
1737 }
1738 }
1739
1740 return false;
1741}
1742
1745 assert(MI.getDesc().isBranch() && "Unexpected opcode!");
1746 // The branch target is always the last operand.
1747 int NumOp = MI.getNumExplicitOperands();
1748 return MI.getOperand(NumOp - 1).getMBB();
1749}
1750
1752 int64_t BrOffset) const {
1753 unsigned XLen = STI.getXLen();
1754 // Ideally we could determine the supported branch offset from the
1755 // RISCVII::FormMask, but this can't be used for Pseudo instructions like
1756 // PseudoBR.
1757 switch (BranchOp) {
1758 default:
1759 llvm_unreachable("Unexpected opcode!");
1760 case RISCV::NDS_BBC:
1761 case RISCV::NDS_BBS:
1762 case RISCV::NDS_BEQC:
1763 case RISCV::NDS_BNEC:
1764 return isInt<11>(BrOffset);
1765 case RISCV::BEQ:
1766 case RISCV::BNE:
1767 case RISCV::BLT:
1768 case RISCV::BGE:
1769 case RISCV::BLTU:
1770 case RISCV::BGEU:
1771 case RISCV::BEQI:
1772 case RISCV::BNEI:
1773 case RISCV::CV_BEQIMM:
1774 case RISCV::CV_BNEIMM:
1775 case RISCV::QC_BEQI:
1776 case RISCV::QC_BNEI:
1777 case RISCV::QC_BGEI:
1778 case RISCV::QC_BLTI:
1779 case RISCV::QC_BLTUI:
1780 case RISCV::QC_BGEUI:
1781 case RISCV::QC_E_BEQI:
1782 case RISCV::QC_E_BNEI:
1783 case RISCV::QC_E_BGEI:
1784 case RISCV::QC_E_BLTI:
1785 case RISCV::QC_E_BLTUI:
1786 case RISCV::QC_E_BGEUI:
1787 return isInt<13>(BrOffset);
1788 case RISCV::JAL:
1789 case RISCV::PseudoBR:
1790 return isInt<21>(BrOffset);
1791 case RISCV::PseudoJump:
1792 return isInt<32>(SignExtend64(BrOffset + 0x800, XLen));
1793 }
1794}
1795
1796// If the operation has a predicated pseudo instruction, return the pseudo
1797// instruction opcode. Otherwise, return RISCV::INSTRUCTION_LIST_END.
1798// TODO: Support more operations.
1799unsigned getPredicatedOpcode(unsigned Opcode) {
1800 // clang-format off
1801 switch (Opcode) {
1802 case RISCV::ADD: return RISCV::PseudoCCADD;
1803 case RISCV::SUB: return RISCV::PseudoCCSUB;
1804 case RISCV::SLL: return RISCV::PseudoCCSLL;
1805 case RISCV::SRL: return RISCV::PseudoCCSRL;
1806 case RISCV::SRA: return RISCV::PseudoCCSRA;
1807 case RISCV::AND: return RISCV::PseudoCCAND;
1808 case RISCV::OR: return RISCV::PseudoCCOR;
1809 case RISCV::XOR: return RISCV::PseudoCCXOR;
1810 case RISCV::MAX: return RISCV::PseudoCCMAX;
1811 case RISCV::MAXU: return RISCV::PseudoCCMAXU;
1812 case RISCV::MIN: return RISCV::PseudoCCMIN;
1813 case RISCV::MINU: return RISCV::PseudoCCMINU;
1814 case RISCV::MUL: return RISCV::PseudoCCMUL;
1815 case RISCV::LUI: return RISCV::PseudoCCLUI;
1816 case RISCV::QC_LI: return RISCV::PseudoCCQC_LI;
1817 case RISCV::QC_E_LI: return RISCV::PseudoCCQC_E_LI;
1818
1819 case RISCV::ADDI: return RISCV::PseudoCCADDI;
1820 case RISCV::SLLI: return RISCV::PseudoCCSLLI;
1821 case RISCV::SRLI: return RISCV::PseudoCCSRLI;
1822 case RISCV::SRAI: return RISCV::PseudoCCSRAI;
1823 case RISCV::ANDI: return RISCV::PseudoCCANDI;
1824 case RISCV::ORI: return RISCV::PseudoCCORI;
1825 case RISCV::XORI: return RISCV::PseudoCCXORI;
1826
1827 case RISCV::ADDW: return RISCV::PseudoCCADDW;
1828 case RISCV::SUBW: return RISCV::PseudoCCSUBW;
1829 case RISCV::SLLW: return RISCV::PseudoCCSLLW;
1830 case RISCV::SRLW: return RISCV::PseudoCCSRLW;
1831 case RISCV::SRAW: return RISCV::PseudoCCSRAW;
1832
1833 case RISCV::ADDIW: return RISCV::PseudoCCADDIW;
1834 case RISCV::SLLIW: return RISCV::PseudoCCSLLIW;
1835 case RISCV::SRLIW: return RISCV::PseudoCCSRLIW;
1836 case RISCV::SRAIW: return RISCV::PseudoCCSRAIW;
1837
1838 case RISCV::ANDN: return RISCV::PseudoCCANDN;
1839 case RISCV::ORN: return RISCV::PseudoCCORN;
1840 case RISCV::XNOR: return RISCV::PseudoCCXNOR;
1841
1842 case RISCV::NDS_BFOS: return RISCV::PseudoCCNDS_BFOS;
1843 case RISCV::NDS_BFOZ: return RISCV::PseudoCCNDS_BFOZ;
1844 }
1845 // clang-format on
1846
1847 return RISCV::INSTRUCTION_LIST_END;
1848}
1849
1850/// Identify instructions that can be folded into a CCMOV instruction, and
1851/// return the defining instruction.
1853 const MachineRegisterInfo &MRI,
1854 const TargetInstrInfo *TII,
1855 const RISCVSubtarget &STI) {
1856 if (!Reg.isVirtual())
1857 return nullptr;
1858 if (!MRI.hasOneNonDBGUse(Reg))
1859 return nullptr;
1860 MachineInstr *MI = MRI.getVRegDef(Reg);
1861 if (!MI)
1862 return nullptr;
1863
1864 if (!STI.hasShortForwardBranchIMinMax() &&
1865 (MI->getOpcode() == RISCV::MAX || MI->getOpcode() == RISCV::MIN ||
1866 MI->getOpcode() == RISCV::MINU || MI->getOpcode() == RISCV::MAXU))
1867 return nullptr;
1868
1869 if (!STI.hasShortForwardBranchIMul() && MI->getOpcode() == RISCV::MUL)
1870 return nullptr;
1871
1872 // Check if MI can be predicated and folded into the CCMOV.
1873 if (getPredicatedOpcode(MI->getOpcode()) == RISCV::INSTRUCTION_LIST_END)
1874 return nullptr;
1875 // Don't predicate li idiom.
1876 if (MI->getOpcode() == RISCV::ADDI && MI->getOperand(1).isReg() &&
1877 MI->getOperand(1).getReg() == RISCV::X0)
1878 return nullptr;
1879 // Check if MI has any other defs or physreg uses.
1880 for (const MachineOperand &MO : llvm::drop_begin(MI->operands())) {
1881 // Reject frame index operands, PEI can't handle the predicated pseudos.
1882 if (MO.isFI() || MO.isCPI() || MO.isJTI())
1883 return nullptr;
1884 if (!MO.isReg())
1885 continue;
1886 // MI can't have any tied operands, that would conflict with predication.
1887 if (MO.isTied())
1888 return nullptr;
1889 if (MO.isDef())
1890 return nullptr;
1891 // Allow constant physregs.
1892 if (MO.getReg().isPhysical() && !MRI.isConstantPhysReg(MO.getReg()))
1893 return nullptr;
1894 }
1895 bool DontMoveAcrossStores = true;
1896 if (!MI->isSafeToMove(DontMoveAcrossStores))
1897 return nullptr;
1898 return MI;
1899}
1900
1903 unsigned &TrueOp, unsigned &FalseOp,
1904 bool &Optimizable) const {
1905 assert(MI.getOpcode() == RISCV::PseudoCCMOVGPR &&
1906 "Unknown select instruction");
1907 // CCMOV operands:
1908 // 0: Def.
1909 // 1: LHS of compare.
1910 // 2: RHS of compare.
1911 // 3: Condition code.
1912 // 4: False use.
1913 // 5: True use.
1914 TrueOp = 5;
1915 FalseOp = 4;
1916 Cond.push_back(MI.getOperand(1));
1917 Cond.push_back(MI.getOperand(2));
1918 Cond.push_back(MI.getOperand(3));
1919 // We can only fold when we support short forward branch opt.
1920 Optimizable = STI.hasShortForwardBranchIALU();
1921 return false;
1922}
1923
1927 bool PreferFalse) const {
1928 assert(MI.getOpcode() == RISCV::PseudoCCMOVGPR &&
1929 "Unknown select instruction");
1930 if (!STI.hasShortForwardBranchIALU())
1931 return nullptr;
1932
1933 MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo();
1935 canFoldAsPredicatedOp(MI.getOperand(5).getReg(), MRI, this, STI);
1936 bool Invert = !DefMI;
1937 if (!DefMI)
1938 DefMI = canFoldAsPredicatedOp(MI.getOperand(4).getReg(), MRI, this, STI);
1939 if (!DefMI)
1940 return nullptr;
1941
1942 // Find new register class to use.
1943 MachineOperand FalseReg = MI.getOperand(Invert ? 5 : 4);
1944 Register DestReg = MI.getOperand(0).getReg();
1945 const TargetRegisterClass *PreviousClass = MRI.getRegClass(FalseReg.getReg());
1946 if (!MRI.constrainRegClass(DestReg, PreviousClass))
1947 return nullptr;
1948
1949 unsigned PredOpc = getPredicatedOpcode(DefMI->getOpcode());
1950 assert(PredOpc != RISCV::INSTRUCTION_LIST_END && "Unexpected opcode!");
1951
1952 // Create a new predicated version of DefMI.
1953 MachineInstrBuilder NewMI =
1954 BuildMI(*MI.getParent(), MI, MI.getDebugLoc(), get(PredOpc), DestReg);
1955
1956 // Copy the condition portion.
1957 NewMI.add(MI.getOperand(1));
1958 NewMI.add(MI.getOperand(2));
1959
1960 // Add condition code, inverting if necessary.
1961 auto CC = static_cast<RISCVCC::CondCode>(MI.getOperand(3).getImm());
1962 if (Invert)
1964 NewMI.addImm(CC);
1965
1966 // Copy the false register.
1967 NewMI.add(FalseReg);
1968
1969 // Copy all the DefMI operands.
1970 const MCInstrDesc &DefDesc = DefMI->getDesc();
1971 for (unsigned i = 1, e = DefDesc.getNumOperands(); i != e; ++i)
1972 NewMI.add(DefMI->getOperand(i));
1973
1974 // Update SeenMIs set: register newly created MI and erase removed DefMI.
1975 SeenMIs.insert(NewMI);
1976 SeenMIs.erase(DefMI);
1977
1978 // If MI is inside a loop, and DefMI is outside the loop, then kill flags on
1979 // DefMI would be invalid when transferred inside the loop. Checking for a
1980 // loop is expensive, but at least remove kill flags if they are in different
1981 // BBs.
1982 if (DefMI->getParent() != MI.getParent())
1983 NewMI->clearKillInfo();
1984
1985 // The caller will erase MI, but not DefMI.
1986 DefMI->eraseFromParent();
1987 return NewMI;
1988}
1989
1991 if (MI.isMetaInstruction())
1992 return 0;
1993
1994 unsigned Opcode = MI.getOpcode();
1995
1996 if (Opcode == TargetOpcode::INLINEASM ||
1997 Opcode == TargetOpcode::INLINEASM_BR) {
1998 const MachineFunction &MF = *MI.getParent()->getParent();
1999 return getInlineAsmLength(MI.getOperand(0).getSymbolName(),
2000 *MF.getTarget().getMCAsmInfo());
2001 }
2002
2003 if (!MI.memoperands_empty()) {
2004 MachineMemOperand *MMO = *(MI.memoperands_begin());
2005 if (STI.hasStdExtZihintntl() && MMO->isNonTemporal()) {
2006 if (STI.hasStdExtZca()) {
2007 if (isCompressibleInst(MI, STI))
2008 return 4; // c.ntl.all + c.load/c.store
2009 return 6; // c.ntl.all + load/store
2010 }
2011 return 8; // ntl.all + load/store
2012 }
2013 }
2014
2015 if (Opcode == TargetOpcode::BUNDLE)
2016 return getInstBundleLength(MI);
2017
2018 if (MI.getParent() && MI.getParent()->getParent()) {
2019 if (isCompressibleInst(MI, STI))
2020 return 2;
2021 }
2022
2023 switch (Opcode) {
2024 case RISCV::PseudoMV_FPR16INX:
2025 case RISCV::PseudoMV_FPR32INX:
2026 // MV is always compressible to either c.mv or c.li rd, 0.
2027 return STI.hasStdExtZca() ? 2 : 4;
2028 case TargetOpcode::STACKMAP:
2029 // The upper bound for a stackmap intrinsic is the full length of its shadow
2031 case TargetOpcode::PATCHPOINT:
2032 // The size of the patchpoint intrinsic is the number of bytes requested
2034 case TargetOpcode::STATEPOINT: {
2035 // The size of the statepoint intrinsic is the number of bytes requested
2036 unsigned NumBytes = StatepointOpers(&MI).getNumPatchBytes();
2037 // No patch bytes means at most a PseudoCall is emitted
2038 return std::max(NumBytes, 8U);
2039 }
2040 case TargetOpcode::PATCHABLE_FUNCTION_ENTER:
2041 case TargetOpcode::PATCHABLE_FUNCTION_EXIT:
2042 case TargetOpcode::PATCHABLE_TAIL_CALL: {
2043 const MachineFunction &MF = *MI.getParent()->getParent();
2044 const Function &F = MF.getFunction();
2045 if (Opcode == TargetOpcode::PATCHABLE_FUNCTION_ENTER &&
2046 F.hasFnAttribute("patchable-function-entry")) {
2047 unsigned Num;
2048 if (F.getFnAttribute("patchable-function-entry")
2049 .getValueAsString()
2050 .getAsInteger(10, Num))
2051 return get(Opcode).getSize();
2052
2053 // Number of C.NOP or NOP
2054 return (STI.hasStdExtZca() ? 2 : 4) * Num;
2055 }
2056 // XRay uses C.JAL + 21 or 33 C.NOP for each sled in RV32 and RV64,
2057 // respectively.
2058 return STI.is64Bit() ? 68 : 44;
2059 }
2060 default:
2061 return get(Opcode).getSize();
2062 }
2063}
2064
2065unsigned RISCVInstrInfo::getInstBundleLength(const MachineInstr &MI) const {
2066 unsigned Size = 0;
2068 MachineBasicBlock::const_instr_iterator E = MI.getParent()->instr_end();
2069 while (++I != E && I->isInsideBundle()) {
2070 assert(!I->isBundle() && "No nested bundle!");
2072 }
2073 return Size;
2074}
2075
2077 const unsigned Opcode = MI.getOpcode();
2078 switch (Opcode) {
2079 default:
2080 break;
2081 case RISCV::FSGNJ_D:
2082 case RISCV::FSGNJ_S:
2083 case RISCV::FSGNJ_H:
2084 case RISCV::FSGNJ_D_INX:
2085 case RISCV::FSGNJ_D_IN32X:
2086 case RISCV::FSGNJ_S_INX:
2087 case RISCV::FSGNJ_H_INX:
2088 // The canonical floating-point move is fsgnj rd, rs, rs.
2089 return MI.getOperand(1).isReg() && MI.getOperand(2).isReg() &&
2090 MI.getOperand(1).getReg() == MI.getOperand(2).getReg();
2091 case RISCV::ADDI:
2092 case RISCV::ORI:
2093 case RISCV::XORI:
2094 return (MI.getOperand(1).isReg() &&
2095 MI.getOperand(1).getReg() == RISCV::X0) ||
2096 (MI.getOperand(2).isImm() && MI.getOperand(2).getImm() == 0);
2097 }
2098 return MI.isAsCheapAsAMove();
2099}
2100
2101std::optional<DestSourcePair>
2103 if (MI.isMoveReg())
2104 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2105 switch (MI.getOpcode()) {
2106 default:
2107 break;
2108 case RISCV::ADD:
2109 case RISCV::OR:
2110 case RISCV::XOR:
2111 if (MI.getOperand(1).isReg() && MI.getOperand(1).getReg() == RISCV::X0 &&
2112 MI.getOperand(2).isReg())
2113 return DestSourcePair{MI.getOperand(0), MI.getOperand(2)};
2114 if (MI.getOperand(2).isReg() && MI.getOperand(2).getReg() == RISCV::X0 &&
2115 MI.getOperand(1).isReg())
2116 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2117 break;
2118 case RISCV::ADDI:
2119 // Operand 1 can be a frameindex but callers expect registers
2120 if (MI.getOperand(1).isReg() && MI.getOperand(2).isImm() &&
2121 MI.getOperand(2).getImm() == 0)
2122 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2123 break;
2124 case RISCV::SUB:
2125 if (MI.getOperand(2).isReg() && MI.getOperand(2).getReg() == RISCV::X0 &&
2126 MI.getOperand(1).isReg())
2127 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2128 break;
2129 case RISCV::SH1ADD:
2130 case RISCV::SH1ADD_UW:
2131 case RISCV::SH2ADD:
2132 case RISCV::SH2ADD_UW:
2133 case RISCV::SH3ADD:
2134 case RISCV::SH3ADD_UW:
2135 if (MI.getOperand(1).isReg() && MI.getOperand(1).getReg() == RISCV::X0 &&
2136 MI.getOperand(2).isReg())
2137 return DestSourcePair{MI.getOperand(0), MI.getOperand(2)};
2138 break;
2139 case RISCV::FSGNJ_D:
2140 case RISCV::FSGNJ_S:
2141 case RISCV::FSGNJ_H:
2142 case RISCV::FSGNJ_D_INX:
2143 case RISCV::FSGNJ_D_IN32X:
2144 case RISCV::FSGNJ_S_INX:
2145 case RISCV::FSGNJ_H_INX:
2146 // The canonical floating-point move is fsgnj rd, rs, rs.
2147 if (MI.getOperand(1).isReg() && MI.getOperand(2).isReg() &&
2148 MI.getOperand(1).getReg() == MI.getOperand(2).getReg())
2149 return DestSourcePair{MI.getOperand(0), MI.getOperand(1)};
2150 break;
2151 }
2152 return std::nullopt;
2153}
2154
2156 if (ForceMachineCombinerStrategy.getNumOccurrences() == 0) {
2157 // The option is unused. Choose Local strategy only for in-order cores. When
2158 // scheduling model is unspecified, use MinInstrCount strategy as more
2159 // generic one.
2160 const auto &SchedModel = STI.getSchedModel();
2161 return (!SchedModel.hasInstrSchedModel() || SchedModel.isOutOfOrder())
2164 }
2165 // The strategy was forced by the option.
2167}
2168
2170 MachineInstr &Root, unsigned &Pattern,
2171 SmallVectorImpl<MachineInstr *> &InsInstrs) const {
2172 int16_t FrmOpIdx =
2173 RISCV::getNamedOperandIdx(Root.getOpcode(), RISCV::OpName::frm);
2174 if (FrmOpIdx < 0) {
2175 assert(all_of(InsInstrs,
2176 [](MachineInstr *MI) {
2177 return RISCV::getNamedOperandIdx(MI->getOpcode(),
2178 RISCV::OpName::frm) < 0;
2179 }) &&
2180 "New instructions require FRM whereas the old one does not have it");
2181 return;
2182 }
2183
2184 const MachineOperand &FRM = Root.getOperand(FrmOpIdx);
2185 MachineFunction &MF = *Root.getMF();
2186
2187 for (auto *NewMI : InsInstrs) {
2188 // We'd already added the FRM operand.
2189 if (static_cast<unsigned>(RISCV::getNamedOperandIdx(
2190 NewMI->getOpcode(), RISCV::OpName::frm)) != NewMI->getNumOperands())
2191 continue;
2192 MachineInstrBuilder MIB(MF, NewMI);
2193 MIB.add(FRM);
2194 if (FRM.getImm() == RISCVFPRndMode::DYN)
2195 MIB.addUse(RISCV::FRM, RegState::Implicit);
2196 }
2197}
2198
2199static bool isFADD(unsigned Opc) {
2200 switch (Opc) {
2201 default:
2202 return false;
2203 case RISCV::FADD_H:
2204 case RISCV::FADD_S:
2205 case RISCV::FADD_D:
2206 return true;
2207 }
2208}
2209
2210static bool isFSUB(unsigned Opc) {
2211 switch (Opc) {
2212 default:
2213 return false;
2214 case RISCV::FSUB_H:
2215 case RISCV::FSUB_S:
2216 case RISCV::FSUB_D:
2217 return true;
2218 }
2219}
2220
2221static bool isFMUL(unsigned Opc) {
2222 switch (Opc) {
2223 default:
2224 return false;
2225 case RISCV::FMUL_H:
2226 case RISCV::FMUL_S:
2227 case RISCV::FMUL_D:
2228 return true;
2229 }
2230}
2231
2232bool RISCVInstrInfo::isVectorAssociativeAndCommutative(const MachineInstr &Inst,
2233 bool Invert) const {
2234#define OPCODE_LMUL_CASE(OPC) \
2235 case RISCV::OPC##_M1: \
2236 case RISCV::OPC##_M2: \
2237 case RISCV::OPC##_M4: \
2238 case RISCV::OPC##_M8: \
2239 case RISCV::OPC##_MF2: \
2240 case RISCV::OPC##_MF4: \
2241 case RISCV::OPC##_MF8
2242
2243#define OPCODE_LMUL_MASK_CASE(OPC) \
2244 case RISCV::OPC##_M1_MASK: \
2245 case RISCV::OPC##_M2_MASK: \
2246 case RISCV::OPC##_M4_MASK: \
2247 case RISCV::OPC##_M8_MASK: \
2248 case RISCV::OPC##_MF2_MASK: \
2249 case RISCV::OPC##_MF4_MASK: \
2250 case RISCV::OPC##_MF8_MASK
2251
2252 unsigned Opcode = Inst.getOpcode();
2253 if (Invert) {
2254 if (auto InvOpcode = getInverseOpcode(Opcode))
2255 Opcode = *InvOpcode;
2256 else
2257 return false;
2258 }
2259
2260 // clang-format off
2261 switch (Opcode) {
2262 default:
2263 return false;
2264 OPCODE_LMUL_CASE(PseudoVADD_VV):
2265 OPCODE_LMUL_MASK_CASE(PseudoVADD_VV):
2266 OPCODE_LMUL_CASE(PseudoVMUL_VV):
2267 OPCODE_LMUL_MASK_CASE(PseudoVMUL_VV):
2268 return true;
2269 }
2270 // clang-format on
2271
2272#undef OPCODE_LMUL_MASK_CASE
2273#undef OPCODE_LMUL_CASE
2274}
2275
2276bool RISCVInstrInfo::areRVVInstsReassociable(const MachineInstr &Root,
2277 const MachineInstr &Prev) const {
2278 if (!areOpcodesEqualOrInverse(Root.getOpcode(), Prev.getOpcode()))
2279 return false;
2280
2281 assert(Root.getMF() == Prev.getMF());
2282 const MachineRegisterInfo *MRI = &Root.getMF()->getRegInfo();
2283 const TargetRegisterInfo *TRI = MRI->getTargetRegisterInfo();
2284
2285 // Make sure vtype operands are also the same.
2286 const MCInstrDesc &Desc = get(Root.getOpcode());
2287 const uint64_t TSFlags = Desc.TSFlags;
2288
2289 auto checkImmOperand = [&](unsigned OpIdx) {
2290 return Root.getOperand(OpIdx).getImm() == Prev.getOperand(OpIdx).getImm();
2291 };
2292
2293 auto checkRegOperand = [&](unsigned OpIdx) {
2294 return Root.getOperand(OpIdx).getReg() == Prev.getOperand(OpIdx).getReg();
2295 };
2296
2297 // PassThru
2298 // TODO: Potentially we can loosen the condition to consider Root to be
2299 // associable with Prev if Root has NoReg as passthru. In which case we
2300 // also need to loosen the condition on vector policies between these.
2301 if (!checkRegOperand(1))
2302 return false;
2303
2304 // SEW
2305 if (RISCVII::hasSEWOp(TSFlags) &&
2306 !checkImmOperand(RISCVII::getSEWOpNum(Desc)))
2307 return false;
2308
2309 // Mask
2310 if (RISCVII::usesMaskPolicy(TSFlags)) {
2311 const MachineBasicBlock *MBB = Root.getParent();
2314 Register MI1VReg;
2315
2316 bool SeenMI2 = false;
2317 for (auto End = MBB->rend(), It = It1; It != End; ++It) {
2318 if (It == It2) {
2319 SeenMI2 = true;
2320 if (!MI1VReg.isValid())
2321 // There is no V0 def between Root and Prev; they're sharing the
2322 // same V0.
2323 break;
2324 }
2325
2326 if (It->modifiesRegister(RISCV::V0, TRI)) {
2327 Register SrcReg = It->getOperand(1).getReg();
2328 // If it's not VReg it'll be more difficult to track its defs, so
2329 // bailing out here just to be safe.
2330 if (!SrcReg.isVirtual())
2331 return false;
2332
2333 if (!MI1VReg.isValid()) {
2334 // This is the V0 def for Root.
2335 MI1VReg = SrcReg;
2336 continue;
2337 }
2338
2339 // Some random mask updates.
2340 if (!SeenMI2)
2341 continue;
2342
2343 // This is the V0 def for Prev; check if it's the same as that of
2344 // Root.
2345 if (MI1VReg != SrcReg)
2346 return false;
2347 else
2348 break;
2349 }
2350 }
2351
2352 // If we haven't encountered Prev, it's likely that this function was
2353 // called in a wrong way (e.g. Root is before Prev).
2354 assert(SeenMI2 && "Prev is expected to appear before Root");
2355 }
2356
2357 // Tail / Mask policies
2358 if (RISCVII::hasVecPolicyOp(TSFlags) &&
2359 !checkImmOperand(RISCVII::getVecPolicyOpNum(Desc)))
2360 return false;
2361
2362 // VL
2363 if (RISCVII::hasVLOp(TSFlags)) {
2364 unsigned OpIdx = RISCVII::getVLOpNum(Desc);
2365 const MachineOperand &Op1 = Root.getOperand(OpIdx);
2366 const MachineOperand &Op2 = Prev.getOperand(OpIdx);
2367 if (Op1.getType() != Op2.getType())
2368 return false;
2369 switch (Op1.getType()) {
2371 if (Op1.getReg() != Op2.getReg())
2372 return false;
2373 break;
2375 if (Op1.getImm() != Op2.getImm())
2376 return false;
2377 break;
2378 default:
2379 llvm_unreachable("Unrecognized VL operand type");
2380 }
2381 }
2382
2383 // Rounding modes
2384 if (RISCVII::hasRoundModeOp(TSFlags) &&
2385 !checkImmOperand(RISCVII::getVLOpNum(Desc) - 1))
2386 return false;
2387
2388 return true;
2389}
2390
2391// Most of our RVV pseudos have passthru operand, so the real operands
2392// start from index = 2.
2393bool RISCVInstrInfo::hasReassociableVectorSibling(const MachineInstr &Inst,
2394 bool &Commuted) const {
2395 const MachineBasicBlock *MBB = Inst.getParent();
2396 const MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo();
2398 "Expect the present of passthrough operand.");
2399 MachineInstr *MI1 = MRI.getUniqueVRegDef(Inst.getOperand(2).getReg());
2400 MachineInstr *MI2 = MRI.getUniqueVRegDef(Inst.getOperand(3).getReg());
2401
2402 // If only one operand has the same or inverse opcode and it's the second
2403 // source operand, the operands must be commuted.
2404 Commuted = !areRVVInstsReassociable(Inst, *MI1) &&
2405 areRVVInstsReassociable(Inst, *MI2);
2406 if (Commuted)
2407 std::swap(MI1, MI2);
2408
2409 return areRVVInstsReassociable(Inst, *MI1) &&
2410 (isVectorAssociativeAndCommutative(*MI1) ||
2411 isVectorAssociativeAndCommutative(*MI1, /* Invert */ true)) &&
2413 MRI.hasOneNonDBGUse(MI1->getOperand(0).getReg());
2414}
2415
2417 const MachineInstr &Inst, const MachineBasicBlock *MBB) const {
2418 if (!isVectorAssociativeAndCommutative(Inst) &&
2419 !isVectorAssociativeAndCommutative(Inst, /*Invert=*/true))
2421
2422 const MachineOperand &Op1 = Inst.getOperand(2);
2423 const MachineOperand &Op2 = Inst.getOperand(3);
2424 const MachineRegisterInfo &MRI = MBB->getParent()->getRegInfo();
2425
2426 // We need virtual register definitions for the operands that we will
2427 // reassociate.
2428 MachineInstr *MI1 = nullptr;
2429 MachineInstr *MI2 = nullptr;
2430 if (Op1.isReg() && Op1.getReg().isVirtual())
2431 MI1 = MRI.getUniqueVRegDef(Op1.getReg());
2432 if (Op2.isReg() && Op2.getReg().isVirtual())
2433 MI2 = MRI.getUniqueVRegDef(Op2.getReg());
2434
2435 // And at least one operand must be defined in MBB.
2436 return MI1 && MI2 && (MI1->getParent() == MBB || MI2->getParent() == MBB);
2437}
2438
2440 const MachineInstr &Root, unsigned Pattern,
2441 std::array<unsigned, 5> &OperandIndices) const {
2443 if (RISCV::getRVVMCOpcode(Root.getOpcode())) {
2444 // Skip the passthrough operand, so increment all indices by one.
2445 for (unsigned I = 0; I < 5; ++I)
2446 ++OperandIndices[I];
2447 }
2448}
2449
2451 bool &Commuted) const {
2452 if (isVectorAssociativeAndCommutative(Inst) ||
2453 isVectorAssociativeAndCommutative(Inst, /*Invert=*/true))
2454 return hasReassociableVectorSibling(Inst, Commuted);
2455
2456 if (!TargetInstrInfo::hasReassociableSibling(Inst, Commuted))
2457 return false;
2458
2459 const MachineRegisterInfo &MRI = Inst.getMF()->getRegInfo();
2460 unsigned OperandIdx = Commuted ? 2 : 1;
2461 const MachineInstr &Sibling =
2462 *MRI.getVRegDef(Inst.getOperand(OperandIdx).getReg());
2463
2464 int16_t InstFrmOpIdx =
2465 RISCV::getNamedOperandIdx(Inst.getOpcode(), RISCV::OpName::frm);
2466 int16_t SiblingFrmOpIdx =
2467 RISCV::getNamedOperandIdx(Sibling.getOpcode(), RISCV::OpName::frm);
2468
2469 return (InstFrmOpIdx < 0 && SiblingFrmOpIdx < 0) ||
2470 RISCV::hasEqualFRM(Inst, Sibling);
2471}
2472
2474 bool Invert) const {
2475 if (isVectorAssociativeAndCommutative(Inst, Invert))
2476 return true;
2477
2478 unsigned Opc = Inst.getOpcode();
2479 if (Invert) {
2480 auto InverseOpcode = getInverseOpcode(Opc);
2481 if (!InverseOpcode)
2482 return false;
2483 Opc = *InverseOpcode;
2484 }
2485
2486 if (isFADD(Opc) || isFMUL(Opc))
2489
2490 switch (Opc) {
2491 default:
2492 return false;
2493 case RISCV::ADD:
2494 case RISCV::ADDW:
2495 case RISCV::AND:
2496 case RISCV::OR:
2497 case RISCV::XOR:
2498 // From RISC-V ISA spec, if both the high and low bits of the same product
2499 // are required, then the recommended code sequence is:
2500 //
2501 // MULH[[S]U] rdh, rs1, rs2
2502 // MUL rdl, rs1, rs2
2503 // (source register specifiers must be in same order and rdh cannot be the
2504 // same as rs1 or rs2)
2505 //
2506 // Microarchitectures can then fuse these into a single multiply operation
2507 // instead of performing two separate multiplies.
2508 // MachineCombiner may reassociate MUL operands and lose the fusion
2509 // opportunity.
2510 case RISCV::MUL:
2511 case RISCV::MULW:
2512 case RISCV::MIN:
2513 case RISCV::MINU:
2514 case RISCV::MAX:
2515 case RISCV::MAXU:
2516 case RISCV::FMIN_H:
2517 case RISCV::FMIN_S:
2518 case RISCV::FMIN_D:
2519 case RISCV::FMAX_H:
2520 case RISCV::FMAX_S:
2521 case RISCV::FMAX_D:
2522 return true;
2523 }
2524
2525 return false;
2526}
2527
2528std::optional<unsigned>
2529RISCVInstrInfo::getInverseOpcode(unsigned Opcode) const {
2530#define RVV_OPC_LMUL_CASE(OPC, INV) \
2531 case RISCV::OPC##_M1: \
2532 return RISCV::INV##_M1; \
2533 case RISCV::OPC##_M2: \
2534 return RISCV::INV##_M2; \
2535 case RISCV::OPC##_M4: \
2536 return RISCV::INV##_M4; \
2537 case RISCV::OPC##_M8: \
2538 return RISCV::INV##_M8; \
2539 case RISCV::OPC##_MF2: \
2540 return RISCV::INV##_MF2; \
2541 case RISCV::OPC##_MF4: \
2542 return RISCV::INV##_MF4; \
2543 case RISCV::OPC##_MF8: \
2544 return RISCV::INV##_MF8
2545
2546#define RVV_OPC_LMUL_MASK_CASE(OPC, INV) \
2547 case RISCV::OPC##_M1_MASK: \
2548 return RISCV::INV##_M1_MASK; \
2549 case RISCV::OPC##_M2_MASK: \
2550 return RISCV::INV##_M2_MASK; \
2551 case RISCV::OPC##_M4_MASK: \
2552 return RISCV::INV##_M4_MASK; \
2553 case RISCV::OPC##_M8_MASK: \
2554 return RISCV::INV##_M8_MASK; \
2555 case RISCV::OPC##_MF2_MASK: \
2556 return RISCV::INV##_MF2_MASK; \
2557 case RISCV::OPC##_MF4_MASK: \
2558 return RISCV::INV##_MF4_MASK; \
2559 case RISCV::OPC##_MF8_MASK: \
2560 return RISCV::INV##_MF8_MASK
2561
2562 switch (Opcode) {
2563 default:
2564 return std::nullopt;
2565 case RISCV::FADD_H:
2566 return RISCV::FSUB_H;
2567 case RISCV::FADD_S:
2568 return RISCV::FSUB_S;
2569 case RISCV::FADD_D:
2570 return RISCV::FSUB_D;
2571 case RISCV::FSUB_H:
2572 return RISCV::FADD_H;
2573 case RISCV::FSUB_S:
2574 return RISCV::FADD_S;
2575 case RISCV::FSUB_D:
2576 return RISCV::FADD_D;
2577 case RISCV::ADD:
2578 return RISCV::SUB;
2579 case RISCV::SUB:
2580 return RISCV::ADD;
2581 case RISCV::ADDW:
2582 return RISCV::SUBW;
2583 case RISCV::SUBW:
2584 return RISCV::ADDW;
2585 // clang-format off
2586 RVV_OPC_LMUL_CASE(PseudoVADD_VV, PseudoVSUB_VV);
2587 RVV_OPC_LMUL_MASK_CASE(PseudoVADD_VV, PseudoVSUB_VV);
2588 RVV_OPC_LMUL_CASE(PseudoVSUB_VV, PseudoVADD_VV);
2589 RVV_OPC_LMUL_MASK_CASE(PseudoVSUB_VV, PseudoVADD_VV);
2590 // clang-format on
2591 }
2592
2593#undef RVV_OPC_LMUL_MASK_CASE
2594#undef RVV_OPC_LMUL_CASE
2595}
2596
2598 const MachineOperand &MO,
2599 bool DoRegPressureReduce) {
2600 if (!MO.isReg() || !MO.getReg().isVirtual())
2601 return false;
2602 const MachineRegisterInfo &MRI = Root.getMF()->getRegInfo();
2603 MachineInstr *MI = MRI.getVRegDef(MO.getReg());
2604 if (!MI || !isFMUL(MI->getOpcode()))
2605 return false;
2606
2609 return false;
2610
2611 // Try combining even if fmul has more than one use as it eliminates
2612 // dependency between fadd(fsub) and fmul. However, it can extend liveranges
2613 // for fmul operands, so reject the transformation in register pressure
2614 // reduction mode.
2615 if (DoRegPressureReduce && !MRI.hasOneNonDBGUse(MI->getOperand(0).getReg()))
2616 return false;
2617
2618 // Do not combine instructions from different basic blocks.
2619 if (Root.getParent() != MI->getParent())
2620 return false;
2621 return RISCV::hasEqualFRM(Root, *MI);
2622}
2623
2625 SmallVectorImpl<unsigned> &Patterns,
2626 bool DoRegPressureReduce) {
2627 unsigned Opc = Root.getOpcode();
2628 bool IsFAdd = isFADD(Opc);
2629 if (!IsFAdd && !isFSUB(Opc))
2630 return false;
2631 bool Added = false;
2632 if (canCombineFPFusedMultiply(Root, Root.getOperand(1),
2633 DoRegPressureReduce)) {
2636 Added = true;
2637 }
2638 if (canCombineFPFusedMultiply(Root, Root.getOperand(2),
2639 DoRegPressureReduce)) {
2642 Added = true;
2643 }
2644 return Added;
2645}
2646
2647static bool getFPPatterns(MachineInstr &Root,
2648 SmallVectorImpl<unsigned> &Patterns,
2649 bool DoRegPressureReduce) {
2650 return getFPFusedMultiplyPatterns(Root, Patterns, DoRegPressureReduce);
2651}
2652
2653/// Utility routine that checks if \param MO is defined by an
2654/// \param CombineOpc instruction in the basic block \param MBB
2656 const MachineOperand &MO,
2657 unsigned CombineOpc) {
2658 const MachineRegisterInfo &MRI = MBB.getParent()->getRegInfo();
2659 const MachineInstr *MI = nullptr;
2660
2661 if (MO.isReg() && MO.getReg().isVirtual())
2662 MI = MRI.getUniqueVRegDef(MO.getReg());
2663 // And it needs to be in the trace (otherwise, it won't have a depth).
2664 if (!MI || MI->getParent() != &MBB || MI->getOpcode() != CombineOpc)
2665 return nullptr;
2666 // Must only used by the user we combine with.
2667 if (!MRI.hasOneNonDBGUse(MI->getOperand(0).getReg()))
2668 return nullptr;
2669
2670 return MI;
2671}
2672
2673/// Utility routine that checks if \param MO is defined by a SLLI in \param
2674/// MBB that can be combined by splitting across 2 SHXADD instructions. The
2675/// first SHXADD shift amount is given by \param OuterShiftAmt.
2677 const MachineOperand &MO,
2678 unsigned OuterShiftAmt) {
2679 const MachineInstr *ShiftMI = canCombine(MBB, MO, RISCV::SLLI);
2680 if (!ShiftMI)
2681 return false;
2682
2683 unsigned InnerShiftAmt = ShiftMI->getOperand(2).getImm();
2684 if (InnerShiftAmt < OuterShiftAmt || (InnerShiftAmt - OuterShiftAmt) > 3)
2685 return false;
2686
2687 return true;
2688}
2689
2690// Returns the shift amount from a SHXADD instruction. Returns 0 if the
2691// instruction is not a SHXADD.
2692static unsigned getSHXADDShiftAmount(unsigned Opc) {
2693 switch (Opc) {
2694 default:
2695 return 0;
2696 case RISCV::SH1ADD:
2697 return 1;
2698 case RISCV::SH2ADD:
2699 return 2;
2700 case RISCV::SH3ADD:
2701 return 3;
2702 }
2703}
2704
2705// Returns the shift amount from a SHXADD.UW instruction. Returns 0 if the
2706// instruction is not a SHXADD.UW.
2707static unsigned getSHXADDUWShiftAmount(unsigned Opc) {
2708 switch (Opc) {
2709 default:
2710 return 0;
2711 case RISCV::SH1ADD_UW:
2712 return 1;
2713 case RISCV::SH2ADD_UW:
2714 return 2;
2715 case RISCV::SH3ADD_UW:
2716 return 3;
2717 }
2718}
2719
2720// Look for opportunities to combine (sh3add Z, (add X, (slli Y, 5))) into
2721// (sh3add (sh2add Y, Z), X).
2722static bool getSHXADDPatterns(const MachineInstr &Root,
2723 SmallVectorImpl<unsigned> &Patterns) {
2724 unsigned ShiftAmt = getSHXADDShiftAmount(Root.getOpcode());
2725 if (!ShiftAmt)
2726 return false;
2727
2728 const MachineBasicBlock &MBB = *Root.getParent();
2729
2730 const MachineInstr *AddMI = canCombine(MBB, Root.getOperand(2), RISCV::ADD);
2731 if (!AddMI)
2732 return false;
2733
2734 bool Found = false;
2735 if (canCombineShiftIntoShXAdd(MBB, AddMI->getOperand(1), ShiftAmt)) {
2737 Found = true;
2738 }
2739 if (canCombineShiftIntoShXAdd(MBB, AddMI->getOperand(2), ShiftAmt)) {
2741 Found = true;
2742 }
2743
2744 return Found;
2745}
2746
2758
2760 MachineInstr &Root, SmallVectorImpl<unsigned> &Patterns,
2761 bool DoRegPressureReduce) const {
2762
2763 if (getFPPatterns(Root, Patterns, DoRegPressureReduce))
2764 return true;
2765
2766 if (getSHXADDPatterns(Root, Patterns))
2767 return true;
2768
2769 return TargetInstrInfo::getMachineCombinerPatterns(Root, Patterns,
2770 DoRegPressureReduce);
2771}
2772
2773static unsigned getFPFusedMultiplyOpcode(unsigned RootOpc, unsigned Pattern) {
2774 switch (RootOpc) {
2775 default:
2776 llvm_unreachable("Unexpected opcode");
2777 case RISCV::FADD_H:
2778 return RISCV::FMADD_H;
2779 case RISCV::FADD_S:
2780 return RISCV::FMADD_S;
2781 case RISCV::FADD_D:
2782 return RISCV::FMADD_D;
2783 case RISCV::FSUB_H:
2784 return Pattern == RISCVMachineCombinerPattern::FMSUB ? RISCV::FMSUB_H
2785 : RISCV::FNMSUB_H;
2786 case RISCV::FSUB_S:
2787 return Pattern == RISCVMachineCombinerPattern::FMSUB ? RISCV::FMSUB_S
2788 : RISCV::FNMSUB_S;
2789 case RISCV::FSUB_D:
2790 return Pattern == RISCVMachineCombinerPattern::FMSUB ? RISCV::FMSUB_D
2791 : RISCV::FNMSUB_D;
2792 }
2793}
2794
2795static unsigned getAddendOperandIdx(unsigned Pattern) {
2796 switch (Pattern) {
2797 default:
2798 llvm_unreachable("Unexpected pattern");
2801 return 2;
2804 return 1;
2805 }
2806}
2807
2809 unsigned Pattern,
2812 MachineFunction *MF = Root.getMF();
2815
2816 MachineOperand &Mul1 = Prev.getOperand(1);
2817 MachineOperand &Mul2 = Prev.getOperand(2);
2818 MachineOperand &Dst = Root.getOperand(0);
2820
2821 Register DstReg = Dst.getReg();
2822 unsigned FusedOpc = getFPFusedMultiplyOpcode(Root.getOpcode(), Pattern);
2823 uint32_t IntersectedFlags = Root.getFlags() & Prev.getFlags();
2824 DebugLoc MergedLoc =
2826
2827 bool Mul1IsKill = Mul1.isKill();
2828 bool Mul2IsKill = Mul2.isKill();
2829 bool AddendIsKill = Addend.isKill();
2830
2831 // We need to clear kill flags since we may be extending the live range past
2832 // a kill. If the mul had kill flags, we can preserve those since we know
2833 // where the previous range stopped.
2834 MRI.clearKillFlags(Mul1.getReg());
2835 MRI.clearKillFlags(Mul2.getReg());
2836
2838 BuildMI(*MF, MergedLoc, TII->get(FusedOpc), DstReg)
2839 .addReg(Mul1.getReg(), getKillRegState(Mul1IsKill))
2840 .addReg(Mul2.getReg(), getKillRegState(Mul2IsKill))
2841 .addReg(Addend.getReg(), getKillRegState(AddendIsKill))
2842 .setMIFlags(IntersectedFlags);
2843
2844 InsInstrs.push_back(MIB);
2845 if (MRI.hasOneNonDBGUse(Prev.getOperand(0).getReg()))
2846 DelInstrs.push_back(&Prev);
2847 DelInstrs.push_back(&Root);
2848}
2849
2850// Combine patterns like (sh3add Z, (add X, (slli Y, 5))) to
2851// (sh3add (sh2add Y, Z), X) if the shift amount can be split across two
2852// shXadd instructions. The outer shXadd keeps its original opcode.
2853static void
2854genShXAddAddShift(MachineInstr &Root, unsigned AddOpIdx,
2857 DenseMap<Register, unsigned> &InstrIdxForVirtReg) {
2858 MachineFunction *MF = Root.getMF();
2861
2862 unsigned OuterShiftAmt = getSHXADDShiftAmount(Root.getOpcode());
2863 assert(OuterShiftAmt != 0 && "Unexpected opcode");
2864
2865 MachineInstr *AddMI = MRI.getUniqueVRegDef(Root.getOperand(2).getReg());
2866 MachineInstr *ShiftMI =
2867 MRI.getUniqueVRegDef(AddMI->getOperand(AddOpIdx).getReg());
2868
2869 unsigned InnerShiftAmt = ShiftMI->getOperand(2).getImm();
2870 assert(InnerShiftAmt >= OuterShiftAmt && "Unexpected shift amount");
2871
2872 unsigned InnerOpc;
2873 switch (InnerShiftAmt - OuterShiftAmt) {
2874 default:
2875 llvm_unreachable("Unexpected shift amount");
2876 case 0:
2877 InnerOpc = RISCV::ADD;
2878 break;
2879 case 1:
2880 InnerOpc = RISCV::SH1ADD;
2881 break;
2882 case 2:
2883 InnerOpc = RISCV::SH2ADD;
2884 break;
2885 case 3:
2886 InnerOpc = RISCV::SH3ADD;
2887 break;
2888 }
2889
2890 const MachineOperand &X = AddMI->getOperand(3 - AddOpIdx);
2891 const MachineOperand &Y = ShiftMI->getOperand(1);
2892 const MachineOperand &Z = Root.getOperand(1);
2893
2894 Register NewVR = MRI.createVirtualRegister(&RISCV::GPRRegClass);
2895
2896 auto MIB1 = BuildMI(*MF, MIMetadata(Root), TII->get(InnerOpc), NewVR)
2897 .addReg(Y.getReg(), getKillRegState(Y.isKill()))
2898 .addReg(Z.getReg(), getKillRegState(Z.isKill()));
2899 auto MIB2 = BuildMI(*MF, MIMetadata(Root), TII->get(Root.getOpcode()),
2900 Root.getOperand(0).getReg())
2901 .addReg(NewVR, RegState::Kill)
2902 .addReg(X.getReg(), getKillRegState(X.isKill()));
2903
2904 InstrIdxForVirtReg.insert(std::make_pair(NewVR, 0));
2905 InsInstrs.push_back(MIB1);
2906 InsInstrs.push_back(MIB2);
2907 DelInstrs.push_back(ShiftMI);
2908 DelInstrs.push_back(AddMI);
2909 DelInstrs.push_back(&Root);
2910}
2911
2913 MachineInstr &Root, unsigned Pattern,
2916 DenseMap<Register, unsigned> &InstrIdxForVirtReg) const {
2918 switch (Pattern) {
2919 default:
2921 DelInstrs, InstrIdxForVirtReg);
2922 return;
2925 MachineInstr &Prev = *MRI.getVRegDef(Root.getOperand(1).getReg());
2926 combineFPFusedMultiply(Root, Prev, Pattern, InsInstrs, DelInstrs);
2927 return;
2928 }
2931 MachineInstr &Prev = *MRI.getVRegDef(Root.getOperand(2).getReg());
2932 combineFPFusedMultiply(Root, Prev, Pattern, InsInstrs, DelInstrs);
2933 return;
2934 }
2936 genShXAddAddShift(Root, 1, InsInstrs, DelInstrs, InstrIdxForVirtReg);
2937 return;
2939 genShXAddAddShift(Root, 2, InsInstrs, DelInstrs, InstrIdxForVirtReg);
2940 return;
2941 }
2942}
2943
2945 StringRef &ErrInfo) const {
2946 MCInstrDesc const &Desc = MI.getDesc();
2947
2948 for (const auto &[Index, Operand] : enumerate(Desc.operands())) {
2949 const MachineOperand &MO = MI.getOperand(Index);
2950 unsigned OpType = Operand.OperandType;
2951 switch (OpType) {
2952 default:
2953 if (OpType >= RISCVOp::OPERAND_FIRST_RISCV_IMM &&
2955 if (!MO.isImm()) {
2956 ErrInfo = "Expected an immediate operand.";
2957 return false;
2958 }
2959 int64_t Imm = MO.getImm();
2960 bool Ok;
2961 switch (OpType) {
2962 default:
2963 llvm_unreachable("Unexpected operand type");
2964
2965 // clang-format off
2966#define CASE_OPERAND_UIMM(NUM) \
2967 case RISCVOp::OPERAND_UIMM##NUM: \
2968 Ok = isUInt<NUM>(Imm); \
2969 break;
2970#define CASE_OPERAND_SIMM(NUM) \
2971 case RISCVOp::OPERAND_SIMM##NUM: \
2972 Ok = isInt<NUM>(Imm); \
2973 break;
2989 // clang-format on
2991 Ok = isShiftedUInt<1, 1>(Imm);
2992 break;
2994 Ok = isShiftedUInt<4, 1>(Imm);
2995 break;
2997 Ok = isUInt<5>(Imm) && (Imm != 0);
2998 break;
3000 Ok = isUInt<5>(Imm) && (Imm > 3);
3001 break;
3003 Ok = Imm >= 1 && Imm <= 32;
3004 break;
3006 Ok = isShiftedUInt<5, 1>(Imm);
3007 break;
3009 Ok = isShiftedUInt<5, 2>(Imm);
3010 break;
3012 Ok = isShiftedUInt<4, 3>(Imm);
3013 break;
3015 Ok = isShiftedUInt<6, 2>(Imm);
3016 break;
3018 Ok = isShiftedUInt<5, 3>(Imm);
3019 break;
3021 Ok = isUInt<8>(Imm) && Imm >= 32;
3022 break;
3024 Ok = isShiftedUInt<6, 3>(Imm);
3025 break;
3027 Ok = isInt<8>(Imm);
3028 break;
3030 Ok = isShiftedInt<6, 4>(Imm) && (Imm != 0);
3031 break;
3033 Ok = isShiftedUInt<8, 2>(Imm) && (Imm != 0);
3034 break;
3036 Ok = isUInt<16>(Imm) && (Imm != 0);
3037 break;
3039 Ok = Imm == 3;
3040 break;
3042 Ok = Imm == 4;
3043 break;
3045 Ok = (isUInt<5>(Imm) && Imm != 0) || Imm == -1;
3046 break;
3047 // clang-format off
3053 // clang-format on
3055 Ok = Imm >= -15 && Imm <= 16;
3056 break;
3058 Ok = isInt<5>(Imm) && (Imm != 0);
3059 break;
3061 Ok = Imm != 0 && isInt<6>(Imm);
3062 break;
3064 Ok = isUInt<10>(Imm);
3065 break;
3067 Ok = isUInt<11>(Imm);
3068 break;
3070 Ok = isShiftedInt<7, 5>(Imm);
3071 break;
3073 Ok = isInt<16>(Imm) && (Imm != 0);
3074 break;
3076 Ok = isInt<20>(Imm);
3077 break;
3079 Ok = STI.is64Bit() ? isUInt<6>(Imm) : isUInt<5>(Imm);
3080 break;
3082 Ok = STI.is64Bit() ? isUInt<6>(Imm) : isUInt<5>(Imm);
3083 Ok = Ok && Imm != 0;
3084 break;
3086 Ok = (isUInt<5>(Imm) && Imm != 0) || (Imm >= 0xfffe0 && Imm <= 0xfffff);
3087 break;
3089 Ok = Imm >= 0 && Imm <= 10;
3090 break;
3092 Ok = Imm >= 0 && Imm <= 7;
3093 break;
3095 Ok = Imm >= 1 && Imm <= 10;
3096 break;
3098 Ok = Imm >= 2 && Imm <= 14;
3099 break;
3101 Ok = Imm >= RISCVZC::RA && Imm <= RISCVZC::RA_S0_S11;
3102 break;
3104 Ok = Imm >= RISCVZC::RA_S0 && Imm <= RISCVZC::RA_S0_S11;
3105 break;
3107 Ok = Imm >= 0 && Imm <= 48 && Imm % 16 == 0;
3108 break;
3111 break;
3113 Ok = Imm == RISCVFPRndMode::RTZ;
3114 break;
3116 Ok = Imm >= 0 && Imm < RISCVCC::COND_INVALID;
3117 break;
3119 Ok = isValidAtomicOrdering(Imm);
3120 break;
3123 Imm;
3124 break;
3126 Ok = (isUInt<5>(Imm) && RISCVVType::isValidSEW(1 << Imm));
3127 break;
3129 Ok = Imm == 0;
3130 break;
3133 if (RISCVII::usesVXRM(Desc.TSFlags))
3134 Ok = isUInt<2>(Imm);
3135 else
3137 break;
3140 break;
3142 Ok = Imm == 1 || Imm == 2 || Imm == 4;
3143 break;
3144 }
3145 if (!Ok) {
3146 ErrInfo = "Invalid immediate";
3147 return false;
3148 }
3149 }
3150 break;
3152 // TODO: We could be stricter about what non-register operands are
3153 // allowed.
3154 if (MO.isReg()) {
3155 ErrInfo = "Expected a non-register operand.";
3156 return false;
3157 }
3158 if (MO.isImm() && !isInt<12>(MO.getImm())) {
3159 ErrInfo = "Invalid immediate";
3160 return false;
3161 }
3162 break;
3165 // TODO: We could be stricter about what non-register operands are
3166 // allowed.
3167 if (MO.isReg()) {
3168 ErrInfo = "Expected a non-register operand.";
3169 return false;
3170 }
3171 if (MO.isImm() && !isUInt<20>(MO.getImm())) {
3172 ErrInfo = "Invalid immediate";
3173 return false;
3174 }
3175 break;
3177 // TODO: We could be stricter about what non-register operands are
3178 // allowed.
3179 if (MO.isReg()) {
3180 ErrInfo = "Expected a non-register operand.";
3181 return false;
3182 }
3183 if (MO.isImm() && !isInt<32>(MO.getImm())) {
3184 ErrInfo = "Invalid immediate";
3185 return false;
3186 }
3187 break;
3189 if (MO.isImm()) {
3190 int64_t Imm = MO.getImm();
3191 // VLMAX is represented as -1.
3192 if (!isUInt<5>(Imm) && Imm != -1) {
3193 ErrInfo = "Invalid immediate";
3194 return false;
3195 }
3196 } else if (!MO.isReg()) {
3197 ErrInfo = "Expected a register or immediate operand.";
3198 return false;
3199 }
3200 break;
3201 }
3202 }
3203
3204 const uint64_t TSFlags = Desc.TSFlags;
3205 if (RISCVII::hasVLOp(TSFlags)) {
3206 const MachineOperand &Op = MI.getOperand(RISCVII::getVLOpNum(Desc));
3207 if (!Op.isImm() && !Op.isReg()) {
3208 ErrInfo = "Invalid operand type for VL operand";
3209 return false;
3210 }
3211 if (Op.isReg() && Op.getReg().isValid()) {
3212 const MachineRegisterInfo &MRI = MI.getParent()->getParent()->getRegInfo();
3213 auto *RC = MRI.getRegClass(Op.getReg());
3214 if (!RISCV::GPRNoX0RegClass.hasSubClassEq(RC)) {
3215 ErrInfo = "Invalid register class for VL operand";
3216 return false;
3217 }
3218 }
3219 if (!RISCVII::hasSEWOp(TSFlags)) {
3220 ErrInfo = "VL operand w/o SEW operand?";
3221 return false;
3222 }
3223 }
3224 if (RISCVII::hasSEWOp(TSFlags)) {
3225 unsigned OpIdx = RISCVII::getSEWOpNum(Desc);
3226 if (!MI.getOperand(OpIdx).isImm()) {
3227 ErrInfo = "SEW value expected to be an immediate";
3228 return false;
3229 }
3230 uint64_t Log2SEW = MI.getOperand(OpIdx).getImm();
3231 if (Log2SEW > 31) {
3232 ErrInfo = "Unexpected SEW value";
3233 return false;
3234 }
3235 unsigned SEW = Log2SEW ? 1 << Log2SEW : 8;
3236 if (!RISCVVType::isValidSEW(SEW)) {
3237 ErrInfo = "Unexpected SEW value";
3238 return false;
3239 }
3240 }
3241 if (RISCVII::hasVecPolicyOp(TSFlags)) {
3243 if (!MI.getOperand(OpIdx).isImm()) {
3244 ErrInfo = "Policy operand expected to be an immediate";
3245 return false;
3246 }
3247 uint64_t Policy = MI.getOperand(OpIdx).getImm();
3249 ErrInfo = "Invalid Policy Value";
3250 return false;
3251 }
3252 if (!RISCVII::hasVLOp(TSFlags)) {
3253 ErrInfo = "policy operand w/o VL operand?";
3254 return false;
3255 }
3256
3257 // VecPolicy operands can only exist on instructions with passthru/merge
3258 // arguments. Note that not all arguments with passthru have vec policy
3259 // operands- some instructions have implicit policies.
3260 unsigned UseOpIdx;
3261 if (!MI.isRegTiedToUseOperand(0, &UseOpIdx)) {
3262 ErrInfo = "policy operand w/o tied operand?";
3263 return false;
3264 }
3265 }
3266
3267 if (int Idx = RISCVII::getFRMOpNum(Desc);
3268 Idx >= 0 && MI.getOperand(Idx).getImm() == RISCVFPRndMode::DYN &&
3269 !MI.readsRegister(RISCV::FRM, /*TRI=*/nullptr)) {
3270 ErrInfo = "dynamic rounding mode should read FRM";
3271 return false;
3272 }
3273
3274 return true;
3275}
3276
3278 const MachineInstr &AddrI,
3279 ExtAddrMode &AM) const {
3280 switch (MemI.getOpcode()) {
3281 default:
3282 return false;
3283 case RISCV::LB:
3284 case RISCV::LBU:
3285 case RISCV::LH:
3286 case RISCV::LH_INX:
3287 case RISCV::LHU:
3288 case RISCV::LW:
3289 case RISCV::LW_INX:
3290 case RISCV::LWU:
3291 case RISCV::LD:
3292 case RISCV::LD_RV32:
3293 case RISCV::FLH:
3294 case RISCV::FLW:
3295 case RISCV::FLD:
3296 case RISCV::SB:
3297 case RISCV::SH:
3298 case RISCV::SH_INX:
3299 case RISCV::SW:
3300 case RISCV::SW_INX:
3301 case RISCV::SD:
3302 case RISCV::SD_RV32:
3303 case RISCV::FSH:
3304 case RISCV::FSW:
3305 case RISCV::FSD:
3306 break;
3307 }
3308
3309 if (MemI.getOperand(0).getReg() == Reg)
3310 return false;
3311
3312 if (AddrI.getOpcode() != RISCV::ADDI || !AddrI.getOperand(1).isReg() ||
3313 !AddrI.getOperand(2).isImm())
3314 return false;
3315
3316 int64_t OldOffset = MemI.getOperand(2).getImm();
3317 int64_t Disp = AddrI.getOperand(2).getImm();
3318 int64_t NewOffset = OldOffset + Disp;
3319 if (!STI.is64Bit())
3320 NewOffset = SignExtend64<32>(NewOffset);
3321
3322 if (!isInt<12>(NewOffset))
3323 return false;
3324
3325 AM.BaseReg = AddrI.getOperand(1).getReg();
3326 AM.ScaledReg = 0;
3327 AM.Scale = 0;
3328 AM.Displacement = NewOffset;
3330 return true;
3331}
3332
3334 const ExtAddrMode &AM) const {
3335
3336 const DebugLoc &DL = MemI.getDebugLoc();
3337 MachineBasicBlock &MBB = *MemI.getParent();
3338
3339 assert(AM.ScaledReg == 0 && AM.Scale == 0 &&
3340 "Addressing mode not supported for folding");
3341
3342 return BuildMI(MBB, MemI, DL, get(MemI.getOpcode()))
3343 .addReg(MemI.getOperand(0).getReg(), getDefRegState(MemI.mayLoad()))
3344 .addReg(AM.BaseReg)
3345 .addImm(AM.Displacement)
3346 .setMemRefs(MemI.memoperands())
3347 .setMIFlags(MemI.getFlags());
3348}
3349
3350// TODO: At the moment, MIPS introduced paring of instructions operating with
3351// word or double word. This should be extended with more instructions when more
3352// vendors support load/store pairing.
3354 switch (Opc) {
3355 default:
3356 return false;
3357 case RISCV::SW:
3358 case RISCV::SD:
3359 case RISCV::LD:
3360 case RISCV::LW:
3361 return true;
3362 }
3363}
3364
3366 const TargetRegisterInfo *TRI) {
3367 // If this is a volatile load/store, don't mess with it.
3368 if (LdSt.hasOrderedMemoryRef() || LdSt.getNumExplicitOperands() != 3)
3369 return false;
3370
3371 if (LdSt.getOperand(1).isFI())
3372 return true;
3373
3374 assert(LdSt.getOperand(1).isReg() && "Expected a reg operand.");
3375 // Can't cluster if the instruction modifies the base register
3376 // or it is update form. e.g. ld x5,8(x5)
3377 if (LdSt.modifiesRegister(LdSt.getOperand(1).getReg(), TRI))
3378 return false;
3379
3380 if (!LdSt.getOperand(2).isImm())
3381 return false;
3382
3383 return true;
3384}
3385
3388 int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width,
3389 const TargetRegisterInfo *TRI) const {
3390 if (!LdSt.mayLoadOrStore())
3391 return false;
3392
3393 // Conservatively, only handle scalar loads/stores for now.
3394 switch (LdSt.getOpcode()) {
3395 case RISCV::LB:
3396 case RISCV::LBU:
3397 case RISCV::SB:
3398 case RISCV::LH:
3399 case RISCV::LH_INX:
3400 case RISCV::LHU:
3401 case RISCV::FLH:
3402 case RISCV::SH:
3403 case RISCV::SH_INX:
3404 case RISCV::FSH:
3405 case RISCV::LW:
3406 case RISCV::LW_INX:
3407 case RISCV::LWU:
3408 case RISCV::FLW:
3409 case RISCV::SW:
3410 case RISCV::SW_INX:
3411 case RISCV::FSW:
3412 case RISCV::LD:
3413 case RISCV::LD_RV32:
3414 case RISCV::FLD:
3415 case RISCV::SD:
3416 case RISCV::SD_RV32:
3417 case RISCV::FSD:
3418 break;
3419 default:
3420 return false;
3421 }
3422 const MachineOperand *BaseOp;
3423 OffsetIsScalable = false;
3424 if (!getMemOperandWithOffsetWidth(LdSt, BaseOp, Offset, Width, TRI))
3425 return false;
3426 BaseOps.push_back(BaseOp);
3427 return true;
3428}
3429
3430// TODO: This was copied from SIInstrInfo. Could it be lifted to a common
3431// helper?
3434 const MachineInstr &MI2,
3436 // Only examine the first "base" operand of each instruction, on the
3437 // assumption that it represents the real base address of the memory access.
3438 // Other operands are typically offsets or indices from this base address.
3439 if (BaseOps1.front()->isIdenticalTo(*BaseOps2.front()))
3440 return true;
3441
3442 if (!MI1.hasOneMemOperand() || !MI2.hasOneMemOperand())
3443 return false;
3444
3445 auto MO1 = *MI1.memoperands_begin();
3446 auto MO2 = *MI2.memoperands_begin();
3447 if (MO1->getAddrSpace() != MO2->getAddrSpace())
3448 return false;
3449
3450 auto Base1 = MO1->getValue();
3451 auto Base2 = MO2->getValue();
3452 if (!Base1 || !Base2)
3453 return false;
3454 Base1 = getUnderlyingObject(Base1);
3455 Base2 = getUnderlyingObject(Base2);
3456
3457 if (isa<UndefValue>(Base1) || isa<UndefValue>(Base2))
3458 return false;
3459
3460 return Base1 == Base2;
3461}
3462
3464 ArrayRef<const MachineOperand *> BaseOps1, int64_t Offset1,
3465 bool OffsetIsScalable1, ArrayRef<const MachineOperand *> BaseOps2,
3466 int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize,
3467 unsigned NumBytes) const {
3468 // If the mem ops (to be clustered) do not have the same base ptr, then they
3469 // should not be clustered
3470 if (!BaseOps1.empty() && !BaseOps2.empty()) {
3471 const MachineInstr &FirstLdSt = *BaseOps1.front()->getParent();
3472 const MachineInstr &SecondLdSt = *BaseOps2.front()->getParent();
3473 if (!memOpsHaveSameBasePtr(FirstLdSt, BaseOps1, SecondLdSt, BaseOps2))
3474 return false;
3475 } else if (!BaseOps1.empty() || !BaseOps2.empty()) {
3476 // If only one base op is empty, they do not have the same base ptr
3477 return false;
3478 }
3479
3480 unsigned CacheLineSize =
3481 BaseOps1.front()->getParent()->getMF()->getSubtarget().getCacheLineSize();
3482 // Assume a cache line size of 64 bytes if no size is set in RISCVSubtarget.
3484 // Cluster if the memory operations are on the same or a neighbouring cache
3485 // line, but limit the maximum ClusterSize to avoid creating too much
3486 // additional register pressure.
3487 return ClusterSize <= 4 && std::abs(Offset1 - Offset2) < CacheLineSize;
3488}
3489
3490// Set BaseReg (the base register operand), Offset (the byte offset being
3491// accessed) and the access Width of the passed instruction that reads/writes
3492// memory. Returns false if the instruction does not read/write memory or the
3493// BaseReg/Offset/Width can't be determined. Is not guaranteed to always
3494// recognise base operands and offsets in all cases.
3495// TODO: Add an IsScalable bool ref argument (like the equivalent AArch64
3496// function) and set it as appropriate.
3498 const MachineInstr &LdSt, const MachineOperand *&BaseReg, int64_t &Offset,
3499 LocationSize &Width, const TargetRegisterInfo *TRI) const {
3500 if (!LdSt.mayLoadOrStore())
3501 return false;
3502
3503 // Here we assume the standard RISC-V ISA, which uses a base+offset
3504 // addressing mode. You'll need to relax these conditions to support custom
3505 // load/store instructions.
3506 if (LdSt.getNumExplicitOperands() != 3)
3507 return false;
3508 if ((!LdSt.getOperand(1).isReg() && !LdSt.getOperand(1).isFI()) ||
3509 !LdSt.getOperand(2).isImm())
3510 return false;
3511
3512 if (!LdSt.hasOneMemOperand())
3513 return false;
3514
3515 Width = (*LdSt.memoperands_begin())->getSize();
3516 BaseReg = &LdSt.getOperand(1);
3517 Offset = LdSt.getOperand(2).getImm();
3518 return true;
3519}
3520
3522 const MachineInstr &MIa, const MachineInstr &MIb) const {
3523 assert(MIa.mayLoadOrStore() && "MIa must be a load or store.");
3524 assert(MIb.mayLoadOrStore() && "MIb must be a load or store.");
3525
3528 return false;
3529
3530 // Retrieve the base register, offset from the base register and width. Width
3531 // is the size of memory that is being loaded/stored (e.g. 1, 2, 4). If
3532 // base registers are identical, and the offset of a lower memory access +
3533 // the width doesn't overlap the offset of a higher memory access,
3534 // then the memory accesses are different.
3535 const TargetRegisterInfo *TRI = STI.getRegisterInfo();
3536 const MachineOperand *BaseOpA = nullptr, *BaseOpB = nullptr;
3537 int64_t OffsetA = 0, OffsetB = 0;
3539 WidthB = LocationSize::precise(0);
3540 if (getMemOperandWithOffsetWidth(MIa, BaseOpA, OffsetA, WidthA, TRI) &&
3541 getMemOperandWithOffsetWidth(MIb, BaseOpB, OffsetB, WidthB, TRI)) {
3542 if (BaseOpA->isIdenticalTo(*BaseOpB)) {
3543 int LowOffset = std::min(OffsetA, OffsetB);
3544 int HighOffset = std::max(OffsetA, OffsetB);
3545 LocationSize LowWidth = (LowOffset == OffsetA) ? WidthA : WidthB;
3546 if (LowWidth.hasValue() &&
3547 LowOffset + (int)LowWidth.getValue() <= HighOffset)
3548 return true;
3549 }
3550 }
3551 return false;
3552}
3553
3554std::pair<unsigned, unsigned>
3556 const unsigned Mask = RISCVII::MO_DIRECT_FLAG_MASK;
3557 return std::make_pair(TF & Mask, TF & ~Mask);
3558}
3559
3562 using namespace RISCVII;
3563 static const std::pair<unsigned, const char *> TargetFlags[] = {
3564 {MO_CALL, "riscv-call"},
3565 {MO_LO, "riscv-lo"},
3566 {MO_HI, "riscv-hi"},
3567 {MO_PCREL_LO, "riscv-pcrel-lo"},
3568 {MO_PCREL_HI, "riscv-pcrel-hi"},
3569 {MO_GOT_HI, "riscv-got-hi"},
3570 {MO_TPREL_LO, "riscv-tprel-lo"},
3571 {MO_TPREL_HI, "riscv-tprel-hi"},
3572 {MO_TPREL_ADD, "riscv-tprel-add"},
3573 {MO_TLS_GOT_HI, "riscv-tls-got-hi"},
3574 {MO_TLS_GD_HI, "riscv-tls-gd-hi"},
3575 {MO_TLSDESC_HI, "riscv-tlsdesc-hi"},
3576 {MO_TLSDESC_LOAD_LO, "riscv-tlsdesc-load-lo"},
3577 {MO_TLSDESC_ADD_LO, "riscv-tlsdesc-add-lo"},
3578 {MO_TLSDESC_CALL, "riscv-tlsdesc-call"}};
3579 return ArrayRef(TargetFlags);
3580}
3582 MachineFunction &MF, bool OutlineFromLinkOnceODRs) const {
3583 const Function &F = MF.getFunction();
3584
3585 // Can F be deduplicated by the linker? If it can, don't outline from it.
3586 if (!OutlineFromLinkOnceODRs && F.hasLinkOnceODRLinkage())
3587 return false;
3588
3589 // Don't outline from functions with section markings; the program could
3590 // expect that all the code is in the named section.
3591 if (F.hasSection())
3592 return false;
3593
3594 // It's safe to outline from MF.
3595 return true;
3596}
3597
3599 unsigned &Flags) const {
3600 // More accurate safety checking is done in getOutliningCandidateInfo.
3602}
3603
3604// Enum values indicating how an outlined call should be constructed.
3609
3614
3616 const MachineFunction *MF = MBB.getParent();
3617 const Function &F = MF->getFunction();
3618 return F.getFnAttribute("fentry-call").getValueAsBool() ||
3619 F.hasFnAttribute("patchable-function-entry");
3620}
3621
3623 MCRegister RegNo) {
3624 return MI.readsRegister(RegNo, TRI) ||
3625 MI.getDesc().hasImplicitUseOfPhysReg(RegNo);
3626}
3627
3629 const TargetRegisterInfo *TRI, MCRegister RegNo) {
3630 return MI.modifiesRegister(RegNo, TRI) ||
3631 MI.getDesc().hasImplicitDefOfPhysReg(RegNo);
3632}
3633
3635 if (!MBB.back().isReturn())
3636 return true;
3638 return true;
3639
3640 // If the candidate reads the pre-set register
3641 // that can be used for expanding PseudoTAIL instruction,
3642 // then we cannot insert tail call.
3643 const TargetSubtargetInfo &STI = MBB.getParent()->getSubtarget();
3644 MCRegister TailExpandUseRegNo =
3646 for (const MachineInstr &MI : MBB) {
3647 if (isMIReadsReg(MI, STI.getRegisterInfo(), TailExpandUseRegNo))
3648 return true;
3649 if (isMIModifiesReg(MI, STI.getRegisterInfo(), TailExpandUseRegNo))
3650 break;
3651 }
3652 return false;
3653}
3654
3656 // If last instruction is return then we can rely on
3657 // the verification already performed in the getOutliningTypeImpl.
3658 if (C.back().isReturn()) {
3659 assert(!cannotInsertTailCall(*C.getMBB()) &&
3660 "The candidate who uses return instruction must be outlined "
3661 "using tail call");
3662 return false;
3663 }
3664
3665 // Filter out candidates where the X5 register (t0) can't be used to setup
3666 // the function call.
3667 const TargetRegisterInfo *TRI = C.getMF()->getSubtarget().getRegisterInfo();
3668 if (llvm::any_of(C, [TRI](const MachineInstr &MI) {
3669 return isMIModifiesReg(MI, TRI, RISCV::X5);
3670 }))
3671 return true;
3672
3673 return !C.isAvailableAcrossAndOutOfSeq(RISCV::X5, *TRI);
3674}
3675
3676std::optional<std::unique_ptr<outliner::OutlinedFunction>>
3678 const MachineModuleInfo &MMI,
3679 std::vector<outliner::Candidate> &RepeatedSequenceLocs,
3680 unsigned MinRepeats) const {
3681
3682 // Analyze each candidate and erase the ones that are not viable.
3683 llvm::erase_if(RepeatedSequenceLocs, analyzeCandidate);
3684
3685 // If the sequence doesn't have enough candidates left, then we're done.
3686 if (RepeatedSequenceLocs.size() < MinRepeats)
3687 return std::nullopt;
3688
3689 // Each RepeatedSequenceLoc is identical.
3690 outliner::Candidate &Candidate = RepeatedSequenceLocs[0];
3691 unsigned InstrSizeCExt =
3692 Candidate.getMF()->getSubtarget<RISCVSubtarget>().hasStdExtZca() ? 2 : 4;
3693 unsigned CallOverhead = 0, FrameOverhead = 0;
3694
3695 // Count the number of CFI instructions in the candidate, if present.
3696 unsigned CFICount = 0;
3697 for (auto &I : Candidate) {
3698 if (I.isCFIInstruction())
3699 CFICount++;
3700 }
3701
3702 // Ensure CFI coverage matches: comparing the number of CFIs in the candidate
3703 // with the total number of CFIs in the parent function for each candidate.
3704 // Outlining only a subset of a function’s CFIs would split the unwind state
3705 // across two code regions and lead to incorrect address offsets between the
3706 // outlined body and the remaining code. To preserve correct unwind info, we
3707 // only outline when all CFIs in the function can be outlined together.
3708 for (outliner::Candidate &C : RepeatedSequenceLocs) {
3709 std::vector<MCCFIInstruction> CFIInstructions =
3710 C.getMF()->getFrameInstructions();
3711
3712 if (CFICount > 0 && CFICount != CFIInstructions.size())
3713 return std::nullopt;
3714 }
3715
3717 if (Candidate.back().isReturn()) {
3719 // tail call = auipc + jalr in the worst case without linker relaxation.
3720 // FIXME: This code suggests the JALR can be compressed - how?
3721 CallOverhead = 4 + InstrSizeCExt;
3722 // Using tail call we move ret instruction from caller to callee.
3723 FrameOverhead = 0;
3724 } else {
3725 // call t0, function = 8 bytes.
3726 CallOverhead = 8;
3727 // jr t0 = 4 bytes, 2 bytes if compressed instructions are enabled.
3728 FrameOverhead = InstrSizeCExt;
3729 }
3730
3731 // If we have CFI instructions, we can only outline if the outlined section
3732 // can be a tail call.
3733 if (MOCI != MachineOutlinerTailCall && CFICount > 0)
3734 return std::nullopt;
3735
3736 for (auto &C : RepeatedSequenceLocs)
3737 C.setCallInfo(MOCI, CallOverhead);
3738
3739 unsigned SequenceSize = 0;
3740 for (auto &MI : Candidate)
3741 SequenceSize += getInstSizeInBytes(MI);
3742
3743 return std::make_unique<outliner::OutlinedFunction>(
3744 RepeatedSequenceLocs, SequenceSize, FrameOverhead, MOCI);
3745}
3746
3750 unsigned Flags) const {
3751 MachineInstr &MI = *MBBI;
3752 MachineBasicBlock *MBB = MI.getParent();
3753 const TargetRegisterInfo *TRI =
3754 MBB->getParent()->getSubtarget().getRegisterInfo();
3755 const auto &F = MI.getMF()->getFunction();
3756
3757 // We can only outline CFI instructions if we will tail call the outlined
3758 // function, or fix up the CFI offsets. Currently, CFI instructions are
3759 // outlined only if in a tail call.
3760 if (MI.isCFIInstruction())
3762
3763 if (cannotInsertTailCall(*MBB) &&
3764 (MI.isReturn() || isMIModifiesReg(MI, TRI, RISCV::X5)))
3766
3767 // Make sure the operands don't reference something unsafe.
3768 for (const auto &MO : MI.operands()) {
3769
3770 // pcrel-hi and pcrel-lo can't put in separate sections, filter that out
3771 // if any possible.
3772 if (MO.getTargetFlags() == RISCVII::MO_PCREL_LO &&
3773 (MI.getMF()->getTarget().getFunctionSections() || F.hasComdat() ||
3774 F.hasSection() || F.getSectionPrefix()))
3776 }
3777
3778 if (isLPAD(MI))
3780
3782}
3783
3786 const outliner::OutlinedFunction &OF) const {
3787
3788 if (OF.FrameConstructionID == MachineOutlinerTailCall)
3789 return;
3790
3791 MBB.addLiveIn(RISCV::X5);
3792
3793 // Add in a return instruction to the end of the outlined frame.
3794 MBB.insert(MBB.end(), BuildMI(MF, DebugLoc(), get(RISCV::JALR))
3795 .addReg(RISCV::X0, RegState::Define)
3796 .addReg(RISCV::X5)
3797 .addImm(0));
3798}
3799
3803
3804 if (C.CallConstructionID == MachineOutlinerTailCall) {
3805 It = MBB.insert(It, BuildMI(MF, DebugLoc(), get(RISCV::PseudoTAIL))
3806 .addGlobalAddress(M.getNamedValue(MF.getName()),
3807 /*Offset=*/0, RISCVII::MO_CALL));
3808 return It;
3809 }
3810
3811 // Add in a call instruction to the outlined function at the given location.
3812 It = MBB.insert(It,
3813 BuildMI(MF, DebugLoc(), get(RISCV::PseudoCALLReg), RISCV::X5)
3814 .addGlobalAddress(M.getNamedValue(MF.getName()), 0,
3816 return It;
3817}
3818
3819std::optional<RegImmPair> RISCVInstrInfo::isAddImmediate(const MachineInstr &MI,
3820 Register Reg) const {
3821 // TODO: Handle cases where Reg is a super- or sub-register of the
3822 // destination register.
3823 const MachineOperand &Op0 = MI.getOperand(0);
3824 if (!Op0.isReg() || Reg != Op0.getReg())
3825 return std::nullopt;
3826
3827 // Don't consider ADDIW as a candidate because the caller may not be aware
3828 // of its sign extension behaviour.
3829 if (MI.getOpcode() == RISCV::ADDI && MI.getOperand(1).isReg() &&
3830 MI.getOperand(2).isImm())
3831 return RegImmPair{MI.getOperand(1).getReg(), MI.getOperand(2).getImm()};
3832
3833 return std::nullopt;
3834}
3835
3836// MIR printer helper function to annotate Operands with a comment.
3838 const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx,
3839 const TargetRegisterInfo *TRI) const {
3840 // Print a generic comment for this operand if there is one.
3841 std::string GenericComment =
3843 if (!GenericComment.empty())
3844 return GenericComment;
3845
3846 // If not, we must have an immediate operand.
3847 if (!Op.isImm())
3848 return std::string();
3849
3850 const MCInstrDesc &Desc = MI.getDesc();
3851 if (OpIdx >= Desc.getNumOperands())
3852 return std::string();
3853
3854 std::string Comment;
3855 raw_string_ostream OS(Comment);
3856
3857 const MCOperandInfo &OpInfo = Desc.operands()[OpIdx];
3858
3859 // Print the full VType operand of vsetvli/vsetivli instructions, and the SEW
3860 // operand of vector codegen pseudos.
3861 switch (OpInfo.OperandType) {
3864 unsigned Imm = Op.getImm();
3865 RISCVVType::printVType(Imm, OS);
3866 break;
3867 }
3869 unsigned Imm = Op.getImm();
3871 break;
3872 }
3874 unsigned Imm = Op.getImm();
3875 OS << "w" << Imm;
3876 break;
3877 }
3880 unsigned Log2SEW = Op.getImm();
3881 unsigned SEW = Log2SEW ? 1 << Log2SEW : 8;
3882 assert(RISCVVType::isValidSEW(SEW) && "Unexpected SEW");
3883 OS << "e" << SEW;
3884 break;
3885 }
3887 unsigned Policy = Op.getImm();
3889 "Invalid Policy Value");
3890 OS << (Policy & RISCVVType::TAIL_AGNOSTIC ? "ta" : "tu") << ", "
3891 << (Policy & RISCVVType::MASK_AGNOSTIC ? "ma" : "mu");
3892 break;
3893 }
3894
3895 return Comment;
3896}
3897
3898// clang-format off
3899#define CASE_RVV_OPCODE_UNMASK_LMUL(OP, LMUL) \
3900 RISCV::Pseudo##OP##_##LMUL
3901
3902#define CASE_RVV_OPCODE_MASK_LMUL(OP, LMUL) \
3903 RISCV::Pseudo##OP##_##LMUL##_MASK
3904
3905#define CASE_RVV_OPCODE_LMUL(OP, LMUL) \
3906 CASE_RVV_OPCODE_UNMASK_LMUL(OP, LMUL): \
3907 case CASE_RVV_OPCODE_MASK_LMUL(OP, LMUL)
3908
3909#define CASE_RVV_OPCODE_UNMASK_WIDEN(OP) \
3910 CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF8): \
3911 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF4): \
3912 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, MF2): \
3913 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M1): \
3914 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M2): \
3915 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M4)
3916
3917#define CASE_RVV_OPCODE_UNMASK(OP) \
3918 CASE_RVV_OPCODE_UNMASK_WIDEN(OP): \
3919 case CASE_RVV_OPCODE_UNMASK_LMUL(OP, M8)
3920
3921#define CASE_RVV_OPCODE_MASK_WIDEN(OP) \
3922 CASE_RVV_OPCODE_MASK_LMUL(OP, MF8): \
3923 case CASE_RVV_OPCODE_MASK_LMUL(OP, MF4): \
3924 case CASE_RVV_OPCODE_MASK_LMUL(OP, MF2): \
3925 case CASE_RVV_OPCODE_MASK_LMUL(OP, M1): \
3926 case CASE_RVV_OPCODE_MASK_LMUL(OP, M2): \
3927 case CASE_RVV_OPCODE_MASK_LMUL(OP, M4)
3928
3929#define CASE_RVV_OPCODE_MASK(OP) \
3930 CASE_RVV_OPCODE_MASK_WIDEN(OP): \
3931 case CASE_RVV_OPCODE_MASK_LMUL(OP, M8)
3932
3933#define CASE_RVV_OPCODE_WIDEN(OP) \
3934 CASE_RVV_OPCODE_UNMASK_WIDEN(OP): \
3935 case CASE_RVV_OPCODE_MASK_WIDEN(OP)
3936
3937#define CASE_RVV_OPCODE(OP) \
3938 CASE_RVV_OPCODE_UNMASK(OP): \
3939 case CASE_RVV_OPCODE_MASK(OP)
3940// clang-format on
3941
3942// clang-format off
3943#define CASE_VMA_OPCODE_COMMON(OP, TYPE, LMUL) \
3944 RISCV::PseudoV##OP##_##TYPE##_##LMUL
3945
3946#define CASE_VMA_OPCODE_LMULS(OP, TYPE) \
3947 CASE_VMA_OPCODE_COMMON(OP, TYPE, MF8): \
3948 case CASE_VMA_OPCODE_COMMON(OP, TYPE, MF4): \
3949 case CASE_VMA_OPCODE_COMMON(OP, TYPE, MF2): \
3950 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M1): \
3951 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M2): \
3952 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M4): \
3953 case CASE_VMA_OPCODE_COMMON(OP, TYPE, M8)
3954
3955// VFMA instructions are SEW specific.
3956#define CASE_VFMA_OPCODE_COMMON(OP, TYPE, LMUL, SEW) \
3957 RISCV::PseudoV##OP##_##TYPE##_##LMUL##_##SEW
3958
3959#define CASE_VFMA_OPCODE_LMULS_M1(OP, TYPE, SEW) \
3960 CASE_VFMA_OPCODE_COMMON(OP, TYPE, M1, SEW): \
3961 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M2, SEW): \
3962 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M4, SEW): \
3963 case CASE_VFMA_OPCODE_COMMON(OP, TYPE, M8, SEW)
3964
3965#define CASE_VFMA_OPCODE_LMULS_MF2(OP, TYPE, SEW) \
3966 CASE_VFMA_OPCODE_COMMON(OP, TYPE, MF2, SEW): \
3967 case CASE_VFMA_OPCODE_LMULS_M1(OP, TYPE, SEW)
3968
3969#define CASE_VFMA_OPCODE_LMULS_MF4(OP, TYPE, SEW) \
3970 CASE_VFMA_OPCODE_COMMON(OP, TYPE, MF4, SEW): \
3971 case CASE_VFMA_OPCODE_LMULS_MF2(OP, TYPE, SEW)
3972
3973#define CASE_VFMA_OPCODE_VV(OP) \
3974 CASE_VFMA_OPCODE_LMULS_MF4(OP, VV, E16): \
3975 case CASE_VFMA_OPCODE_LMULS_MF4(OP##_ALT, VV, E16): \
3976 case CASE_VFMA_OPCODE_LMULS_MF2(OP, VV, E32): \
3977 case CASE_VFMA_OPCODE_LMULS_M1(OP, VV, E64)
3978
3979#define CASE_VFMA_SPLATS(OP) \
3980 CASE_VFMA_OPCODE_LMULS_MF4(OP, VFPR16, E16): \
3981 case CASE_VFMA_OPCODE_LMULS_MF4(OP##_ALT, VFPR16, E16): \
3982 case CASE_VFMA_OPCODE_LMULS_MF2(OP, VFPR32, E32): \
3983 case CASE_VFMA_OPCODE_LMULS_M1(OP, VFPR64, E64)
3984// clang-format on
3985
3987 unsigned &SrcOpIdx1,
3988 unsigned &SrcOpIdx2) const {
3989 const MCInstrDesc &Desc = MI.getDesc();
3990 if (!Desc.isCommutable())
3991 return false;
3992
3993 switch (MI.getOpcode()) {
3994 case RISCV::TH_MVEQZ:
3995 case RISCV::TH_MVNEZ:
3996 // We can't commute operands if operand 2 (i.e., rs1 in
3997 // mveqz/mvnez rd,rs1,rs2) is the zero-register (as it is
3998 // not valid as the in/out-operand 1).
3999 if (MI.getOperand(2).getReg() == RISCV::X0)
4000 return false;
4001 // Operands 1 and 2 are commutable, if we switch the opcode.
4002 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 2);
4003 case RISCV::QC_SELECTIEQ:
4004 case RISCV::QC_SELECTINE:
4005 case RISCV::QC_SELECTIIEQ:
4006 case RISCV::QC_SELECTIINE:
4007 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 2);
4008 case RISCV::QC_MVEQ:
4009 case RISCV::QC_MVNE:
4010 case RISCV::QC_MVLT:
4011 case RISCV::QC_MVGE:
4012 case RISCV::QC_MVLTU:
4013 case RISCV::QC_MVGEU:
4014 case RISCV::QC_MVEQI:
4015 case RISCV::QC_MVNEI:
4016 case RISCV::QC_MVLTI:
4017 case RISCV::QC_MVGEI:
4018 case RISCV::QC_MVLTUI:
4019 case RISCV::QC_MVGEUI:
4020 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 1, 4);
4021 case RISCV::TH_MULA:
4022 case RISCV::TH_MULAW:
4023 case RISCV::TH_MULAH:
4024 case RISCV::TH_MULS:
4025 case RISCV::TH_MULSW:
4026 case RISCV::TH_MULSH:
4027 // Operands 2 and 3 are commutable.
4028 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 2, 3);
4029 case RISCV::PseudoCCMOVGPRNoX0:
4030 case RISCV::PseudoCCMOVGPR:
4031 // Operands 4 and 5 are commutable.
4032 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 4, 5);
4033 case CASE_RVV_OPCODE(VADD_VV):
4034 case CASE_RVV_OPCODE(VAND_VV):
4035 case CASE_RVV_OPCODE(VOR_VV):
4036 case CASE_RVV_OPCODE(VXOR_VV):
4037 case CASE_RVV_OPCODE_MASK(VMSEQ_VV):
4038 case CASE_RVV_OPCODE_MASK(VMSNE_VV):
4039 case CASE_RVV_OPCODE(VMIN_VV):
4040 case CASE_RVV_OPCODE(VMINU_VV):
4041 case CASE_RVV_OPCODE(VMAX_VV):
4042 case CASE_RVV_OPCODE(VMAXU_VV):
4043 case CASE_RVV_OPCODE(VMUL_VV):
4044 case CASE_RVV_OPCODE(VMULH_VV):
4045 case CASE_RVV_OPCODE(VMULHU_VV):
4046 case CASE_RVV_OPCODE_WIDEN(VWADD_VV):
4047 case CASE_RVV_OPCODE_WIDEN(VWADDU_VV):
4048 case CASE_RVV_OPCODE_WIDEN(VWMUL_VV):
4049 case CASE_RVV_OPCODE_WIDEN(VWMULU_VV):
4050 case CASE_RVV_OPCODE_WIDEN(VWMACC_VV):
4051 case CASE_RVV_OPCODE_WIDEN(VWMACCU_VV):
4052 case CASE_RVV_OPCODE_UNMASK(VADC_VVM):
4053 case CASE_RVV_OPCODE(VSADD_VV):
4054 case CASE_RVV_OPCODE(VSADDU_VV):
4055 case CASE_RVV_OPCODE(VAADD_VV):
4056 case CASE_RVV_OPCODE(VAADDU_VV):
4057 case CASE_RVV_OPCODE(VSMUL_VV):
4058 // Operands 2 and 3 are commutable.
4059 return fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, 2, 3);
4060 case CASE_VFMA_SPLATS(FMADD):
4061 case CASE_VFMA_SPLATS(FMSUB):
4062 case CASE_VFMA_SPLATS(FMACC):
4063 case CASE_VFMA_SPLATS(FMSAC):
4066 case CASE_VFMA_SPLATS(FNMACC):
4067 case CASE_VFMA_SPLATS(FNMSAC):
4068 case CASE_VFMA_OPCODE_VV(FMACC):
4069 case CASE_VFMA_OPCODE_VV(FMSAC):
4070 case CASE_VFMA_OPCODE_VV(FNMACC):
4071 case CASE_VFMA_OPCODE_VV(FNMSAC):
4072 case CASE_VMA_OPCODE_LMULS(MADD, VX):
4073 case CASE_VMA_OPCODE_LMULS(NMSUB, VX):
4074 case CASE_VMA_OPCODE_LMULS(MACC, VX):
4075 case CASE_VMA_OPCODE_LMULS(NMSAC, VX):
4076 case CASE_VMA_OPCODE_LMULS(MACC, VV):
4077 case CASE_VMA_OPCODE_LMULS(NMSAC, VV): {
4078 // If the tail policy is undisturbed we can't commute.
4079 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags));
4080 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4081 1) == 0)
4082 return false;
4083
4084 // For these instructions we can only swap operand 1 and operand 3 by
4085 // changing the opcode.
4086 unsigned CommutableOpIdx1 = 1;
4087 unsigned CommutableOpIdx2 = 3;
4088 if (!fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, CommutableOpIdx1,
4089 CommutableOpIdx2))
4090 return false;
4091 return true;
4092 }
4093 case CASE_VFMA_OPCODE_VV(FMADD):
4097 case CASE_VMA_OPCODE_LMULS(MADD, VV):
4098 case CASE_VMA_OPCODE_LMULS(NMSUB, VV): {
4099 // If the tail policy is undisturbed we can't commute.
4100 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags));
4101 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4102 1) == 0)
4103 return false;
4104
4105 // For these instructions we have more freedom. We can commute with the
4106 // other multiplicand or with the addend/subtrahend/minuend.
4107
4108 // Any fixed operand must be from source 1, 2 or 3.
4109 if (SrcOpIdx1 != CommuteAnyOperandIndex && SrcOpIdx1 > 3)
4110 return false;
4111 if (SrcOpIdx2 != CommuteAnyOperandIndex && SrcOpIdx2 > 3)
4112 return false;
4113
4114 // It both ops are fixed one must be the tied source.
4115 if (SrcOpIdx1 != CommuteAnyOperandIndex &&
4116 SrcOpIdx2 != CommuteAnyOperandIndex && SrcOpIdx1 != 1 && SrcOpIdx2 != 1)
4117 return false;
4118
4119 // Look for two different register operands assumed to be commutable
4120 // regardless of the FMA opcode. The FMA opcode is adjusted later if
4121 // needed.
4122 if (SrcOpIdx1 == CommuteAnyOperandIndex ||
4123 SrcOpIdx2 == CommuteAnyOperandIndex) {
4124 // At least one of operands to be commuted is not specified and
4125 // this method is free to choose appropriate commutable operands.
4126 unsigned CommutableOpIdx1 = SrcOpIdx1;
4127 if (SrcOpIdx1 == SrcOpIdx2) {
4128 // Both of operands are not fixed. Set one of commutable
4129 // operands to the tied source.
4130 CommutableOpIdx1 = 1;
4131 } else if (SrcOpIdx1 == CommuteAnyOperandIndex) {
4132 // Only one of the operands is not fixed.
4133 CommutableOpIdx1 = SrcOpIdx2;
4134 }
4135
4136 // CommutableOpIdx1 is well defined now. Let's choose another commutable
4137 // operand and assign its index to CommutableOpIdx2.
4138 unsigned CommutableOpIdx2;
4139 if (CommutableOpIdx1 != 1) {
4140 // If we haven't already used the tied source, we must use it now.
4141 CommutableOpIdx2 = 1;
4142 } else {
4143 Register Op1Reg = MI.getOperand(CommutableOpIdx1).getReg();
4144
4145 // The commuted operands should have different registers.
4146 // Otherwise, the commute transformation does not change anything and
4147 // is useless. We use this as a hint to make our decision.
4148 if (Op1Reg != MI.getOperand(2).getReg())
4149 CommutableOpIdx2 = 2;
4150 else
4151 CommutableOpIdx2 = 3;
4152 }
4153
4154 // Assign the found pair of commutable indices to SrcOpIdx1 and
4155 // SrcOpIdx2 to return those values.
4156 if (!fixCommutedOpIndices(SrcOpIdx1, SrcOpIdx2, CommutableOpIdx1,
4157 CommutableOpIdx2))
4158 return false;
4159 }
4160
4161 return true;
4162 }
4163 }
4164
4165 return TargetInstrInfo::findCommutedOpIndices(MI, SrcOpIdx1, SrcOpIdx2);
4166}
4167
4168// clang-format off
4169#define CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, LMUL) \
4170 case RISCV::PseudoV##OLDOP##_##TYPE##_##LMUL: \
4171 Opc = RISCV::PseudoV##NEWOP##_##TYPE##_##LMUL; \
4172 break;
4173
4174#define CASE_VMA_CHANGE_OPCODE_LMULS(OLDOP, NEWOP, TYPE) \
4175 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF8) \
4176 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF4) \
4177 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF2) \
4178 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M1) \
4179 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M2) \
4180 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M4) \
4181 CASE_VMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M8)
4182
4183// VFMA depends on SEW.
4184#define CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, LMUL, SEW) \
4185 case RISCV::PseudoV##OLDOP##_##TYPE##_##LMUL##_##SEW: \
4186 Opc = RISCV::PseudoV##NEWOP##_##TYPE##_##LMUL##_##SEW; \
4187 break;
4188
4189#define CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, TYPE, SEW) \
4190 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M1, SEW) \
4191 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M2, SEW) \
4192 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M4, SEW) \
4193 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, M8, SEW)
4194
4195#define CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, TYPE, SEW) \
4196 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF2, SEW) \
4197 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, TYPE, SEW)
4198
4199#define CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, TYPE, SEW) \
4200 CASE_VFMA_CHANGE_OPCODE_COMMON(OLDOP, NEWOP, TYPE, MF4, SEW) \
4201 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, TYPE, SEW)
4202
4203#define CASE_VFMA_CHANGE_OPCODE_VV(OLDOP, NEWOP) \
4204 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, VV, E16) \
4205 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP##_ALT, NEWOP##_ALT, VV, E16) \
4206 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, VV, E32) \
4207 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, VV, E64)
4208
4209#define CASE_VFMA_CHANGE_OPCODE_SPLATS(OLDOP, NEWOP) \
4210 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP, NEWOP, VFPR16, E16) \
4211 CASE_VFMA_CHANGE_OPCODE_LMULS_MF4(OLDOP##_ALT, NEWOP##_ALT, VFPR16, E16) \
4212 CASE_VFMA_CHANGE_OPCODE_LMULS_MF2(OLDOP, NEWOP, VFPR32, E32) \
4213 CASE_VFMA_CHANGE_OPCODE_LMULS_M1(OLDOP, NEWOP, VFPR64, E64)
4214// clang-format on
4215
4217 bool NewMI,
4218 unsigned OpIdx1,
4219 unsigned OpIdx2) const {
4220 auto cloneIfNew = [NewMI](MachineInstr &MI) -> MachineInstr & {
4221 if (NewMI)
4222 return *MI.getParent()->getParent()->CloneMachineInstr(&MI);
4223 return MI;
4224 };
4225
4226 switch (MI.getOpcode()) {
4227 case RISCV::TH_MVEQZ:
4228 case RISCV::TH_MVNEZ: {
4229 auto &WorkingMI = cloneIfNew(MI);
4230 WorkingMI.setDesc(get(MI.getOpcode() == RISCV::TH_MVEQZ ? RISCV::TH_MVNEZ
4231 : RISCV::TH_MVEQZ));
4232 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, false, OpIdx1,
4233 OpIdx2);
4234 }
4235 case RISCV::QC_SELECTIEQ:
4236 case RISCV::QC_SELECTINE:
4237 case RISCV::QC_SELECTIIEQ:
4238 case RISCV::QC_SELECTIINE:
4239 return TargetInstrInfo::commuteInstructionImpl(MI, NewMI, OpIdx1, OpIdx2);
4240 case RISCV::QC_MVEQ:
4241 case RISCV::QC_MVNE:
4242 case RISCV::QC_MVLT:
4243 case RISCV::QC_MVGE:
4244 case RISCV::QC_MVLTU:
4245 case RISCV::QC_MVGEU:
4246 case RISCV::QC_MVEQI:
4247 case RISCV::QC_MVNEI:
4248 case RISCV::QC_MVLTI:
4249 case RISCV::QC_MVGEI:
4250 case RISCV::QC_MVLTUI:
4251 case RISCV::QC_MVGEUI: {
4252 auto &WorkingMI = cloneIfNew(MI);
4253 WorkingMI.setDesc(get(getInverseXqcicmOpcode(MI.getOpcode())));
4254 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, false, OpIdx1,
4255 OpIdx2);
4256 }
4257 case RISCV::PseudoCCMOVGPRNoX0:
4258 case RISCV::PseudoCCMOVGPR: {
4259 // CCMOV can be commuted by inverting the condition.
4260 auto CC = static_cast<RISCVCC::CondCode>(MI.getOperand(3).getImm());
4262 auto &WorkingMI = cloneIfNew(MI);
4263 WorkingMI.getOperand(3).setImm(CC);
4264 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, /*NewMI*/ false,
4265 OpIdx1, OpIdx2);
4266 }
4267 case CASE_VFMA_SPLATS(FMACC):
4268 case CASE_VFMA_SPLATS(FMADD):
4269 case CASE_VFMA_SPLATS(FMSAC):
4270 case CASE_VFMA_SPLATS(FMSUB):
4271 case CASE_VFMA_SPLATS(FNMACC):
4273 case CASE_VFMA_SPLATS(FNMSAC):
4275 case CASE_VFMA_OPCODE_VV(FMACC):
4276 case CASE_VFMA_OPCODE_VV(FMSAC):
4277 case CASE_VFMA_OPCODE_VV(FNMACC):
4278 case CASE_VFMA_OPCODE_VV(FNMSAC):
4279 case CASE_VMA_OPCODE_LMULS(MADD, VX):
4280 case CASE_VMA_OPCODE_LMULS(NMSUB, VX):
4281 case CASE_VMA_OPCODE_LMULS(MACC, VX):
4282 case CASE_VMA_OPCODE_LMULS(NMSAC, VX):
4283 case CASE_VMA_OPCODE_LMULS(MACC, VV):
4284 case CASE_VMA_OPCODE_LMULS(NMSAC, VV): {
4285 // It only make sense to toggle these between clobbering the
4286 // addend/subtrahend/minuend one of the multiplicands.
4287 assert((OpIdx1 == 1 || OpIdx2 == 1) && "Unexpected opcode index");
4288 assert((OpIdx1 == 3 || OpIdx2 == 3) && "Unexpected opcode index");
4289 unsigned Opc;
4290 switch (MI.getOpcode()) {
4291 default:
4292 llvm_unreachable("Unexpected opcode");
4293 CASE_VFMA_CHANGE_OPCODE_SPLATS(FMACC, FMADD)
4294 CASE_VFMA_CHANGE_OPCODE_SPLATS(FMADD, FMACC)
4301 CASE_VFMA_CHANGE_OPCODE_VV(FMACC, FMADD)
4305 CASE_VMA_CHANGE_OPCODE_LMULS(MACC, MADD, VX)
4306 CASE_VMA_CHANGE_OPCODE_LMULS(MADD, MACC, VX)
4307 CASE_VMA_CHANGE_OPCODE_LMULS(NMSAC, NMSUB, VX)
4308 CASE_VMA_CHANGE_OPCODE_LMULS(NMSUB, NMSAC, VX)
4309 CASE_VMA_CHANGE_OPCODE_LMULS(MACC, MADD, VV)
4310 CASE_VMA_CHANGE_OPCODE_LMULS(NMSAC, NMSUB, VV)
4311 }
4312
4313 auto &WorkingMI = cloneIfNew(MI);
4314 WorkingMI.setDesc(get(Opc));
4315 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, /*NewMI=*/false,
4316 OpIdx1, OpIdx2);
4317 }
4318 case CASE_VFMA_OPCODE_VV(FMADD):
4322 case CASE_VMA_OPCODE_LMULS(MADD, VV):
4323 case CASE_VMA_OPCODE_LMULS(NMSUB, VV): {
4324 assert((OpIdx1 == 1 || OpIdx2 == 1) && "Unexpected opcode index");
4325 // If one of the operands, is the addend we need to change opcode.
4326 // Otherwise we're just swapping 2 of the multiplicands.
4327 if (OpIdx1 == 3 || OpIdx2 == 3) {
4328 unsigned Opc;
4329 switch (MI.getOpcode()) {
4330 default:
4331 llvm_unreachable("Unexpected opcode");
4332 CASE_VFMA_CHANGE_OPCODE_VV(FMADD, FMACC)
4336 CASE_VMA_CHANGE_OPCODE_LMULS(MADD, MACC, VV)
4337 CASE_VMA_CHANGE_OPCODE_LMULS(NMSUB, NMSAC, VV)
4338 }
4339
4340 auto &WorkingMI = cloneIfNew(MI);
4341 WorkingMI.setDesc(get(Opc));
4342 return TargetInstrInfo::commuteInstructionImpl(WorkingMI, /*NewMI=*/false,
4343 OpIdx1, OpIdx2);
4344 }
4345 // Let the default code handle it.
4346 break;
4347 }
4348 }
4349
4350 return TargetInstrInfo::commuteInstructionImpl(MI, NewMI, OpIdx1, OpIdx2);
4351}
4352
4353#undef CASE_VMA_CHANGE_OPCODE_COMMON
4354#undef CASE_VMA_CHANGE_OPCODE_LMULS
4355#undef CASE_VFMA_CHANGE_OPCODE_COMMON
4356#undef CASE_VFMA_CHANGE_OPCODE_LMULS_M1
4357#undef CASE_VFMA_CHANGE_OPCODE_LMULS_MF2
4358#undef CASE_VFMA_CHANGE_OPCODE_LMULS_MF4
4359#undef CASE_VFMA_CHANGE_OPCODE_VV
4360#undef CASE_VFMA_CHANGE_OPCODE_SPLATS
4361
4362#undef CASE_RVV_OPCODE_UNMASK_LMUL
4363#undef CASE_RVV_OPCODE_MASK_LMUL
4364#undef CASE_RVV_OPCODE_LMUL
4365#undef CASE_RVV_OPCODE_UNMASK_WIDEN
4366#undef CASE_RVV_OPCODE_UNMASK
4367#undef CASE_RVV_OPCODE_MASK_WIDEN
4368#undef CASE_RVV_OPCODE_MASK
4369#undef CASE_RVV_OPCODE_WIDEN
4370#undef CASE_RVV_OPCODE
4371
4372#undef CASE_VMA_OPCODE_COMMON
4373#undef CASE_VMA_OPCODE_LMULS
4374#undef CASE_VFMA_OPCODE_COMMON
4375#undef CASE_VFMA_OPCODE_LMULS_M1
4376#undef CASE_VFMA_OPCODE_LMULS_MF2
4377#undef CASE_VFMA_OPCODE_LMULS_MF4
4378#undef CASE_VFMA_OPCODE_VV
4379#undef CASE_VFMA_SPLATS
4380
4382 switch (MI.getOpcode()) {
4383 default:
4384 break;
4385 case RISCV::ADD:
4386 case RISCV::OR:
4387 case RISCV::XOR:
4388 // Normalize (so we hit the next if clause).
4389 // add/[x]or rd, zero, rs => add/[x]or rd, rs, zero
4390 if (MI.getOperand(1).getReg() == RISCV::X0)
4391 commuteInstruction(MI);
4392 // add/[x]or rd, rs, zero => addi rd, rs, 0
4393 if (MI.getOperand(2).getReg() == RISCV::X0) {
4394 MI.getOperand(2).ChangeToImmediate(0);
4395 MI.setDesc(get(RISCV::ADDI));
4396 return true;
4397 }
4398 // xor rd, rs, rs => addi rd, zero, 0
4399 if (MI.getOpcode() == RISCV::XOR &&
4400 MI.getOperand(1).getReg() == MI.getOperand(2).getReg()) {
4401 MI.getOperand(1).setReg(RISCV::X0);
4402 MI.getOperand(2).ChangeToImmediate(0);
4403 MI.setDesc(get(RISCV::ADDI));
4404 return true;
4405 }
4406 break;
4407 case RISCV::ORI:
4408 case RISCV::XORI:
4409 // [x]ori rd, zero, N => addi rd, zero, N
4410 if (MI.getOperand(1).getReg() == RISCV::X0) {
4411 MI.setDesc(get(RISCV::ADDI));
4412 return true;
4413 }
4414 break;
4415 case RISCV::SUB:
4416 // sub rd, rs, zero => addi rd, rs, 0
4417 if (MI.getOperand(2).getReg() == RISCV::X0) {
4418 MI.getOperand(2).ChangeToImmediate(0);
4419 MI.setDesc(get(RISCV::ADDI));
4420 return true;
4421 }
4422 break;
4423 case RISCV::SUBW:
4424 // subw rd, rs, zero => addiw rd, rs, 0
4425 if (MI.getOperand(2).getReg() == RISCV::X0) {
4426 MI.getOperand(2).ChangeToImmediate(0);
4427 MI.setDesc(get(RISCV::ADDIW));
4428 return true;
4429 }
4430 break;
4431 case RISCV::ADDW:
4432 // Normalize (so we hit the next if clause).
4433 // addw rd, zero, rs => addw rd, rs, zero
4434 if (MI.getOperand(1).getReg() == RISCV::X0)
4435 commuteInstruction(MI);
4436 // addw rd, rs, zero => addiw rd, rs, 0
4437 if (MI.getOperand(2).getReg() == RISCV::X0) {
4438 MI.getOperand(2).ChangeToImmediate(0);
4439 MI.setDesc(get(RISCV::ADDIW));
4440 return true;
4441 }
4442 break;
4443 case RISCV::SH1ADD:
4444 case RISCV::SH1ADD_UW:
4445 case RISCV::SH2ADD:
4446 case RISCV::SH2ADD_UW:
4447 case RISCV::SH3ADD:
4448 case RISCV::SH3ADD_UW:
4449 // shNadd[.uw] rd, zero, rs => addi rd, rs, 0
4450 if (MI.getOperand(1).getReg() == RISCV::X0) {
4451 MI.removeOperand(1);
4452 MI.addOperand(MachineOperand::CreateImm(0));
4453 MI.setDesc(get(RISCV::ADDI));
4454 return true;
4455 }
4456 // shNadd[.uw] rd, rs, zero => slli[.uw] rd, rs, N
4457 if (MI.getOperand(2).getReg() == RISCV::X0) {
4458 MI.removeOperand(2);
4459 unsigned Opc = MI.getOpcode();
4460 if (Opc == RISCV::SH1ADD_UW || Opc == RISCV::SH2ADD_UW ||
4461 Opc == RISCV::SH3ADD_UW) {
4463 MI.setDesc(get(RISCV::SLLI_UW));
4464 return true;
4465 }
4467 MI.setDesc(get(RISCV::SLLI));
4468 return true;
4469 }
4470 break;
4471 case RISCV::AND:
4472 case RISCV::MUL:
4473 case RISCV::MULH:
4474 case RISCV::MULHSU:
4475 case RISCV::MULHU:
4476 case RISCV::MULW:
4477 // and rd, zero, rs => addi rd, zero, 0
4478 // mul* rd, zero, rs => addi rd, zero, 0
4479 // and rd, rs, zero => addi rd, zero, 0
4480 // mul* rd, rs, zero => addi rd, zero, 0
4481 if (MI.getOperand(1).getReg() == RISCV::X0 ||
4482 MI.getOperand(2).getReg() == RISCV::X0) {
4483 MI.getOperand(1).setReg(RISCV::X0);
4484 MI.getOperand(2).ChangeToImmediate(0);
4485 MI.setDesc(get(RISCV::ADDI));
4486 return true;
4487 }
4488 break;
4489 case RISCV::ANDI:
4490 // andi rd, zero, C => addi rd, zero, 0
4491 if (MI.getOperand(1).getReg() == RISCV::X0) {
4492 MI.getOperand(2).setImm(0);
4493 MI.setDesc(get(RISCV::ADDI));
4494 return true;
4495 }
4496 break;
4497 case RISCV::SLL:
4498 case RISCV::SRL:
4499 case RISCV::SRA:
4500 // shift rd, zero, rs => addi rd, zero, 0
4501 if (MI.getOperand(1).getReg() == RISCV::X0) {
4502 MI.getOperand(2).ChangeToImmediate(0);
4503 MI.setDesc(get(RISCV::ADDI));
4504 return true;
4505 }
4506 // shift rd, rs, zero => addi rd, rs, 0
4507 if (MI.getOperand(2).getReg() == RISCV::X0) {
4508 MI.getOperand(2).ChangeToImmediate(0);
4509 MI.setDesc(get(RISCV::ADDI));
4510 return true;
4511 }
4512 break;
4513 case RISCV::SLLW:
4514 case RISCV::SRLW:
4515 case RISCV::SRAW:
4516 // shiftw rd, zero, rs => addi rd, zero, 0
4517 if (MI.getOperand(1).getReg() == RISCV::X0) {
4518 MI.getOperand(2).ChangeToImmediate(0);
4519 MI.setDesc(get(RISCV::ADDI));
4520 return true;
4521 }
4522 break;
4523 case RISCV::SLLI:
4524 case RISCV::SRLI:
4525 case RISCV::SRAI:
4526 case RISCV::SLLIW:
4527 case RISCV::SRLIW:
4528 case RISCV::SRAIW:
4529 case RISCV::SLLI_UW:
4530 // shiftimm rd, zero, N => addi rd, zero, 0
4531 if (MI.getOperand(1).getReg() == RISCV::X0) {
4532 MI.getOperand(2).setImm(0);
4533 MI.setDesc(get(RISCV::ADDI));
4534 return true;
4535 }
4536 break;
4537 case RISCV::SLTU:
4538 case RISCV::ADD_UW:
4539 // sltu rd, zero, zero => addi rd, zero, 0
4540 // add.uw rd, zero, zero => addi rd, zero, 0
4541 if (MI.getOperand(1).getReg() == RISCV::X0 &&
4542 MI.getOperand(2).getReg() == RISCV::X0) {
4543 MI.getOperand(2).ChangeToImmediate(0);
4544 MI.setDesc(get(RISCV::ADDI));
4545 return true;
4546 }
4547 // add.uw rd, zero, rs => addi rd, rs, 0
4548 if (MI.getOpcode() == RISCV::ADD_UW &&
4549 MI.getOperand(1).getReg() == RISCV::X0) {
4550 MI.removeOperand(1);
4551 MI.addOperand(MachineOperand::CreateImm(0));
4552 MI.setDesc(get(RISCV::ADDI));
4553 }
4554 break;
4555 case RISCV::SLTIU:
4556 // sltiu rd, zero, NZC => addi rd, zero, 1
4557 // sltiu rd, zero, 0 => addi rd, zero, 0
4558 if (MI.getOperand(1).getReg() == RISCV::X0) {
4559 MI.getOperand(2).setImm(MI.getOperand(2).getImm() != 0);
4560 MI.setDesc(get(RISCV::ADDI));
4561 return true;
4562 }
4563 break;
4564 case RISCV::SEXT_H:
4565 case RISCV::SEXT_B:
4566 case RISCV::ZEXT_H_RV32:
4567 case RISCV::ZEXT_H_RV64:
4568 // sext.[hb] rd, zero => addi rd, zero, 0
4569 // zext.h rd, zero => addi rd, zero, 0
4570 if (MI.getOperand(1).getReg() == RISCV::X0) {
4571 MI.addOperand(MachineOperand::CreateImm(0));
4572 MI.setDesc(get(RISCV::ADDI));
4573 return true;
4574 }
4575 break;
4576 case RISCV::MIN:
4577 case RISCV::MINU:
4578 case RISCV::MAX:
4579 case RISCV::MAXU:
4580 // min|max rd, rs, rs => addi rd, rs, 0
4581 if (MI.getOperand(1).getReg() == MI.getOperand(2).getReg()) {
4582 MI.getOperand(2).ChangeToImmediate(0);
4583 MI.setDesc(get(RISCV::ADDI));
4584 return true;
4585 }
4586 break;
4587 case RISCV::BEQ:
4588 case RISCV::BNE:
4589 // b{eq,ne} zero, rs, imm => b{eq,ne} rs, zero, imm
4590 if (MI.getOperand(0).getReg() == RISCV::X0) {
4591 MachineOperand MO0 = MI.getOperand(0);
4592 MI.removeOperand(0);
4593 MI.insert(MI.operands_begin() + 1, {MO0});
4594 }
4595 break;
4596 case RISCV::BLTU:
4597 // bltu zero, rs, imm => bne rs, zero, imm
4598 if (MI.getOperand(0).getReg() == RISCV::X0) {
4599 MachineOperand MO0 = MI.getOperand(0);
4600 MI.removeOperand(0);
4601 MI.insert(MI.operands_begin() + 1, {MO0});
4602 MI.setDesc(get(RISCV::BNE));
4603 }
4604 break;
4605 case RISCV::BGEU:
4606 // bgeu zero, rs, imm => beq rs, zero, imm
4607 if (MI.getOperand(0).getReg() == RISCV::X0) {
4608 MachineOperand MO0 = MI.getOperand(0);
4609 MI.removeOperand(0);
4610 MI.insert(MI.operands_begin() + 1, {MO0});
4611 MI.setDesc(get(RISCV::BEQ));
4612 }
4613 break;
4614 }
4615 return false;
4616}
4617
4618// clang-format off
4619#define CASE_WIDEOP_OPCODE_COMMON(OP, LMUL) \
4620 RISCV::PseudoV##OP##_##LMUL##_TIED
4621
4622#define CASE_WIDEOP_OPCODE_LMULS(OP) \
4623 CASE_WIDEOP_OPCODE_COMMON(OP, MF8): \
4624 case CASE_WIDEOP_OPCODE_COMMON(OP, MF4): \
4625 case CASE_WIDEOP_OPCODE_COMMON(OP, MF2): \
4626 case CASE_WIDEOP_OPCODE_COMMON(OP, M1): \
4627 case CASE_WIDEOP_OPCODE_COMMON(OP, M2): \
4628 case CASE_WIDEOP_OPCODE_COMMON(OP, M4)
4629
4630#define CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, LMUL) \
4631 case RISCV::PseudoV##OP##_##LMUL##_TIED: \
4632 NewOpc = RISCV::PseudoV##OP##_##LMUL; \
4633 break;
4634
4635#define CASE_WIDEOP_CHANGE_OPCODE_LMULS(OP) \
4636 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF8) \
4637 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4) \
4638 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2) \
4639 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1) \
4640 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2) \
4641 CASE_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4)
4642
4643// FP Widening Ops may by SEW aware. Create SEW aware cases for these cases.
4644#define CASE_FP_WIDEOP_OPCODE_COMMON(OP, LMUL, SEW) \
4645 RISCV::PseudoV##OP##_##LMUL##_##SEW##_TIED
4646
4647#define CASE_FP_WIDEOP_OPCODE_LMULS(OP) \
4648 CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF4, E16): \
4649 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E16): \
4650 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E32): \
4651 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E16): \
4652 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E32): \
4653 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E16): \
4654 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E32): \
4655 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E16): \
4656 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E32) \
4657
4658#define CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, LMUL, SEW) \
4659 case RISCV::PseudoV##OP##_##LMUL##_##SEW##_TIED: \
4660 NewOpc = RISCV::PseudoV##OP##_##LMUL##_##SEW; \
4661 break;
4662
4663#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS(OP) \
4664 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4, E16) \
4665 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E16) \
4666 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E32) \
4667 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E16) \
4668 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E32) \
4669 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E16) \
4670 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E32) \
4671 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E16) \
4672 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E32) \
4673
4674#define CASE_FP_WIDEOP_OPCODE_LMULS_ALT(OP) \
4675 CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF4, E16): \
4676 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, MF2, E16): \
4677 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M1, E16): \
4678 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M2, E16): \
4679 case CASE_FP_WIDEOP_OPCODE_COMMON(OP, M4, E16)
4680
4681#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS_ALT(OP) \
4682 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF4, E16) \
4683 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, MF2, E16) \
4684 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M1, E16) \
4685 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M2, E16) \
4686 CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON(OP, M4, E16)
4687// clang-format on
4688
4690 LiveVariables *LV,
4691 LiveIntervals *LIS) const {
4693 switch (MI.getOpcode()) {
4694 default:
4695 return nullptr;
4696 case CASE_FP_WIDEOP_OPCODE_LMULS_ALT(FWADD_ALT_WV):
4697 case CASE_FP_WIDEOP_OPCODE_LMULS_ALT(FWSUB_ALT_WV):
4698 case CASE_FP_WIDEOP_OPCODE_LMULS(FWADD_WV):
4699 case CASE_FP_WIDEOP_OPCODE_LMULS(FWSUB_WV): {
4700 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags) &&
4701 MI.getNumExplicitOperands() == 7 &&
4702 "Expect 7 explicit operands rd, rs2, rs1, rm, vl, sew, policy");
4703 // If the tail policy is undisturbed we can't convert.
4704 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4705 1) == 0)
4706 return nullptr;
4707 // clang-format off
4708 unsigned NewOpc;
4709 switch (MI.getOpcode()) {
4710 default:
4711 llvm_unreachable("Unexpected opcode");
4716 }
4717 // clang-format on
4718
4719 MachineBasicBlock &MBB = *MI.getParent();
4720 MIB = BuildMI(MBB, MI, MI.getDebugLoc(), get(NewOpc))
4721 .add(MI.getOperand(0))
4722 .addReg(MI.getOperand(0).getReg(), RegState::Undef)
4723 .add(MI.getOperand(1))
4724 .add(MI.getOperand(2))
4725 .add(MI.getOperand(3))
4726 .add(MI.getOperand(4))
4727 .add(MI.getOperand(5))
4728 .add(MI.getOperand(6));
4729 break;
4730 }
4731 case CASE_WIDEOP_OPCODE_LMULS(WADD_WV):
4732 case CASE_WIDEOP_OPCODE_LMULS(WADDU_WV):
4733 case CASE_WIDEOP_OPCODE_LMULS(WSUB_WV):
4734 case CASE_WIDEOP_OPCODE_LMULS(WSUBU_WV): {
4735 // If the tail policy is undisturbed we can't convert.
4736 assert(RISCVII::hasVecPolicyOp(MI.getDesc().TSFlags) &&
4737 MI.getNumExplicitOperands() == 6);
4738 if ((MI.getOperand(RISCVII::getVecPolicyOpNum(MI.getDesc())).getImm() &
4739 1) == 0)
4740 return nullptr;
4741
4742 // clang-format off
4743 unsigned NewOpc;
4744 switch (MI.getOpcode()) {
4745 default:
4746 llvm_unreachable("Unexpected opcode");
4751 }
4752 // clang-format on
4753
4754 MachineBasicBlock &MBB = *MI.getParent();
4755 MIB = BuildMI(MBB, MI, MI.getDebugLoc(), get(NewOpc))
4756 .add(MI.getOperand(0))
4757 .addReg(MI.getOperand(0).getReg(), RegState::Undef)
4758 .add(MI.getOperand(1))
4759 .add(MI.getOperand(2))
4760 .add(MI.getOperand(3))
4761 .add(MI.getOperand(4))
4762 .add(MI.getOperand(5));
4763 break;
4764 }
4765 }
4766 MIB.copyImplicitOps(MI);
4767
4768 if (LV) {
4769 unsigned NumOps = MI.getNumOperands();
4770 for (unsigned I = 1; I < NumOps; ++I) {
4771 MachineOperand &Op = MI.getOperand(I);
4772 if (Op.isReg() && Op.isKill())
4773 LV->replaceKillInstruction(Op.getReg(), MI, *MIB);
4774 }
4775 }
4776
4777 if (LIS) {
4778 SlotIndex Idx = LIS->ReplaceMachineInstrInMaps(MI, *MIB);
4779
4780 if (MI.getOperand(0).isEarlyClobber()) {
4781 // Use operand 1 was tied to early-clobber def operand 0, so its live
4782 // interval could have ended at an early-clobber slot. Now they are not
4783 // tied we need to update it to the normal register slot.
4784 LiveInterval &LI = LIS->getInterval(MI.getOperand(1).getReg());
4786 if (S->end == Idx.getRegSlot(true))
4787 S->end = Idx.getRegSlot();
4788 }
4789 }
4790
4791 return MIB;
4792}
4793
4794#undef CASE_WIDEOP_OPCODE_COMMON
4795#undef CASE_WIDEOP_OPCODE_LMULS
4796#undef CASE_WIDEOP_CHANGE_OPCODE_COMMON
4797#undef CASE_WIDEOP_CHANGE_OPCODE_LMULS
4798#undef CASE_FP_WIDEOP_OPCODE_COMMON
4799#undef CASE_FP_WIDEOP_OPCODE_LMULS
4800#undef CASE_FP_WIDEOP_CHANGE_OPCODE_COMMON
4801#undef CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS
4802
4805 Register DestReg, uint32_t Amount,
4806 MachineInstr::MIFlag Flag) const {
4808 if (llvm::has_single_bit<uint32_t>(Amount)) {
4809 uint32_t ShiftAmount = Log2_32(Amount);
4810 if (ShiftAmount == 0)
4811 return;
4812 BuildMI(MBB, II, DL, get(RISCV::SLLI), DestReg)
4813 .addReg(DestReg, RegState::Kill)
4814 .addImm(ShiftAmount)
4815 .setMIFlag(Flag);
4816 } else if (int ShXAmount, ShiftAmount;
4817 STI.hasShlAdd(3) &&
4818 (ShXAmount = isShifted359(Amount, ShiftAmount)) != 0) {
4819 // We can use Zba SHXADD+SLLI instructions for multiply in some cases.
4820 unsigned Opc;
4821 switch (ShXAmount) {
4822 case 1:
4823 Opc = RISCV::SH1ADD;
4824 break;
4825 case 2:
4826 Opc = RISCV::SH2ADD;
4827 break;
4828 case 3:
4829 Opc = RISCV::SH3ADD;
4830 break;
4831 default:
4832 llvm_unreachable("unexpected result of isShifted359");
4833 }
4834 if (ShiftAmount)
4835 BuildMI(MBB, II, DL, get(RISCV::SLLI), DestReg)
4836 .addReg(DestReg, RegState::Kill)
4837 .addImm(ShiftAmount)
4838 .setMIFlag(Flag);
4839 BuildMI(MBB, II, DL, get(Opc), DestReg)
4840 .addReg(DestReg, RegState::Kill)
4841 .addReg(DestReg)
4842 .setMIFlag(Flag);
4843 } else if (llvm::has_single_bit<uint32_t>(Amount - 1)) {
4844 Register ScaledRegister = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4845 uint32_t ShiftAmount = Log2_32(Amount - 1);
4846 BuildMI(MBB, II, DL, get(RISCV::SLLI), ScaledRegister)
4847 .addReg(DestReg)
4848 .addImm(ShiftAmount)
4849 .setMIFlag(Flag);
4850 BuildMI(MBB, II, DL, get(RISCV::ADD), DestReg)
4851 .addReg(ScaledRegister, RegState::Kill)
4852 .addReg(DestReg, RegState::Kill)
4853 .setMIFlag(Flag);
4854 } else if (llvm::has_single_bit<uint32_t>(Amount + 1)) {
4855 Register ScaledRegister = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4856 uint32_t ShiftAmount = Log2_32(Amount + 1);
4857 BuildMI(MBB, II, DL, get(RISCV::SLLI), ScaledRegister)
4858 .addReg(DestReg)
4859 .addImm(ShiftAmount)
4860 .setMIFlag(Flag);
4861 BuildMI(MBB, II, DL, get(RISCV::SUB), DestReg)
4862 .addReg(ScaledRegister, RegState::Kill)
4863 .addReg(DestReg, RegState::Kill)
4864 .setMIFlag(Flag);
4865 } else if (STI.hasStdExtZmmul()) {
4866 Register N = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4867 movImm(MBB, II, DL, N, Amount, Flag);
4868 BuildMI(MBB, II, DL, get(RISCV::MUL), DestReg)
4869 .addReg(DestReg, RegState::Kill)
4871 .setMIFlag(Flag);
4872 } else {
4873 Register Acc;
4874 uint32_t PrevShiftAmount = 0;
4875 for (uint32_t ShiftAmount = 0; Amount >> ShiftAmount; ShiftAmount++) {
4876 if (Amount & (1U << ShiftAmount)) {
4877 if (ShiftAmount)
4878 BuildMI(MBB, II, DL, get(RISCV::SLLI), DestReg)
4879 .addReg(DestReg, RegState::Kill)
4880 .addImm(ShiftAmount - PrevShiftAmount)
4881 .setMIFlag(Flag);
4882 if (Amount >> (ShiftAmount + 1)) {
4883 // If we don't have an accmulator yet, create it and copy DestReg.
4884 if (!Acc) {
4885 Acc = MRI.createVirtualRegister(&RISCV::GPRRegClass);
4886 BuildMI(MBB, II, DL, get(TargetOpcode::COPY), Acc)
4887 .addReg(DestReg)
4888 .setMIFlag(Flag);
4889 } else {
4890 BuildMI(MBB, II, DL, get(RISCV::ADD), Acc)
4891 .addReg(Acc, RegState::Kill)
4892 .addReg(DestReg)
4893 .setMIFlag(Flag);
4894 }
4895 }
4896 PrevShiftAmount = ShiftAmount;
4897 }
4898 }
4899 assert(Acc && "Expected valid accumulator");
4900 BuildMI(MBB, II, DL, get(RISCV::ADD), DestReg)
4901 .addReg(DestReg, RegState::Kill)
4902 .addReg(Acc, RegState::Kill)
4903 .setMIFlag(Flag);
4904 }
4905}
4906
4909 static const std::pair<MachineMemOperand::Flags, const char *> TargetFlags[] =
4910 {{MONontemporalBit0, "riscv-nontemporal-domain-bit-0"},
4911 {MONontemporalBit1, "riscv-nontemporal-domain-bit-1"}};
4912 return ArrayRef(TargetFlags);
4913}
4914
4916 return OptLevel >= CodeGenOptLevel::Aggressive
4917 ? STI.getTailDupAggressiveThreshold()
4918 : 2;
4919}
4920
4922 // RVV lacks any support for immediate addressing for stack addresses, so be
4923 // conservative.
4924 unsigned Opcode = MI.getOpcode();
4925 if (!RISCVVPseudosTable::getPseudoInfo(Opcode) &&
4927 return false;
4928 return true;
4929}
4930
4931/// Return true if \p MI is a copy that will be lowered to one or more vmvNr.vs.
4933 const MachineInstr &MI) {
4934 return MI.isCopy() && MI.getOperand(0).getReg().isPhysical() &&
4936 TRI->getMinimalPhysRegClass(MI.getOperand(0).getReg()));
4937}
4938
4939std::optional<std::pair<unsigned, unsigned>>
4941 switch (Opcode) {
4942 default:
4943 return std::nullopt;
4944 case RISCV::PseudoVSPILL2_M1:
4945 case RISCV::PseudoVRELOAD2_M1:
4946 return std::make_pair(2u, 1u);
4947 case RISCV::PseudoVSPILL2_M2:
4948 case RISCV::PseudoVRELOAD2_M2:
4949 return std::make_pair(2u, 2u);
4950 case RISCV::PseudoVSPILL2_M4:
4951 case RISCV::PseudoVRELOAD2_M4:
4952 return std::make_pair(2u, 4u);
4953 case RISCV::PseudoVSPILL3_M1:
4954 case RISCV::PseudoVRELOAD3_M1:
4955 return std::make_pair(3u, 1u);
4956 case RISCV::PseudoVSPILL3_M2:
4957 case RISCV::PseudoVRELOAD3_M2:
4958 return std::make_pair(3u, 2u);
4959 case RISCV::PseudoVSPILL4_M1:
4960 case RISCV::PseudoVRELOAD4_M1:
4961 return std::make_pair(4u, 1u);
4962 case RISCV::PseudoVSPILL4_M2:
4963 case RISCV::PseudoVRELOAD4_M2:
4964 return std::make_pair(4u, 2u);
4965 case RISCV::PseudoVSPILL5_M1:
4966 case RISCV::PseudoVRELOAD5_M1:
4967 return std::make_pair(5u, 1u);
4968 case RISCV::PseudoVSPILL6_M1:
4969 case RISCV::PseudoVRELOAD6_M1:
4970 return std::make_pair(6u, 1u);
4971 case RISCV::PseudoVSPILL7_M1:
4972 case RISCV::PseudoVRELOAD7_M1:
4973 return std::make_pair(7u, 1u);
4974 case RISCV::PseudoVSPILL8_M1:
4975 case RISCV::PseudoVRELOAD8_M1:
4976 return std::make_pair(8u, 1u);
4977 }
4978}
4979
4980bool RISCV::hasEqualFRM(const MachineInstr &MI1, const MachineInstr &MI2) {
4981 int16_t MI1FrmOpIdx =
4982 RISCV::getNamedOperandIdx(MI1.getOpcode(), RISCV::OpName::frm);
4983 int16_t MI2FrmOpIdx =
4984 RISCV::getNamedOperandIdx(MI2.getOpcode(), RISCV::OpName::frm);
4985 if (MI1FrmOpIdx < 0 || MI2FrmOpIdx < 0)
4986 return false;
4987 MachineOperand FrmOp1 = MI1.getOperand(MI1FrmOpIdx);
4988 MachineOperand FrmOp2 = MI2.getOperand(MI2FrmOpIdx);
4989 return FrmOp1.getImm() == FrmOp2.getImm();
4990}
4991
4992std::optional<unsigned>
4993RISCV::getVectorLowDemandedScalarBits(unsigned Opcode, unsigned Log2SEW) {
4994 switch (Opcode) {
4995 default:
4996 return std::nullopt;
4997
4998 // 11.6. Vector Single-Width Shift Instructions
4999 case RISCV::VSLL_VX:
5000 case RISCV::VSRL_VX:
5001 case RISCV::VSRA_VX:
5002 // 12.4. Vector Single-Width Scaling Shift Instructions
5003 case RISCV::VSSRL_VX:
5004 case RISCV::VSSRA_VX:
5005 // Zvbb
5006 case RISCV::VROL_VX:
5007 case RISCV::VROR_VX:
5008 // Only the low lg2(SEW) bits of the shift-amount value are used.
5009 return Log2SEW;
5010
5011 // 11.7 Vector Narrowing Integer Right Shift Instructions
5012 case RISCV::VNSRL_WX:
5013 case RISCV::VNSRA_WX:
5014 // 12.5. Vector Narrowing Fixed-Point Clip Instructions
5015 case RISCV::VNCLIPU_WX:
5016 case RISCV::VNCLIP_WX:
5017 // Zvbb
5018 case RISCV::VWSLL_VX:
5019 // Only the low lg2(2*SEW) bits of the shift-amount value are used.
5020 return Log2SEW + 1;
5021
5022 // 11.1. Vector Single-Width Integer Add and Subtract
5023 case RISCV::VADD_VX:
5024 case RISCV::VSUB_VX:
5025 case RISCV::VRSUB_VX:
5026 // 11.2. Vector Widening Integer Add/Subtract
5027 case RISCV::VWADDU_VX:
5028 case RISCV::VWSUBU_VX:
5029 case RISCV::VWADD_VX:
5030 case RISCV::VWSUB_VX:
5031 case RISCV::VWADDU_WX:
5032 case RISCV::VWSUBU_WX:
5033 case RISCV::VWADD_WX:
5034 case RISCV::VWSUB_WX:
5035 // 11.4. Vector Integer Add-with-Carry / Subtract-with-Borrow Instructions
5036 case RISCV::VADC_VXM:
5037 case RISCV::VADC_VIM:
5038 case RISCV::VMADC_VXM:
5039 case RISCV::VMADC_VIM:
5040 case RISCV::VMADC_VX:
5041 case RISCV::VSBC_VXM:
5042 case RISCV::VMSBC_VXM:
5043 case RISCV::VMSBC_VX:
5044 // 11.5 Vector Bitwise Logical Instructions
5045 case RISCV::VAND_VX:
5046 case RISCV::VOR_VX:
5047 case RISCV::VXOR_VX:
5048 // 11.8. Vector Integer Compare Instructions
5049 case RISCV::VMSEQ_VX:
5050 case RISCV::VMSNE_VX:
5051 case RISCV::VMSLTU_VX:
5052 case RISCV::VMSLT_VX:
5053 case RISCV::VMSLEU_VX:
5054 case RISCV::VMSLE_VX:
5055 case RISCV::VMSGTU_VX:
5056 case RISCV::VMSGT_VX:
5057 // 11.9. Vector Integer Min/Max Instructions
5058 case RISCV::VMINU_VX:
5059 case RISCV::VMIN_VX:
5060 case RISCV::VMAXU_VX:
5061 case RISCV::VMAX_VX:
5062 // 11.10. Vector Single-Width Integer Multiply Instructions
5063 case RISCV::VMUL_VX:
5064 case RISCV::VMULH_VX:
5065 case RISCV::VMULHU_VX:
5066 case RISCV::VMULHSU_VX:
5067 // 11.11. Vector Integer Divide Instructions
5068 case RISCV::VDIVU_VX:
5069 case RISCV::VDIV_VX:
5070 case RISCV::VREMU_VX:
5071 case RISCV::VREM_VX:
5072 // 11.12. Vector Widening Integer Multiply Instructions
5073 case RISCV::VWMUL_VX:
5074 case RISCV::VWMULU_VX:
5075 case RISCV::VWMULSU_VX:
5076 // 11.13. Vector Single-Width Integer Multiply-Add Instructions
5077 case RISCV::VMACC_VX:
5078 case RISCV::VNMSAC_VX:
5079 case RISCV::VMADD_VX:
5080 case RISCV::VNMSUB_VX:
5081 // 11.14. Vector Widening Integer Multiply-Add Instructions
5082 case RISCV::VWMACCU_VX:
5083 case RISCV::VWMACC_VX:
5084 case RISCV::VWMACCSU_VX:
5085 case RISCV::VWMACCUS_VX:
5086 // 11.15. Vector Integer Merge Instructions
5087 case RISCV::VMERGE_VXM:
5088 // 11.16. Vector Integer Move Instructions
5089 case RISCV::VMV_V_X:
5090 // 12.1. Vector Single-Width Saturating Add and Subtract
5091 case RISCV::VSADDU_VX:
5092 case RISCV::VSADD_VX:
5093 case RISCV::VSSUBU_VX:
5094 case RISCV::VSSUB_VX:
5095 // 12.2. Vector Single-Width Averaging Add and Subtract
5096 case RISCV::VAADDU_VX:
5097 case RISCV::VAADD_VX:
5098 case RISCV::VASUBU_VX:
5099 case RISCV::VASUB_VX:
5100 // 12.3. Vector Single-Width Fractional Multiply with Rounding and Saturation
5101 case RISCV::VSMUL_VX:
5102 // 16.1. Integer Scalar Move Instructions
5103 case RISCV::VMV_S_X:
5104 // Zvbb
5105 case RISCV::VANDN_VX:
5106 return 1U << Log2SEW;
5107 }
5108}
5109
5110unsigned RISCV::getRVVMCOpcode(unsigned RVVPseudoOpcode) {
5112 RISCVVPseudosTable::getPseudoInfo(RVVPseudoOpcode);
5113 if (!RVV)
5114 return 0;
5115 return RVV->BaseInstr;
5116}
5117
5118unsigned RISCV::getDestLog2EEW(const MCInstrDesc &Desc, unsigned Log2SEW) {
5119 unsigned DestEEW =
5121 // EEW = 1
5122 if (DestEEW == 0)
5123 return 0;
5124 // EEW = SEW * n
5125 unsigned Scaled = Log2SEW + (DestEEW - 1);
5126 assert(Scaled >= 3 && Scaled <= 6);
5127 return Scaled;
5128}
5129
5130static std::optional<int64_t> getEffectiveImm(const MachineOperand &MO) {
5131 assert(MO.isImm() || MO.getReg().isVirtual());
5132 if (MO.isImm())
5133 return MO.getImm();
5134 const MachineInstr *Def =
5135 MO.getParent()->getMF()->getRegInfo().getVRegDef(MO.getReg());
5136 int64_t Imm;
5137 if (isLoadImm(Def, Imm))
5138 return Imm;
5139 return std::nullopt;
5140}
5141
5142/// Given two VL operands, do we know that LHS <= RHS? Must be used in SSA form.
5144 assert((LHS.isImm() || LHS.getParent()->getMF()->getRegInfo().isSSA()) &&
5145 (RHS.isImm() || RHS.getParent()->getMF()->getRegInfo().isSSA()));
5146 if (LHS.isReg() && RHS.isReg() && LHS.getReg().isVirtual() &&
5147 LHS.getReg() == RHS.getReg())
5148 return true;
5149 if (RHS.isImm() && RHS.getImm() == RISCV::VLMaxSentinel)
5150 return true;
5151 if (LHS.isImm() && LHS.getImm() == 0)
5152 return true;
5153 if (LHS.isImm() && LHS.getImm() == RISCV::VLMaxSentinel)
5154 return false;
5155 std::optional<int64_t> LHSImm = getEffectiveImm(LHS),
5156 RHSImm = getEffectiveImm(RHS);
5157 if (!LHSImm || !RHSImm)
5158 return false;
5159 return LHSImm <= RHSImm;
5160}
5161
5162namespace {
5163class RISCVPipelinerLoopInfo : public TargetInstrInfo::PipelinerLoopInfo {
5164 const MachineInstr *LHS;
5165 const MachineInstr *RHS;
5167
5168public:
5169 RISCVPipelinerLoopInfo(const MachineInstr *LHS, const MachineInstr *RHS,
5171 : LHS(LHS), RHS(RHS), Cond(Cond.begin(), Cond.end()) {}
5172
5173 bool shouldIgnoreForPipelining(const MachineInstr *MI) const override {
5174 // Make the instructions for loop control be placed in stage 0.
5175 // The predecessors of LHS/RHS are considered by the caller.
5176 if (LHS && MI == LHS)
5177 return true;
5178 if (RHS && MI == RHS)
5179 return true;
5180 return false;
5181 }
5182
5183 std::optional<bool> createTripCountGreaterCondition(
5184 int TC, MachineBasicBlock &MBB,
5185 SmallVectorImpl<MachineOperand> &CondParam) override {
5186 // A branch instruction will be inserted as "if (Cond) goto epilogue".
5187 // Cond is normalized for such use.
5188 // The predecessors of the branch are assumed to have already been inserted.
5189 CondParam = Cond;
5190 return {};
5191 }
5192
5193 void setPreheader(MachineBasicBlock *NewPreheader) override {}
5194
5195 void adjustTripCount(int TripCountAdjust) override {}
5196};
5197} // namespace
5198
5199std::unique_ptr<TargetInstrInfo::PipelinerLoopInfo>
5201 MachineBasicBlock *TBB = nullptr, *FBB = nullptr;
5203 if (analyzeBranch(*LoopBB, TBB, FBB, Cond, /*AllowModify=*/false))
5204 return nullptr;
5205
5206 // Infinite loops are not supported
5207 if (TBB == LoopBB && FBB == LoopBB)
5208 return nullptr;
5209
5210 // Must be conditional branch
5211 if (FBB == nullptr)
5212 return nullptr;
5213
5214 assert((TBB == LoopBB || FBB == LoopBB) &&
5215 "The Loop must be a single-basic-block loop");
5216
5217 // Normalization for createTripCountGreaterCondition()
5218 if (TBB == LoopBB)
5220
5221 const MachineRegisterInfo &MRI = LoopBB->getParent()->getRegInfo();
5222 auto FindRegDef = [&MRI](MachineOperand &Op) -> const MachineInstr * {
5223 if (!Op.isReg())
5224 return nullptr;
5225 Register Reg = Op.getReg();
5226 if (!Reg.isVirtual())
5227 return nullptr;
5228 return MRI.getVRegDef(Reg);
5229 };
5230
5231 const MachineInstr *LHS = FindRegDef(Cond[1]);
5232 const MachineInstr *RHS = FindRegDef(Cond[2]);
5233 if (LHS && LHS->isPHI())
5234 return nullptr;
5235 if (RHS && RHS->isPHI())
5236 return nullptr;
5237
5238 return std::make_unique<RISCVPipelinerLoopInfo>(LHS, RHS, Cond);
5239}
5240
5241// FIXME: We should remove this if we have a default generic scheduling model.
5243 unsigned RVVMCOpcode = RISCV::getRVVMCOpcode(Opc);
5244 Opc = RVVMCOpcode ? RVVMCOpcode : Opc;
5245 switch (Opc) {
5246 default:
5247 return false;
5248 // Integer div/rem.
5249 case RISCV::DIV:
5250 case RISCV::DIVW:
5251 case RISCV::DIVU:
5252 case RISCV::DIVUW:
5253 case RISCV::REM:
5254 case RISCV::REMW:
5255 case RISCV::REMU:
5256 case RISCV::REMUW:
5257 // Floating-point div/sqrt.
5258 case RISCV::FDIV_H:
5259 case RISCV::FDIV_S:
5260 case RISCV::FDIV_D:
5261 case RISCV::FDIV_H_INX:
5262 case RISCV::FDIV_S_INX:
5263 case RISCV::FDIV_D_INX:
5264 case RISCV::FDIV_D_IN32X:
5265 case RISCV::FSQRT_H:
5266 case RISCV::FSQRT_S:
5267 case RISCV::FSQRT_D:
5268 case RISCV::FSQRT_H_INX:
5269 case RISCV::FSQRT_S_INX:
5270 case RISCV::FSQRT_D_INX:
5271 case RISCV::FSQRT_D_IN32X:
5272 // Vector integer div/rem
5273 case RISCV::VDIV_VV:
5274 case RISCV::VDIV_VX:
5275 case RISCV::VDIVU_VV:
5276 case RISCV::VDIVU_VX:
5277 case RISCV::VREM_VV:
5278 case RISCV::VREM_VX:
5279 case RISCV::VREMU_VV:
5280 case RISCV::VREMU_VX:
5281 // Vector floating-point div/sqrt.
5282 case RISCV::VFDIV_VV:
5283 case RISCV::VFDIV_VF:
5284 case RISCV::VFRDIV_VF:
5285 case RISCV::VFSQRT_V:
5286 case RISCV::VFRSQRT7_V:
5287 return true;
5288 }
5289}
5290
5291bool RISCVInstrInfo::isVRegCopy(const MachineInstr *MI, unsigned LMul) const {
5292 if (MI->getOpcode() != TargetOpcode::COPY)
5293 return false;
5294 const MachineRegisterInfo &MRI = MI->getMF()->getRegInfo();
5295 const TargetRegisterInfo *TRI = MRI.getTargetRegisterInfo();
5296
5297 Register DstReg = MI->getOperand(0).getReg();
5298 const TargetRegisterClass *RC = DstReg.isVirtual()
5299 ? MRI.getRegClass(DstReg)
5300 : TRI->getMinimalPhysRegClass(DstReg);
5301
5303 return false;
5304
5305 if (!LMul)
5306 return true;
5307
5308 // TODO: Perhaps we could distinguish segment register classes (e.g. VRN3M2)
5309 // in the future.
5310 auto [RCLMul, RCFractional] =
5312 return (!RCFractional && LMul == RCLMul) || (RCFractional && LMul == 1);
5313}
unsigned SubReg
unsigned const MachineRegisterInfo * MRI
MachineInstrBuilder MachineInstrBuilder & DefMI
static bool forwardCopyWillClobberTuple(unsigned DestReg, unsigned SrcReg, unsigned NumRegs)
static void parseCondBranch(MachineInstr *LastInst, MachineBasicBlock *&Target, SmallVectorImpl< MachineOperand > &Cond)
@ MachineOutlinerTailCall
Emit a save, restore, call, and return.
@ MachineOutlinerDefault
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
SmallVector< int16_t, MAX_SRC_OPERANDS_NUM > OperandIndices
@ Scaled
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
basic Basic Alias true
#define clEnumValN(ENUMVAL, FLAGNAME, DESC)
const HexagonInstrInfo * TII
#define _
IRTranslator LLVM IR MI
Module.h This file contains the declarations for the Module class.
const size_t AbstractManglingParser< Derived, Alloc >::NumOps
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
Register Reg
Register const TargetRegisterInfo * TRI
Promote Memory to Register
Definition Mem2Reg.cpp:110
This file provides utility analysis objects describing memory locations.
MachineInstr unsigned OpIdx
uint64_t IntrinsicInst * II
static bool cannotInsertTailCall(const MachineBasicBlock &MBB)
#define CASE_VFMA_CHANGE_OPCODE_SPLATS(OLDOP, NEWOP)
#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS_ALT(OP)
#define CASE_FP_WIDEOP_OPCODE_LMULS(OP)
#define CASE_OPERAND_SIMM(NUM)
static std::optional< unsigned > getLMULForRVVWholeLoadStore(unsigned Opcode)
#define CASE_VFMA_CHANGE_OPCODE_VV(OLDOP, NEWOP)
static bool analyzeCandidate(outliner::Candidate &C)
static unsigned getFPFusedMultiplyOpcode(unsigned RootOpc, unsigned Pattern)
std::optional< unsigned > getFoldedOpcode(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, const RISCVSubtarget &ST)
#define RVV_OPC_LMUL_CASE(OPC, INV)
#define CASE_FP_WIDEOP_CHANGE_OPCODE_LMULS(OP)
static void combineFPFusedMultiply(MachineInstr &Root, MachineInstr &Prev, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs)
static unsigned getAddendOperandIdx(unsigned Pattern)
#define CASE_RVV_OPCODE_UNMASK(OP)
#define CASE_WIDEOP_CHANGE_OPCODE_LMULS(OP)
static cl::opt< bool > PreferWholeRegisterMove("riscv-prefer-whole-register-move", cl::init(false), cl::Hidden, cl::desc("Prefer whole register move for vector registers."))
#define CASE_VFMA_SPLATS(OP)
unsigned getPredicatedOpcode(unsigned Opcode)
#define CASE_FP_WIDEOP_OPCODE_LMULS_ALT(OP)
#define CASE_WIDEOP_OPCODE_LMULS(OP)
static bool isMIReadsReg(const MachineInstr &MI, const TargetRegisterInfo *TRI, MCRegister RegNo)
#define OPCODE_LMUL_MASK_CASE(OPC)
static bool isFSUB(unsigned Opc)
#define CASE_VMA_CHANGE_OPCODE_LMULS(OLDOP, NEWOP, TYPE)
#define CASE_RVV_OPCODE(OP)
static std::optional< int64_t > getEffectiveImm(const MachineOperand &MO)
#define CASE_VFMA_OPCODE_VV(OP)
MachineOutlinerConstructionID
#define CASE_RVV_OPCODE_WIDEN(OP)
static unsigned getLoadPredicatedOpcode(unsigned Opcode)
static unsigned getSHXADDUWShiftAmount(unsigned Opc)
#define CASE_VMA_OPCODE_LMULS(OP, TYPE)
static bool isConvertibleToVMV_V_V(const RISCVSubtarget &STI, const MachineBasicBlock &MBB, MachineBasicBlock::const_iterator MBBI, MachineBasicBlock::const_iterator &DefMBBI, RISCVVType::VLMUL LMul)
static bool isFMUL(unsigned Opc)
static unsigned getInverseXqcicmOpcode(unsigned Opcode)
static bool getFPPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce)
#define OPCODE_LMUL_CASE(OPC)
#define CASE_OPERAND_UIMM(NUM)
static bool canCombineShiftIntoShXAdd(const MachineBasicBlock &MBB, const MachineOperand &MO, unsigned OuterShiftAmt)
Utility routine that checks if.
static bool isCandidatePatchable(const MachineBasicBlock &MBB)
static bool isFADD(unsigned Opc)
static void genShXAddAddShift(MachineInstr &Root, unsigned AddOpIdx, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg)
static bool isLoadImm(const MachineInstr *MI, int64_t &Imm)
static bool isMIModifiesReg(const MachineInstr &MI, const TargetRegisterInfo *TRI, MCRegister RegNo)
static bool canCombineFPFusedMultiply(const MachineInstr &Root, const MachineOperand &MO, bool DoRegPressureReduce)
static bool getSHXADDPatterns(const MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns)
static bool getFPFusedMultiplyPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce)
static cl::opt< MachineTraceStrategy > ForceMachineCombinerStrategy("riscv-force-machine-combiner-strategy", cl::Hidden, cl::desc("Force machine combiner to use a specific strategy for machine " "trace metrics evaluation."), cl::init(MachineTraceStrategy::TS_NumStrategies), cl::values(clEnumValN(MachineTraceStrategy::TS_Local, "local", "Local strategy."), clEnumValN(MachineTraceStrategy::TS_MinInstrCount, "min-instr", "MinInstrCount strategy.")))
static unsigned getSHXADDShiftAmount(unsigned Opc)
#define CASE_RVV_OPCODE_MASK(OP)
#define RVV_OPC_LMUL_MASK_CASE(OPC, INV)
static MachineInstr * canFoldAsPredicatedOp(Register Reg, const MachineRegisterInfo &MRI, const TargetInstrInfo *TII, const RISCVSubtarget &STI)
Identify instructions that can be folded into a CCMOV instruction, and return the defining instructio...
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
This file declares the machine register scavenger class.
static bool memOpsHaveSameBasePtr(const MachineInstr &MI1, ArrayRef< const MachineOperand * > BaseOps1, const MachineInstr &MI2, ArrayRef< const MachineOperand * > BaseOps2)
This file contains some templates that are useful if you are working with the STL at all.
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
Definition Value.cpp:487
This file defines the SmallVector class.
This file defines the 'Statistic' class, which is designed to be an easy way to expose various metric...
#define STATISTIC(VARNAME, DESC)
Definition Statistic.h:171
static TableGen::Emitter::Opt Y("gen-skeleton-entry", EmitSkeleton, "Generate example skeleton entry")
static TableGen::Emitter::OptClass< SkeletonEmitter > X("gen-skeleton-class", "Generate example skeleton class")
static bool canCombine(MachineBasicBlock &MBB, MachineOperand &MO, unsigned CombineOpc=0)
static cl::opt< unsigned > CacheLineSize("cache-line-size", cl::init(0), cl::Hidden, cl::desc("Use this to override the target cache line size when " "specified by the user."))
Value * RHS
Value * LHS
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
const T & front() const
front - Get the first element.
Definition ArrayRef.h:145
bool empty() const
empty - Check if the array is empty.
Definition ArrayRef.h:137
static LLVM_ABI DILocation * getMergedLocation(DILocation *LocA, DILocation *LocB)
Attempts to merge LocA and LocB into a single location; see DebugLoc::getMergedLocation for more deta...
bool isBigEndian() const
Definition DataLayout.h:215
A debug info location.
Definition DebugLoc.h:123
std::pair< iterator, bool > insert(const std::pair< KeyT, ValueT > &KV)
Definition DenseMap.h:241
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Definition Function.h:703
LiveInterval - This class represents the liveness of a register, or stack slot.
LiveInterval & getInterval(Register Reg)
SlotIndex ReplaceMachineInstrInMaps(MachineInstr &MI, MachineInstr &NewMI)
const Segment * getSegmentContaining(SlotIndex Idx) const
Return the segment that contains the specified index, or null if there is none.
LLVM_ABI void replaceKillInstruction(Register Reg, MachineInstr &OldMI, MachineInstr &NewMI)
replaceKillInstruction - Update register kill info by replacing a kill instruction with a new one.
bool hasValue() const
static LocationSize precise(uint64_t Value)
TypeSize getValue() const
MCInstBuilder & addReg(MCRegister Reg)
Add a new register operand.
MCInstBuilder & addImm(int64_t Val)
Add a new integer immediate operand.
Instances of this class represent a single low-level machine instruction.
Definition MCInst.h:188
Describe properties that are true of each instruction in the target description file.
unsigned getNumOperands() const
Return the number of declared MachineOperands for this MachineInstruction.
bool isConditionalBranch() const
Return true if this is a branch which may fall through to the next instruction or may transfer contro...
This holds information about one operand of a machine instruction, indicating the register class for ...
Definition MCInstrDesc.h:86
Wrapper class representing physical registers. Should be passed by value.
Definition MCRegister.h:41
const FeatureBitset & getFeatureBits() const
Set of metadata that should be preserved when using BuildMI().
MachineInstrBundleIterator< const MachineInstr > const_iterator
MachineInstrBundleIterator< MachineInstr, true > reverse_iterator
Instructions::const_iterator const_instr_iterator
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
MachineInstrBundleIterator< MachineInstr > iterator
MachineInstrBundleIterator< const MachineInstr, true > const_reverse_iterator
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
void setStackID(int ObjectIdx, uint8_t ID)
Align getObjectAlign(int ObjectIdx) const
Return the alignment of the specified stack object.
int64_t getObjectSize(int ObjectIdx) const
Return the size of the specified object.
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
StringRef getName() const
getName - Return the name of the corresponding LLVM function.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & setMemRefs(ArrayRef< MachineMemOperand * > MMOs) const
const MachineInstrBuilder & setMIFlag(MachineInstr::MIFlag Flag) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & cloneMemRefs(const MachineInstr &OtherMI) const
const MachineInstrBuilder & addUse(Register RegNo, unsigned Flags=0, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & setMIFlags(unsigned Flags) const
const MachineInstrBuilder & copyImplicitOps(const MachineInstr &OtherMI) const
Copy all the implicit operands from OtherMI onto this one.
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
reverse_iterator getReverse() const
Get a reverse iterator to the same node.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
bool isReturn(QueryType Type=AnyInBundle) const
bool mayLoadOrStore(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read or modify memory.
const MachineBasicBlock * getParent() const
bool getFlag(MIFlag Flag) const
Return whether an MI flag is set.
LLVM_ABI unsigned getNumExplicitOperands() const
Returns the number of non-implicit operands.
bool modifiesRegister(Register Reg, const TargetRegisterInfo *TRI) const
Return true if the MachineInstr modifies (fully define or partially define) the specified register.
bool mayLoad(QueryType Type=AnyInBundle) const
Return true if this instruction could possibly read memory.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
LLVM_ABI bool hasUnmodeledSideEffects() const
Return true if this instruction has side effects that are not modeled by mayLoad / mayStore,...
bool hasOneMemOperand() const
Return true if this instruction has exactly one MachineMemOperand.
mmo_iterator memoperands_begin() const
Access to memory operands of the instruction.
LLVM_ABI bool hasOrderedMemoryRef() const
Return true if this instruction may have an ordered or volatile memory reference, or if the informati...
LLVM_ABI const MachineFunction * getMF() const
Return the function that contains the basic block that this instruction belongs to.
ArrayRef< MachineMemOperand * > memoperands() const
Access to memory operands of the instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
const MachineOperand & getOperand(unsigned i) const
uint32_t getFlags() const
Return the MI flags bitvector.
LLVM_ABI void clearKillInfo()
Clears kill flags on all operands.
A description of a memory reference used in the backend.
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
This class contains meta information specific to a module.
MachineOperand class - Representation of each machine instruction operand.
int64_t getImm() const
bool isReg() const
isReg - Tests if this is a MO_Register operand.
MachineBasicBlock * getMBB() const
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
MachineInstr * getParent()
getParent - Return the instruction that this operand belongs to.
static MachineOperand CreateImm(int64_t Val)
MachineOperandType getType() const
getType - Returns the MachineOperandType for this operand.
Register getReg() const
getReg - Returns the register number.
bool isFI() const
isFI - Tests if this is a MO_FrameIndex operand.
LLVM_ABI bool isIdenticalTo(const MachineOperand &Other) const
Returns true if this operand is identical to the specified operand except for liveness related flags ...
@ MO_Immediate
Immediate operand.
@ MO_Register
Register operand.
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
LLVM_ABI MachineInstr * getVRegDef(Register Reg) const
getVRegDef - Return the machine instr that defines the specified virtual register or null if none is ...
A Module instance is used to store all the information related to an LLVM module.
Definition Module.h:67
MI-level patchpoint operands.
Definition StackMaps.h:77
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given patchpoint should emit.
Definition StackMaps.h:105
void storeRegToStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register SrcReg, bool IsKill, int FrameIndex, const TargetRegisterClass *RC, Register VReg, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
MachineInstr * convertToThreeAddress(MachineInstr &MI, LiveVariables *LV, LiveIntervals *LIS) const override
Register isLoadFromStackSlot(const MachineInstr &MI, int &FrameIndex) const override
std::optional< std::unique_ptr< outliner::OutlinedFunction > > getOutliningCandidateInfo(const MachineModuleInfo &MMI, std::vector< outliner::Candidate > &RepeatedSequenceLocs, unsigned MinRepeats) const override
unsigned removeBranch(MachineBasicBlock &MBB, int *BytesRemoved=nullptr) const override
void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstrIdxForVirtReg) const override
void movImm(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, Register DstReg, uint64_t Val, MachineInstr::MIFlag Flag=MachineInstr::NoFlags, bool DstRenamable=false, bool DstIsDead=false) const
MachineInstr * emitLdStWithAddr(MachineInstr &MemI, const ExtAddrMode &AM) const override
void mulImm(MachineFunction &MF, MachineBasicBlock &MBB, MachineBasicBlock::iterator II, const DebugLoc &DL, Register DestReg, uint32_t Amt, MachineInstr::MIFlag Flag) const
Generate code to multiply the value in DestReg by Amt - handles all the common optimizations for this...
static bool isPairableLdStInstOpc(unsigned Opc)
Return true if pairing the given load or store may be paired with another.
RISCVInstrInfo(const RISCVSubtarget &STI)
void loadRegFromStackSlot(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, Register DstReg, int FrameIndex, const TargetRegisterClass *RC, Register VReg, unsigned SubReg=0, MachineInstr::MIFlag Flags=MachineInstr::NoFlags) const override
bool isFunctionSafeToOutlineFrom(MachineFunction &MF, bool OutlineFromLinkOnceODRs) const override
std::unique_ptr< TargetInstrInfo::PipelinerLoopInfo > analyzeLoopForPipelining(MachineBasicBlock *LoopBB) const override
unsigned insertBranch(MachineBasicBlock &MBB, MachineBasicBlock *TBB, MachineBasicBlock *FBB, ArrayRef< MachineOperand > Cond, const DebugLoc &dl, int *BytesAdded=nullptr) const override
bool hasReassociableSibling(const MachineInstr &Inst, bool &Commuted) const override
static bool isLdStSafeToPair(const MachineInstr &LdSt, const TargetRegisterInfo *TRI)
void copyPhysRegVector(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, MCRegister DstReg, MCRegister SrcReg, bool KillSrc, const TargetRegisterClass *RegClass) const
bool isReMaterializableImpl(const MachineInstr &MI) const override
MachineInstr * optimizeSelect(MachineInstr &MI, SmallPtrSetImpl< MachineInstr * > &SeenMIs, bool) const override
bool isVRegCopy(const MachineInstr *MI, unsigned LMul=0) const
Return true if MI is a COPY to a vector register of a specific LMul, or any kind of vector registers ...
bool canFoldIntoAddrMode(const MachineInstr &MemI, Register Reg, const MachineInstr &AddrI, ExtAddrMode &AM) const override
void insertIndirectBranch(MachineBasicBlock &MBB, MachineBasicBlock &NewDestBB, MachineBasicBlock &RestoreBB, const DebugLoc &DL, int64_t BrOffset, RegScavenger *RS) const override
bool isAsCheapAsAMove(const MachineInstr &MI) const override
bool verifyInstruction(const MachineInstr &MI, StringRef &ErrInfo) const override
bool getMemOperandWithOffsetWidth(const MachineInstr &LdSt, const MachineOperand *&BaseOp, int64_t &Offset, LocationSize &Width, const TargetRegisterInfo *TRI) const
unsigned getTailDuplicateSize(CodeGenOptLevel OptLevel) const override
void getReassociateOperandIndices(const MachineInstr &Root, unsigned Pattern, std::array< unsigned, 5 > &OperandIndices) const override
const RISCVSubtarget & STI
Register isStoreToStackSlot(const MachineInstr &MI, int &FrameIndex) const override
std::optional< unsigned > getInverseOpcode(unsigned Opcode) const override
bool simplifyInstruction(MachineInstr &MI) const override
ArrayRef< std::pair< unsigned, const char * > > getSerializableDirectMachineOperandTargetFlags() const override
outliner::InstrType getOutliningTypeImpl(const MachineModuleInfo &MMI, MachineBasicBlock::iterator &MBBI, unsigned Flags) const override
MachineTraceStrategy getMachineCombinerTraceStrategy() const override
unsigned getInstSizeInBytes(const MachineInstr &MI) const override
std::optional< RegImmPair > isAddImmediate(const MachineInstr &MI, Register Reg) const override
bool reverseBranchCondition(SmallVectorImpl< MachineOperand > &Cond) const override
ArrayRef< std::pair< MachineMemOperand::Flags, const char * > > getSerializableMachineMemOperandTargetFlags() const override
MCInst getNop() const override
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
bool isMBBSafeToOutlineFrom(MachineBasicBlock &MBB, unsigned &Flags) const override
bool getMemOperandsWithOffsetWidth(const MachineInstr &MI, SmallVectorImpl< const MachineOperand * > &BaseOps, int64_t &Offset, bool &OffsetIsScalable, LocationSize &Width, const TargetRegisterInfo *TRI) const override
void buildOutlinedFrame(MachineBasicBlock &MBB, MachineFunction &MF, const outliner::OutlinedFunction &OF) const override
void finalizeInsInstrs(MachineInstr &Root, unsigned &Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs) const override
std::pair< unsigned, unsigned > decomposeMachineOperandsTargetFlags(unsigned TF) const override
MachineInstr * commuteInstructionImpl(MachineInstr &MI, bool NewMI, unsigned OpIdx1, unsigned OpIdx2) const override
bool hasReassociableOperands(const MachineInstr &Inst, const MachineBasicBlock *MBB) const override
MachineBasicBlock * getBranchDestBlock(const MachineInstr &MI) const override
std::string createMIROperandComment(const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx, const TargetRegisterInfo *TRI) const override
bool shouldOutlineFromFunctionByDefault(MachineFunction &MF) const override
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator MBBI, const DebugLoc &DL, Register DstReg, Register SrcReg, bool KillSrc, bool RenamableDest=false, bool RenamableSrc=false) const override
bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, unsigned &SrcOpIdx2) const override
bool analyzeBranch(MachineBasicBlock &MBB, MachineBasicBlock *&TBB, MachineBasicBlock *&FBB, SmallVectorImpl< MachineOperand > &Cond, bool AllowModify) const override
MachineBasicBlock::iterator insertOutlinedCall(Module &M, MachineBasicBlock &MBB, MachineBasicBlock::iterator &It, MachineFunction &MF, outliner::Candidate &C) const override
bool isBranchOffsetInRange(unsigned BranchOpc, int64_t BrOffset) const override
static RISCVCC::CondCode getCondFromBranchOpc(unsigned Opc)
bool isAssociativeAndCommutative(const MachineInstr &Inst, bool Invert) const override
CombinerObjective getCombinerObjective(unsigned Pattern) const override
bool isHighLatencyDef(int Opc) const override
static bool evaluateCondBranch(RISCVCC::CondCode CC, int64_t C0, int64_t C1)
Return the result of the evaluation of C0 CC C1, where CC is a RISCVCC::CondCode.
bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const override
bool optimizeCondBranch(MachineInstr &MI) const override
std::optional< DestSourcePair > isCopyInstrImpl(const MachineInstr &MI) const override
bool analyzeSelect(const MachineInstr &MI, SmallVectorImpl< MachineOperand > &Cond, unsigned &TrueOp, unsigned &FalseOp, bool &Optimizable) const override
static bool isFromLoadImm(const MachineRegisterInfo &MRI, const MachineOperand &Op, int64_t &Imm)
Return true if the operand is a load immediate instruction and sets Imm to the immediate value.
bool shouldClusterMemOps(ArrayRef< const MachineOperand * > BaseOps1, int64_t Offset1, bool OffsetIsScalable1, ArrayRef< const MachineOperand * > BaseOps2, int64_t Offset2, bool OffsetIsScalable2, unsigned ClusterSize, unsigned NumBytes) const override
bool areMemAccessesTriviallyDisjoint(const MachineInstr &MIa, const MachineInstr &MIb) const override
RISCVMachineFunctionInfo - This class is derived from MachineFunctionInfo and contains private RISCV-...
const RISCVRegisterInfo * getRegisterInfo() const override
Wrapper class representing virtual and physical registers.
Definition Register.h:20
constexpr bool isValid() const
Definition Register.h:112
constexpr bool isVirtual() const
Return true if the specified register number is in the virtual register namespace.
Definition Register.h:79
SlotIndex - An opaque wrapper around machine indexes.
Definition SlotIndexes.h:66
SlotIndex getRegSlot(bool EC=false) const
Returns the register use/def slot in the current instruction for a normal or early-clobber def.
A templated base class for SmallPtrSet which provides the typesafe interface that is common across al...
bool erase(PtrType Ptr)
Remove pointer from the set.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
MI-level stackmap operands.
Definition StackMaps.h:36
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given stackmap should emit.
Definition StackMaps.h:51
MI-level Statepoint operands.
Definition StackMaps.h:159
uint32_t getNumPatchBytes() const
Return the number of patchable bytes the given statepoint should emit.
Definition StackMaps.h:208
StringRef - Represent a constant reference to a string, i.e.
Definition StringRef.h:55
Object returned by analyzeLoopForPipelining.
TargetInstrInfo - Interface to description of machine instruction set.
virtual bool findCommutedOpIndices(const MachineInstr &MI, unsigned &SrcOpIdx1, unsigned &SrcOpIdx2) const
Returns true iff the routine could find two commutable operands in the given machine instruction.
virtual bool hasReassociableOperands(const MachineInstr &Inst, const MachineBasicBlock *MBB) const
Return true when \P Inst has reassociable operands in the same \P MBB.
virtual void genAlternativeCodeSequence(MachineInstr &Root, unsigned Pattern, SmallVectorImpl< MachineInstr * > &InsInstrs, SmallVectorImpl< MachineInstr * > &DelInstrs, DenseMap< Register, unsigned > &InstIdxForVirtReg) const
When getMachineCombinerPatterns() finds patterns, this function generates the instructions that could...
virtual bool getMachineCombinerPatterns(MachineInstr &Root, SmallVectorImpl< unsigned > &Patterns, bool DoRegPressureReduce) const
Return true when there is potentially a faster code sequence for an instruction chain ending in Root.
virtual bool isReMaterializableImpl(const MachineInstr &MI) const
For instructions with opcodes for which the M_REMATERIALIZABLE flag is set, this hook lets the target...
virtual bool isMBBSafeToOutlineFrom(MachineBasicBlock &MBB, unsigned &Flags) const
Optional target hook that returns true if MBB is safe to outline from, and returns any target-specifi...
virtual void getReassociateOperandIndices(const MachineInstr &Root, unsigned Pattern, std::array< unsigned, 5 > &OperandIndices) const
The returned array encodes the operand index for each parameter because the operands may be commuted;...
virtual CombinerObjective getCombinerObjective(unsigned Pattern) const
Return the objective of a combiner pattern.
virtual MachineInstr * commuteInstructionImpl(MachineInstr &MI, bool NewMI, unsigned OpIdx1, unsigned OpIdx2) const
This method commutes the operands of the given machine instruction MI.
virtual bool hasReassociableSibling(const MachineInstr &Inst, bool &Commuted) const
Return true when \P Inst has reassociable sibling.
virtual std::string createMIROperandComment(const MachineInstr &MI, const MachineOperand &Op, unsigned OpIdx, const TargetRegisterInfo *TRI) const
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
const uint8_t TSFlags
Configurable target specific flags.
TargetRegisterInfo base class - We assume that the target defines a static array of TargetRegisterDes...
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const TargetInstrInfo * getInstrInfo() const
virtual const TargetRegisterInfo * getRegisterInfo() const =0
Return the target's register information.
Target - Wrapper for Target specific information.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition TypeSize.h:343
static constexpr TypeSize getZero()
Definition TypeSize.h:349
static constexpr TypeSize getScalable(ScalarTy MinimumSize)
Definition TypeSize.h:346
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
CondCode getInverseBranchCondition(CondCode)
unsigned getBrCond(CondCode CC, unsigned SelectOpc=0)
static bool isValidRoundingMode(unsigned Mode)
static unsigned getVecPolicyOpNum(const MCInstrDesc &Desc)
static bool usesMaskPolicy(uint64_t TSFlags)
static bool hasRoundModeOp(uint64_t TSFlags)
static unsigned getVLOpNum(const MCInstrDesc &Desc)
static bool hasVLOp(uint64_t TSFlags)
static MCRegister getTailExpandUseRegNo(const FeatureBitset &FeatureBits)
static int getFRMOpNum(const MCInstrDesc &Desc)
static bool hasVecPolicyOp(uint64_t TSFlags)
static bool usesVXRM(uint64_t TSFlags)
static bool isRVVWideningReduction(uint64_t TSFlags)
static unsigned getSEWOpNum(const MCInstrDesc &Desc)
static bool hasSEWOp(uint64_t TSFlags)
static bool isFirstDefTiedToFirstUse(const MCInstrDesc &Desc)
InstSeq generateInstSeq(int64_t Val, const MCSubtargetInfo &STI)
SmallVector< Inst, 8 > InstSeq
Definition RISCVMatInt.h:43
@ OPERAND_SIMM10_LSB0000_NONZERO
static unsigned getNF(uint8_t TSFlags)
static RISCVVType::VLMUL getLMul(uint8_t TSFlags)
static bool isTailAgnostic(unsigned VType)
LLVM_ABI void printXSfmmVType(unsigned VType, raw_ostream &OS)
LLVM_ABI std::pair< unsigned, bool > decodeVLMUL(VLMUL VLMul)
static bool isValidSEW(unsigned SEW)
LLVM_ABI void printVType(unsigned VType, raw_ostream &OS)
static bool isValidXSfmmVType(unsigned VTypeI)
static unsigned getSEW(unsigned VType)
static VLMUL getVLMUL(unsigned VType)
bool hasEqualFRM(const MachineInstr &MI1, const MachineInstr &MI2)
bool isVLKnownLE(const MachineOperand &LHS, const MachineOperand &RHS)
Given two VL operands, do we know that LHS <= RHS?
unsigned getRVVMCOpcode(unsigned RVVPseudoOpcode)
unsigned getDestLog2EEW(const MCInstrDesc &Desc, unsigned Log2SEW)
std::optional< unsigned > getVectorLowDemandedScalarBits(unsigned Opcode, unsigned Log2SEW)
std::optional< std::pair< unsigned, unsigned > > isRVVSpillForZvlsseg(unsigned Opcode)
static constexpr unsigned RVVBitsPerBlock
bool isRVVSpill(const MachineInstr &MI)
static constexpr unsigned RVVBytesPerBlock
static constexpr int64_t VLMaxSentinel
bool isVectorCopy(const TargetRegisterInfo *TRI, const MachineInstr &MI)
Return true if MI is a copy that will be lowered to one or more vmvNr.vs.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Dead
Unused definition.
@ Define
Register definition.
@ Kill
The last use of a register.
@ Undef
Value of the register doesn't matter.
ValuesClass values(OptsTy... Options)
Helper to build a ValuesClass by forwarding a variable number of arguments as an initializer list to ...
initializer< Ty > init(const Ty &Val)
InstrType
Represents how an instruction should be mapped by the outliner.
This is an optimization pass for GlobalISel generic memory operations.
Definition Types.h:26
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
Definition STLExtras.h:316
constexpr unsigned getRenamableRegState(bool B)
@ Offset
Definition DWP.cpp:532
constexpr unsigned getDeadRegState(bool B)
@ SHXADD_ADD_SLLI_OP2
@ SHXADD_ADD_SLLI_OP1
MachineTraceStrategy
Strategies for selecting traces.
@ TS_MinInstrCount
Select the trace through a block that has the fewest instructions.
@ TS_Local
Select the trace that contains only the current basic block.
bool all_of(R &&range, UnaryPredicate P)
Provide wrappers to std::all_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1737
static const MachineMemOperand::Flags MONontemporalBit1
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr unsigned getKillRegState(bool B)
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
Definition MathExtras.h:165
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
Definition STLExtras.h:2544
bool isValidAtomicOrdering(Int I)
static const MachineMemOperand::Flags MONontemporalBit0
Op::Description Desc
constexpr bool has_single_bit(T Value) noexcept
Definition bit.h:147
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1744
unsigned Log2_32(uint32_t Value)
Return the floor log base 2 of the specified value, -1 if the value is zero.
Definition MathExtras.h:331
MachineInstr * getImm(const MachineOperand &MO, const MachineRegisterInfo *MRI)
decltype(auto) get(const PointerIntPair< PointerTy, IntBits, IntType, PtrTraits, Info > &Pair)
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
Definition Error.cpp:163
CombinerObjective
The combiner's goal may differ based on which pattern it is attempting to optimize.
constexpr bool isUInt(uint64_t x)
Checks if an unsigned integer fits into the given bit width.
Definition MathExtras.h:189
CodeGenOptLevel
Code generation optimization level.
Definition CodeGen.h:82
int isShifted359(T Value, int &Shift)
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
constexpr unsigned getDefRegState(bool B)
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
constexpr bool isShiftedInt(int64_t x)
Checks if a signed integer is an N bit number shifted left by S.
Definition MathExtras.h:182
void erase_if(Container &C, UnaryPredicate P)
Provide a container algorithm similar to C++ Library Fundamentals v2's erase_if which is equivalent t...
Definition STLExtras.h:2182
constexpr int64_t SignExtend64(uint64_t x)
Sign-extend the number in the bottom B bits of X to a 64-bit integer.
Definition MathExtras.h:572
LLVM_ABI const Value * getUnderlyingObject(const Value *V, unsigned MaxLookup=MaxLookupSearchDepth)
This method strips off any GEP address adjustments, pointer casts or llvm.threadlocal....
constexpr bool isShiftedUInt(uint64_t x)
Checks if a unsigned integer is an N bit number shifted left by S.
Definition MathExtras.h:198
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:872
#define N
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
Used to describe addressing mode similar to ExtAddrMode in CodeGenPrepare.
This represents a simple continuous liveness interval for a value.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
static bool isRVVRegClass(const TargetRegisterClass *RC)
Used to describe a register and immediate addition.
An individual sequence of instructions to be replaced with a call to an outlined function.
MachineFunction * getMF() const
The information necessary to create an outlined function for some class of candidate.