LLVM 23.0.0git
IRTranslator.cpp
Go to the documentation of this file.
1//===- llvm/CodeGen/GlobalISel/IRTranslator.cpp - IRTranslator ---*- C++ -*-==//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8/// \file
9/// This file implements the IRTranslator class.
10//===----------------------------------------------------------------------===//
11
14#include "llvm/ADT/STLExtras.h"
15#include "llvm/ADT/ScopeExit.h"
20#include "llvm/Analysis/Loads.h"
50#include "llvm/IR/BasicBlock.h"
51#include "llvm/IR/CFG.h"
52#include "llvm/IR/Constant.h"
53#include "llvm/IR/Constants.h"
54#include "llvm/IR/DataLayout.h"
57#include "llvm/IR/Function.h"
59#include "llvm/IR/InlineAsm.h"
60#include "llvm/IR/InstrTypes.h"
63#include "llvm/IR/Intrinsics.h"
64#include "llvm/IR/IntrinsicsAMDGPU.h"
65#include "llvm/IR/LLVMContext.h"
66#include "llvm/IR/Metadata.h"
68#include "llvm/IR/Statepoint.h"
69#include "llvm/IR/Type.h"
70#include "llvm/IR/User.h"
71#include "llvm/IR/Value.h"
73#include "llvm/MC/MCContext.h"
74#include "llvm/Pass.h"
77#include "llvm/Support/Debug.h"
84#include <algorithm>
85#include <cassert>
86#include <cstdint>
87#include <iterator>
88#include <optional>
89#include <string>
90#include <utility>
91#include <vector>
92
93#define DEBUG_TYPE "irtranslator"
94
95using namespace llvm;
96
97static cl::opt<bool>
98 EnableCSEInIRTranslator("enable-cse-in-irtranslator",
99 cl::desc("Should enable CSE in irtranslator"),
100 cl::Optional, cl::init(false));
101char IRTranslator::ID = 0;
102
103INITIALIZE_PASS_BEGIN(IRTranslator, DEBUG_TYPE, "IRTranslator LLVM IR -> MI",
104 false, false)
110INITIALIZE_PASS_END(IRTranslator, DEBUG_TYPE, "IRTranslator LLVM IR -> MI",
112
116 MF.getProperties().setFailedISel();
117 bool IsGlobalISelAbortEnabled =
118 MF.getTarget().Options.GlobalISelAbort == GlobalISelAbortMode::Enable;
119
120 // Print the function name explicitly if we don't have a debug location (which
121 // makes the diagnostic less useful) or if we're going to emit a raw error.
122 if (!R.getLocation().isValid() || IsGlobalISelAbortEnabled)
123 R << (" (in function: " + MF.getName() + ")").str();
124
125 if (IsGlobalISelAbortEnabled)
126 report_fatal_error(Twine(R.getMsg()));
127 else
128 ORE.emit(R);
129}
130
132 : MachineFunctionPass(ID), OptLevel(optlevel) {}
133
134#ifndef NDEBUG
135namespace {
136/// Verify that every instruction created has the same DILocation as the
137/// instruction being translated.
138class DILocationVerifier : public GISelChangeObserver {
139 const Instruction *CurrInst = nullptr;
140
141public:
142 DILocationVerifier() = default;
143 ~DILocationVerifier() override = default;
144
145 const Instruction *getCurrentInst() const { return CurrInst; }
146 void setCurrentInst(const Instruction *Inst) { CurrInst = Inst; }
147
148 void erasingInstr(MachineInstr &MI) override {}
149 void changingInstr(MachineInstr &MI) override {}
150 void changedInstr(MachineInstr &MI) override {}
151
152 void createdInstr(MachineInstr &MI) override {
153 assert(getCurrentInst() && "Inserted instruction without a current MI");
154
155 // Only print the check message if we're actually checking it.
156#ifndef NDEBUG
157 LLVM_DEBUG(dbgs() << "Checking DILocation from " << *CurrInst
158 << " was copied to " << MI);
159#endif
160 // We allow insts in the entry block to have no debug loc because
161 // they could have originated from constants, and we don't want a jumpy
162 // debug experience.
163 assert((CurrInst->getDebugLoc() == MI.getDebugLoc() ||
164 (MI.getParent()->isEntryBlock() && !MI.getDebugLoc()) ||
165 (MI.isDebugInstr())) &&
166 "Line info was not transferred to all instructions");
167 }
168};
169} // namespace
170#endif // ifndef NDEBUG
171
172
189
190IRTranslator::ValueToVRegInfo::VRegListT &
191IRTranslator::allocateVRegs(const Value &Val) {
192 auto VRegsIt = VMap.findVRegs(Val);
193 if (VRegsIt != VMap.vregs_end())
194 return *VRegsIt->second;
195 auto *Regs = VMap.getVRegs(Val);
196 auto *Offsets = VMap.getOffsets(Val);
197 SmallVector<LLT, 4> SplitTys;
198 computeValueLLTs(*DL, *Val.getType(), SplitTys,
199 Offsets->empty() ? Offsets : nullptr);
200 for (unsigned i = 0; i < SplitTys.size(); ++i)
201 Regs->push_back(0);
202 return *Regs;
203}
204
205ArrayRef<Register> IRTranslator::getOrCreateVRegs(const Value &Val) {
206 auto VRegsIt = VMap.findVRegs(Val);
207 if (VRegsIt != VMap.vregs_end())
208 return *VRegsIt->second;
209
210 if (Val.getType()->isVoidTy())
211 return *VMap.getVRegs(Val);
212
213 // Create entry for this type.
214 auto *VRegs = VMap.getVRegs(Val);
215 auto *Offsets = VMap.getOffsets(Val);
216
217 if (!Val.getType()->isTokenTy())
218 assert(Val.getType()->isSized() &&
219 "Don't know how to create an empty vreg");
220
221 // Fast-path values that lower to a single vreg.
222 if (!Val.getType()->isAggregateType()) {
223 LLT Ty = getLLTForType(*Val.getType(), *DL);
224 if (Offsets->empty())
225 Offsets->push_back(0);
226 VRegs->push_back(MRI->createGenericVirtualRegister(Ty));
227 if (isa<Constant>(Val)) {
228 bool Success = translate(cast<Constant>(Val), VRegs->front());
229 if (!Success) {
230 OptimizationRemarkMissed R("gisel-irtranslator", "GISelFailure",
232 &MF->getFunction().getEntryBlock());
233 R << "unable to translate constant: " << ore::NV("Type", Val.getType());
234 reportTranslationError(*MF, *ORE, R);
235 }
236 }
237 return *VRegs;
238 }
239
240 SmallVector<LLT, 4> SplitTys;
241 computeValueLLTs(*DL, *Val.getType(), SplitTys,
242 Offsets->empty() ? Offsets : nullptr);
243
244 if (!isa<Constant>(Val)) {
245 for (auto Ty : SplitTys)
246 VRegs->push_back(MRI->createGenericVirtualRegister(Ty));
247 return *VRegs;
248 }
249
250 // UndefValue, ConstantAggregateZero
251 auto &C = cast<Constant>(Val);
252 unsigned Idx = 0;
253 while (auto Elt = C.getAggregateElement(Idx++)) {
254 auto EltRegs = getOrCreateVRegs(*Elt);
255 llvm::append_range(*VRegs, EltRegs);
256 }
257
258 return *VRegs;
259}
260
261int IRTranslator::getOrCreateFrameIndex(const AllocaInst &AI) {
262 auto [MapEntry, Inserted] = FrameIndices.try_emplace(&AI);
263 if (!Inserted)
264 return MapEntry->second;
265
266 TypeSize TySize = AI.getAllocationSize(*DL).value_or(TypeSize::getZero());
267 uint64_t Size = TySize.getKnownMinValue();
268
269 // Always allocate at least one byte.
270 Size = std::max<uint64_t>(Size, 1u);
271
272 int &FI = MapEntry->second;
273 FI = MF->getFrameInfo().CreateStackObject(Size, AI.getAlign(), false, &AI);
274
275 // Scalable vectors and structures that contain scalable vectors may
276 // need a special StackID to distinguish them from other (fixed size)
277 // stack objects.
278 if (TySize.isScalable()) {
279 auto StackID =
280 MF->getSubtarget().getFrameLowering()->getStackIDForScalableVectors();
281 MF->getFrameInfo().setStackID(FI, StackID);
282 }
283
284 return FI;
285}
286
287Align IRTranslator::getMemOpAlign(const Instruction &I) {
288 if (const StoreInst *SI = dyn_cast<StoreInst>(&I))
289 return SI->getAlign();
290 if (const LoadInst *LI = dyn_cast<LoadInst>(&I))
291 return LI->getAlign();
292 if (const AtomicCmpXchgInst *AI = dyn_cast<AtomicCmpXchgInst>(&I))
293 return AI->getAlign();
294 if (const AtomicRMWInst *AI = dyn_cast<AtomicRMWInst>(&I))
295 return AI->getAlign();
296
297 OptimizationRemarkMissed R("gisel-irtranslator", "", &I);
298 R << "unable to translate memop: " << ore::NV("Opcode", &I);
299 reportTranslationError(*MF, *ORE, R);
300 return Align(1);
301}
302
303MachineBasicBlock &IRTranslator::getMBB(const BasicBlock &BB) {
304 MachineBasicBlock *MBB = FuncInfo.getMBB(&BB);
305 assert(MBB && "BasicBlock was not encountered before");
306 return *MBB;
307}
308
309void IRTranslator::addMachineCFGPred(CFGEdge Edge, MachineBasicBlock *NewPred) {
310 assert(NewPred && "new predecessor must be a real MachineBasicBlock");
311 MachinePreds[Edge].push_back(NewPred);
312}
313
314bool IRTranslator::translateBinaryOp(unsigned Opcode, const User &U,
315 MachineIRBuilder &MIRBuilder) {
316 if (!mayTranslateUserTypes(U))
317 return false;
318
319 // Get or create a virtual register for each value.
320 // Unless the value is a Constant => loadimm cst?
321 // or inline constant each time?
322 // Creation of a virtual register needs to have a size.
323 Register Op0 = getOrCreateVReg(*U.getOperand(0));
324 Register Op1 = getOrCreateVReg(*U.getOperand(1));
325 Register Res = getOrCreateVReg(U);
326 uint32_t Flags = 0;
327 if (isa<Instruction>(U)) {
328 const Instruction &I = cast<Instruction>(U);
330 }
331
332 MIRBuilder.buildInstr(Opcode, {Res}, {Op0, Op1}, Flags);
333 return true;
334}
335
336bool IRTranslator::translateUnaryOp(unsigned Opcode, const User &U,
337 MachineIRBuilder &MIRBuilder) {
338 if (!mayTranslateUserTypes(U))
339 return false;
340
341 Register Op0 = getOrCreateVReg(*U.getOperand(0));
342 Register Res = getOrCreateVReg(U);
343 uint32_t Flags = 0;
344 if (isa<Instruction>(U)) {
345 const Instruction &I = cast<Instruction>(U);
347 }
348 MIRBuilder.buildInstr(Opcode, {Res}, {Op0}, Flags);
349 return true;
350}
351
352bool IRTranslator::translateFNeg(const User &U, MachineIRBuilder &MIRBuilder) {
353 return translateUnaryOp(TargetOpcode::G_FNEG, U, MIRBuilder);
354}
355
356bool IRTranslator::translateCompare(const User &U,
357 MachineIRBuilder &MIRBuilder) {
358 if (!mayTranslateUserTypes(U))
359 return false;
360
361 auto *CI = cast<CmpInst>(&U);
362 Register Op0 = getOrCreateVReg(*U.getOperand(0));
363 Register Op1 = getOrCreateVReg(*U.getOperand(1));
364 Register Res = getOrCreateVReg(U);
365 CmpInst::Predicate Pred = CI->getPredicate();
367 if (CmpInst::isIntPredicate(Pred))
368 MIRBuilder.buildICmp(Pred, Res, Op0, Op1, Flags);
369 else if (Pred == CmpInst::FCMP_FALSE)
370 MIRBuilder.buildCopy(
371 Res, getOrCreateVReg(*Constant::getNullValue(U.getType())));
372 else if (Pred == CmpInst::FCMP_TRUE)
373 MIRBuilder.buildCopy(
374 Res, getOrCreateVReg(*Constant::getAllOnesValue(U.getType())));
375 else
376 MIRBuilder.buildFCmp(Pred, Res, Op0, Op1, Flags);
377
378 return true;
379}
380
381bool IRTranslator::translateRet(const User &U, MachineIRBuilder &MIRBuilder) {
382 const ReturnInst &RI = cast<ReturnInst>(U);
383 const Value *Ret = RI.getReturnValue();
384 if (Ret && DL->getTypeStoreSize(Ret->getType()).isZero())
385 Ret = nullptr;
386
387 ArrayRef<Register> VRegs;
388 if (Ret)
389 VRegs = getOrCreateVRegs(*Ret);
390
391 Register SwiftErrorVReg = 0;
392 if (CLI->supportSwiftError() && SwiftError.getFunctionArg()) {
393 SwiftErrorVReg = SwiftError.getOrCreateVRegUseAt(
394 &RI, &MIRBuilder.getMBB(), SwiftError.getFunctionArg());
395 }
396
397 // The target may mess up with the insertion point, but
398 // this is not important as a return is the last instruction
399 // of the block anyway.
400 return CLI->lowerReturn(MIRBuilder, Ret, VRegs, FuncInfo, SwiftErrorVReg);
401}
402
403void IRTranslator::emitBranchForMergedCondition(
405 MachineBasicBlock *CurBB, MachineBasicBlock *SwitchBB,
406 BranchProbability TProb, BranchProbability FProb, bool InvertCond) {
407 // If the leaf of the tree is a comparison, merge the condition into
408 // the caseblock.
409 if (const CmpInst *BOp = dyn_cast<CmpInst>(Cond)) {
410 CmpInst::Predicate Condition;
411 if (const ICmpInst *IC = dyn_cast<ICmpInst>(Cond)) {
412 Condition = InvertCond ? IC->getInversePredicate() : IC->getPredicate();
413 } else {
414 const FCmpInst *FC = cast<FCmpInst>(Cond);
415 Condition = InvertCond ? FC->getInversePredicate() : FC->getPredicate();
416 }
417
418 SwitchCG::CaseBlock CB(Condition, false, BOp->getOperand(0),
419 BOp->getOperand(1), nullptr, TBB, FBB, CurBB,
420 CurBuilder->getDebugLoc(), TProb, FProb);
421 SL->SwitchCases.push_back(CB);
422 return;
423 }
424
425 // Create a CaseBlock record representing this branch.
427 SwitchCG::CaseBlock CB(
428 Pred, false, Cond, ConstantInt::getTrue(MF->getFunction().getContext()),
429 nullptr, TBB, FBB, CurBB, CurBuilder->getDebugLoc(), TProb, FProb);
430 SL->SwitchCases.push_back(CB);
431}
432
433static bool isValInBlock(const Value *V, const BasicBlock *BB) {
434 if (const Instruction *I = dyn_cast<Instruction>(V))
435 return I->getParent() == BB;
436 return true;
437}
438
439void IRTranslator::findMergedConditions(
441 MachineBasicBlock *CurBB, MachineBasicBlock *SwitchBB,
443 BranchProbability FProb, bool InvertCond) {
444 using namespace PatternMatch;
445 assert((Opc == Instruction::And || Opc == Instruction::Or) &&
446 "Expected Opc to be AND/OR");
447 // Skip over not part of the tree and remember to invert op and operands at
448 // next level.
449 Value *NotCond;
450 if (match(Cond, m_OneUse(m_Not(m_Value(NotCond)))) &&
451 isValInBlock(NotCond, CurBB->getBasicBlock())) {
452 findMergedConditions(NotCond, TBB, FBB, CurBB, SwitchBB, Opc, TProb, FProb,
453 !InvertCond);
454 return;
455 }
456
458 const Value *BOpOp0, *BOpOp1;
459 // Compute the effective opcode for Cond, taking into account whether it needs
460 // to be inverted, e.g.
461 // and (not (or A, B)), C
462 // gets lowered as
463 // and (and (not A, not B), C)
465 if (BOp) {
466 BOpc = match(BOp, m_LogicalAnd(m_Value(BOpOp0), m_Value(BOpOp1)))
467 ? Instruction::And
468 : (match(BOp, m_LogicalOr(m_Value(BOpOp0), m_Value(BOpOp1)))
469 ? Instruction::Or
471 if (InvertCond) {
472 if (BOpc == Instruction::And)
473 BOpc = Instruction::Or;
474 else if (BOpc == Instruction::Or)
475 BOpc = Instruction::And;
476 }
477 }
478
479 // If this node is not part of the or/and tree, emit it as a branch.
480 // Note that all nodes in the tree should have same opcode.
481 bool BOpIsInOrAndTree = BOpc && BOpc == Opc && BOp->hasOneUse();
482 if (!BOpIsInOrAndTree || BOp->getParent() != CurBB->getBasicBlock() ||
483 !isValInBlock(BOpOp0, CurBB->getBasicBlock()) ||
484 !isValInBlock(BOpOp1, CurBB->getBasicBlock())) {
485 emitBranchForMergedCondition(Cond, TBB, FBB, CurBB, SwitchBB, TProb, FProb,
486 InvertCond);
487 return;
488 }
489
490 // Create TmpBB after CurBB.
491 MachineFunction::iterator BBI(CurBB);
492 MachineBasicBlock *TmpBB =
493 MF->CreateMachineBasicBlock(CurBB->getBasicBlock());
494 CurBB->getParent()->insert(++BBI, TmpBB);
495
496 if (Opc == Instruction::Or) {
497 // Codegen X | Y as:
498 // BB1:
499 // jmp_if_X TBB
500 // jmp TmpBB
501 // TmpBB:
502 // jmp_if_Y TBB
503 // jmp FBB
504 //
505
506 // We have flexibility in setting Prob for BB1 and Prob for TmpBB.
507 // The requirement is that
508 // TrueProb for BB1 + (FalseProb for BB1 * TrueProb for TmpBB)
509 // = TrueProb for original BB.
510 // Assuming the original probabilities are A and B, one choice is to set
511 // BB1's probabilities to A/2 and A/2+B, and set TmpBB's probabilities to
512 // A/(1+B) and 2B/(1+B). This choice assumes that
513 // TrueProb for BB1 == FalseProb for BB1 * TrueProb for TmpBB.
514 // Another choice is to assume TrueProb for BB1 equals to TrueProb for
515 // TmpBB, but the math is more complicated.
516
517 auto NewTrueProb = TProb / 2;
518 auto NewFalseProb = TProb / 2 + FProb;
519 // Emit the LHS condition.
520 findMergedConditions(BOpOp0, TBB, TmpBB, CurBB, SwitchBB, Opc, NewTrueProb,
521 NewFalseProb, InvertCond);
522
523 // Normalize A/2 and B to get A/(1+B) and 2B/(1+B).
524 SmallVector<BranchProbability, 2> Probs{TProb / 2, FProb};
526 // Emit the RHS condition into TmpBB.
527 findMergedConditions(BOpOp1, TBB, FBB, TmpBB, SwitchBB, Opc, Probs[0],
528 Probs[1], InvertCond);
529 } else {
530 assert(Opc == Instruction::And && "Unknown merge op!");
531 // Codegen X & Y as:
532 // BB1:
533 // jmp_if_X TmpBB
534 // jmp FBB
535 // TmpBB:
536 // jmp_if_Y TBB
537 // jmp FBB
538 //
539 // This requires creation of TmpBB after CurBB.
540
541 // We have flexibility in setting Prob for BB1 and Prob for TmpBB.
542 // The requirement is that
543 // FalseProb for BB1 + (TrueProb for BB1 * FalseProb for TmpBB)
544 // = FalseProb for original BB.
545 // Assuming the original probabilities are A and B, one choice is to set
546 // BB1's probabilities to A+B/2 and B/2, and set TmpBB's probabilities to
547 // 2A/(1+A) and B/(1+A). This choice assumes that FalseProb for BB1 ==
548 // TrueProb for BB1 * FalseProb for TmpBB.
549
550 auto NewTrueProb = TProb + FProb / 2;
551 auto NewFalseProb = FProb / 2;
552 // Emit the LHS condition.
553 findMergedConditions(BOpOp0, TmpBB, FBB, CurBB, SwitchBB, Opc, NewTrueProb,
554 NewFalseProb, InvertCond);
555
556 // Normalize A and B/2 to get 2A/(1+A) and B/(1+A).
557 SmallVector<BranchProbability, 2> Probs{TProb, FProb / 2};
559 // Emit the RHS condition into TmpBB.
560 findMergedConditions(BOpOp1, TBB, FBB, TmpBB, SwitchBB, Opc, Probs[0],
561 Probs[1], InvertCond);
562 }
563}
564
565bool IRTranslator::shouldEmitAsBranches(
566 const std::vector<SwitchCG::CaseBlock> &Cases) {
567 // For multiple cases, it's better to emit as branches.
568 if (Cases.size() != 2)
569 return true;
570
571 // If this is two comparisons of the same values or'd or and'd together, they
572 // will get folded into a single comparison, so don't emit two blocks.
573 if ((Cases[0].CmpLHS == Cases[1].CmpLHS &&
574 Cases[0].CmpRHS == Cases[1].CmpRHS) ||
575 (Cases[0].CmpRHS == Cases[1].CmpLHS &&
576 Cases[0].CmpLHS == Cases[1].CmpRHS)) {
577 return false;
578 }
579
580 // Handle: (X != null) | (Y != null) --> (X|Y) != 0
581 // Handle: (X == null) & (Y == null) --> (X|Y) == 0
582 if (Cases[0].CmpRHS == Cases[1].CmpRHS &&
583 Cases[0].PredInfo.Pred == Cases[1].PredInfo.Pred &&
584 isa<Constant>(Cases[0].CmpRHS) &&
585 cast<Constant>(Cases[0].CmpRHS)->isNullValue()) {
586 if (Cases[0].PredInfo.Pred == CmpInst::ICMP_EQ &&
587 Cases[0].TrueBB == Cases[1].ThisBB)
588 return false;
589 if (Cases[0].PredInfo.Pred == CmpInst::ICMP_NE &&
590 Cases[0].FalseBB == Cases[1].ThisBB)
591 return false;
592 }
593
594 return true;
595}
596
597bool IRTranslator::translateUncondBr(const User &U,
598 MachineIRBuilder &MIRBuilder) {
599 const UncondBrInst &BrInst = cast<UncondBrInst>(U);
600 auto &CurMBB = MIRBuilder.getMBB();
601 auto *Succ0MBB = &getMBB(*BrInst.getSuccessor(0));
602
603 // If the unconditional target is the layout successor, fallthrough.
604 if (OptLevel == CodeGenOptLevel::None || !CurMBB.isLayoutSuccessor(Succ0MBB))
605 MIRBuilder.buildBr(*Succ0MBB);
606
607 // Link successors.
608 for (const BasicBlock *Succ : successors(&BrInst))
609 CurMBB.addSuccessor(&getMBB(*Succ));
610 return true;
611}
612
613bool IRTranslator::translateCondBr(const User &U,
614 MachineIRBuilder &MIRBuilder) {
615 const CondBrInst &BrInst = cast<CondBrInst>(U);
616 auto &CurMBB = MIRBuilder.getMBB();
617 auto *Succ0MBB = &getMBB(*BrInst.getSuccessor(0));
618
619 // If this condition is one of the special cases we handle, do special stuff
620 // now.
621 const Value *CondVal = BrInst.getCondition();
622 MachineBasicBlock *Succ1MBB = &getMBB(*BrInst.getSuccessor(1));
623
624 // If this is a series of conditions that are or'd or and'd together, emit
625 // this as a sequence of branches instead of setcc's with and/or operations.
626 // As long as jumps are not expensive (exceptions for multi-use logic ops,
627 // unpredictable branches, and vector extracts because those jumps are likely
628 // expensive for any target), this should improve performance.
629 // For example, instead of something like:
630 // cmp A, B
631 // C = seteq
632 // cmp D, E
633 // F = setle
634 // or C, F
635 // jnz foo
636 // Emit:
637 // cmp A, B
638 // je foo
639 // cmp D, E
640 // jle foo
641 using namespace PatternMatch;
642 const Instruction *CondI = dyn_cast<Instruction>(CondVal);
643 if (!TLI->isJumpExpensive() && CondI && CondI->hasOneUse() &&
644 !BrInst.hasMetadata(LLVMContext::MD_unpredictable)) {
646 Value *Vec;
647 const Value *BOp0, *BOp1;
648 if (match(CondI, m_LogicalAnd(m_Value(BOp0), m_Value(BOp1))))
649 Opcode = Instruction::And;
650 else if (match(CondI, m_LogicalOr(m_Value(BOp0), m_Value(BOp1))))
651 Opcode = Instruction::Or;
652
653 if (Opcode && !(match(BOp0, m_ExtractElt(m_Value(Vec), m_Value())) &&
654 match(BOp1, m_ExtractElt(m_Specific(Vec), m_Value())))) {
655 findMergedConditions(CondI, Succ0MBB, Succ1MBB, &CurMBB, &CurMBB, Opcode,
656 getEdgeProbability(&CurMBB, Succ0MBB),
657 getEdgeProbability(&CurMBB, Succ1MBB),
658 /*InvertCond=*/false);
659 assert(SL->SwitchCases[0].ThisBB == &CurMBB && "Unexpected lowering!");
660
661 // Allow some cases to be rejected.
662 if (shouldEmitAsBranches(SL->SwitchCases)) {
663 // Emit the branch for this block.
664 emitSwitchCase(SL->SwitchCases[0], &CurMBB, *CurBuilder);
665 SL->SwitchCases.erase(SL->SwitchCases.begin());
666 return true;
667 }
668
669 // Okay, we decided not to do this, remove any inserted MBB's and clear
670 // SwitchCases.
671 for (unsigned I = 1, E = SL->SwitchCases.size(); I != E; ++I)
672 MF->erase(SL->SwitchCases[I].ThisBB);
673
674 SL->SwitchCases.clear();
675 }
676 }
677
678 // Create a CaseBlock record representing this branch.
679 SwitchCG::CaseBlock CB(CmpInst::ICMP_EQ, false, CondVal,
680 ConstantInt::getTrue(MF->getFunction().getContext()),
681 nullptr, Succ0MBB, Succ1MBB, &CurMBB,
682 CurBuilder->getDebugLoc());
683
684 // Use emitSwitchCase to actually insert the fast branch sequence for this
685 // cond branch.
686 emitSwitchCase(CB, &CurMBB, *CurBuilder);
687 return true;
688}
689
690void IRTranslator::addSuccessorWithProb(MachineBasicBlock *Src,
692 BranchProbability Prob) {
693 if (!FuncInfo.BPI) {
694 Src->addSuccessorWithoutProb(Dst);
695 return;
696 }
697 if (Prob.isUnknown())
698 Prob = getEdgeProbability(Src, Dst);
699 Src->addSuccessor(Dst, Prob);
700}
701
703IRTranslator::getEdgeProbability(const MachineBasicBlock *Src,
704 const MachineBasicBlock *Dst) const {
705 const BasicBlock *SrcBB = Src->getBasicBlock();
706 const BasicBlock *DstBB = Dst->getBasicBlock();
707 if (!FuncInfo.BPI) {
708 // If BPI is not available, set the default probability as 1 / N, where N is
709 // the number of successors.
710 auto SuccSize = std::max<uint32_t>(succ_size(SrcBB), 1);
711 return BranchProbability(1, SuccSize);
712 }
713 return FuncInfo.BPI->getEdgeProbability(SrcBB, DstBB);
714}
715
716bool IRTranslator::translateSwitch(const User &U, MachineIRBuilder &MIB) {
717 using namespace SwitchCG;
718 // Extract cases from the switch.
719 const SwitchInst &SI = cast<SwitchInst>(U);
720 BranchProbabilityInfo *BPI = FuncInfo.BPI;
721 CaseClusterVector Clusters;
722 Clusters.reserve(SI.getNumCases());
723 for (const auto &I : SI.cases()) {
724 MachineBasicBlock *Succ = &getMBB(*I.getCaseSuccessor());
725 assert(Succ && "Could not find successor mbb in mapping");
726 const ConstantInt *CaseVal = I.getCaseValue();
727 BranchProbability Prob =
728 BPI ? BPI->getEdgeProbability(SI.getParent(), I.getSuccessorIndex())
729 : BranchProbability(1, SI.getNumCases() + 1);
730 Clusters.push_back(CaseCluster::range(CaseVal, CaseVal, Succ, Prob));
731 }
732
733 MachineBasicBlock *DefaultMBB = &getMBB(*SI.getDefaultDest());
734
735 // Cluster adjacent cases with the same destination. We do this at all
736 // optimization levels because it's cheap to do and will make codegen faster
737 // if there are many clusters.
738 sortAndRangeify(Clusters);
739
740 MachineBasicBlock *SwitchMBB = &getMBB(*SI.getParent());
741
742 // If there is only the default destination, jump there directly.
743 if (Clusters.empty()) {
744 SwitchMBB->addSuccessor(DefaultMBB);
745 if (DefaultMBB != SwitchMBB->getNextNode())
746 MIB.buildBr(*DefaultMBB);
747 return true;
748 }
749
750 SL->findJumpTables(Clusters, &SI, std::nullopt, DefaultMBB, nullptr, nullptr);
751 SL->findBitTestClusters(Clusters, &SI);
752
753 LLVM_DEBUG({
754 dbgs() << "Case clusters: ";
755 for (const CaseCluster &C : Clusters) {
756 if (C.Kind == CC_JumpTable)
757 dbgs() << "JT:";
758 if (C.Kind == CC_BitTests)
759 dbgs() << "BT:";
760
761 C.Low->getValue().print(dbgs(), true);
762 if (C.Low != C.High) {
763 dbgs() << '-';
764 C.High->getValue().print(dbgs(), true);
765 }
766 dbgs() << ' ';
767 }
768 dbgs() << '\n';
769 });
770
771 assert(!Clusters.empty());
772 SwitchWorkList WorkList;
773 CaseClusterIt First = Clusters.begin();
774 CaseClusterIt Last = Clusters.end() - 1;
775 auto DefaultProb = getEdgeProbability(SwitchMBB, DefaultMBB);
776 WorkList.push_back({SwitchMBB, First, Last, nullptr, nullptr, DefaultProb});
777
778 while (!WorkList.empty()) {
779 SwitchWorkListItem W = WorkList.pop_back_val();
780
781 unsigned NumClusters = W.LastCluster - W.FirstCluster + 1;
782 // For optimized builds, lower large range as a balanced binary tree.
783 if (NumClusters > 3 &&
784 MF->getTarget().getOptLevel() != CodeGenOptLevel::None &&
785 !DefaultMBB->getParent()->getFunction().hasMinSize()) {
786 splitWorkItem(WorkList, W, SI.getCondition(), SwitchMBB, MIB);
787 continue;
788 }
789
790 if (!lowerSwitchWorkItem(W, SI.getCondition(), SwitchMBB, DefaultMBB, MIB))
791 return false;
792 }
793 return true;
794}
795
796void IRTranslator::splitWorkItem(SwitchCG::SwitchWorkList &WorkList,
798 Value *Cond, MachineBasicBlock *SwitchMBB,
799 MachineIRBuilder &MIB) {
800 using namespace SwitchCG;
801 assert(W.FirstCluster->Low->getValue().slt(W.LastCluster->Low->getValue()) &&
802 "Clusters not sorted?");
803 assert(W.LastCluster - W.FirstCluster + 1 >= 2 && "Too small to split!");
804
805 auto [LastLeft, FirstRight, LeftProb, RightProb] =
806 SL->computeSplitWorkItemInfo(W);
807
808 // Use the first element on the right as pivot since we will make less-than
809 // comparisons against it.
810 CaseClusterIt PivotCluster = FirstRight;
811 assert(PivotCluster > W.FirstCluster);
812 assert(PivotCluster <= W.LastCluster);
813
814 CaseClusterIt FirstLeft = W.FirstCluster;
815 CaseClusterIt LastRight = W.LastCluster;
816
817 const ConstantInt *Pivot = PivotCluster->Low;
818
819 // New blocks will be inserted immediately after the current one.
821 ++BBI;
822
823 // We will branch to the LHS if Value < Pivot. If LHS is a single cluster,
824 // we can branch to its destination directly if it's squeezed exactly in
825 // between the known lower bound and Pivot - 1.
826 MachineBasicBlock *LeftMBB;
827 if (FirstLeft == LastLeft && FirstLeft->Kind == CC_Range &&
828 FirstLeft->Low == W.GE &&
829 (FirstLeft->High->getValue() + 1LL) == Pivot->getValue()) {
830 LeftMBB = FirstLeft->MBB;
831 } else {
832 LeftMBB = FuncInfo.MF->CreateMachineBasicBlock(W.MBB->getBasicBlock());
833 FuncInfo.MF->insert(BBI, LeftMBB);
834 WorkList.push_back(
835 {LeftMBB, FirstLeft, LastLeft, W.GE, Pivot, W.DefaultProb / 2});
836 }
837
838 // Similarly, we will branch to the RHS if Value >= Pivot. If RHS is a
839 // single cluster, RHS.Low == Pivot, and we can branch to its destination
840 // directly if RHS.High equals the current upper bound.
841 MachineBasicBlock *RightMBB;
842 if (FirstRight == LastRight && FirstRight->Kind == CC_Range && W.LT &&
843 (FirstRight->High->getValue() + 1ULL) == W.LT->getValue()) {
844 RightMBB = FirstRight->MBB;
845 } else {
846 RightMBB = FuncInfo.MF->CreateMachineBasicBlock(W.MBB->getBasicBlock());
847 FuncInfo.MF->insert(BBI, RightMBB);
848 WorkList.push_back(
849 {RightMBB, FirstRight, LastRight, Pivot, W.LT, W.DefaultProb / 2});
850 }
851
852 // Create the CaseBlock record that will be used to lower the branch.
853 CaseBlock CB(ICmpInst::Predicate::ICMP_SLT, false, Cond, Pivot, nullptr,
854 LeftMBB, RightMBB, W.MBB, MIB.getDebugLoc(), LeftProb,
855 RightProb);
856
857 if (W.MBB == SwitchMBB)
858 emitSwitchCase(CB, SwitchMBB, MIB);
859 else
860 SL->SwitchCases.push_back(CB);
861}
862
863void IRTranslator::emitJumpTable(SwitchCG::JumpTable &JT,
865 // Emit the code for the jump table
866 assert(JT.Reg && "Should lower JT Header first!");
867 MachineIRBuilder MIB(*MBB->getParent());
868 MIB.setMBB(*MBB);
869 MIB.setDebugLoc(CurBuilder->getDebugLoc());
870
871 Type *PtrIRTy = PointerType::getUnqual(MF->getFunction().getContext());
872 const LLT PtrTy = getLLTForType(*PtrIRTy, *DL);
873
874 auto Table = MIB.buildJumpTable(PtrTy, JT.JTI);
875 MIB.buildBrJT(Table.getReg(0), JT.JTI, JT.Reg);
876}
877
878bool IRTranslator::emitJumpTableHeader(SwitchCG::JumpTable &JT,
880 MachineBasicBlock *HeaderBB) {
881 MachineIRBuilder MIB(*HeaderBB->getParent());
882 MIB.setMBB(*HeaderBB);
883 MIB.setDebugLoc(CurBuilder->getDebugLoc());
884
885 const Value &SValue = *JTH.SValue;
886 // Subtract the lowest switch case value from the value being switched on.
887 const LLT SwitchTy = getLLTForType(*SValue.getType(), *DL);
888 Register SwitchOpReg = getOrCreateVReg(SValue);
889 auto FirstCst = MIB.buildConstant(SwitchTy, JTH.First);
890 auto Sub = MIB.buildSub({SwitchTy}, SwitchOpReg, FirstCst);
891
892 // This value may be smaller or larger than the target's pointer type, and
893 // therefore require extension or truncating.
894 auto *PtrIRTy = PointerType::getUnqual(SValue.getContext());
895 const LLT PtrScalarTy = LLT::integer(DL->getTypeSizeInBits(PtrIRTy));
896 Sub = MIB.buildZExtOrTrunc(PtrScalarTy, Sub);
897
898 JT.Reg = Sub.getReg(0);
899
900 if (JTH.FallthroughUnreachable) {
901 if (JT.MBB != HeaderBB->getNextNode())
902 MIB.buildBr(*JT.MBB);
903 return true;
904 }
905
906 // Emit the range check for the jump table, and branch to the default block
907 // for the switch statement if the value being switched on exceeds the
908 // largest case in the switch.
909 auto Cst = getOrCreateVReg(
910 *ConstantInt::get(SValue.getType(), JTH.Last - JTH.First));
911 Cst = MIB.buildZExtOrTrunc(PtrScalarTy, Cst).getReg(0);
912 auto Cmp = MIB.buildICmp(CmpInst::ICMP_UGT, LLT::integer(1), Sub, Cst);
913
914 auto BrCond = MIB.buildBrCond(Cmp.getReg(0), *JT.Default);
915
916 // Avoid emitting unnecessary branches to the next block.
917 if (JT.MBB != HeaderBB->getNextNode())
918 BrCond = MIB.buildBr(*JT.MBB);
919 return true;
920}
921
922void IRTranslator::emitSwitchCase(SwitchCG::CaseBlock &CB,
923 MachineBasicBlock *SwitchBB,
924 MachineIRBuilder &MIB) {
925 Register CondLHS = getOrCreateVReg(*CB.CmpLHS);
927 DebugLoc OldDbgLoc = MIB.getDebugLoc();
928 MIB.setDebugLoc(CB.DbgLoc);
929 MIB.setMBB(*CB.ThisBB);
930
931 if (CB.PredInfo.NoCmp) {
932 // Branch or fall through to TrueBB.
933 addSuccessorWithProb(CB.ThisBB, CB.TrueBB, CB.TrueProb);
934 addMachineCFGPred({SwitchBB->getBasicBlock(), CB.TrueBB->getBasicBlock()},
935 CB.ThisBB);
937 if (CB.TrueBB != CB.ThisBB->getNextNode())
938 MIB.buildBr(*CB.TrueBB);
939 MIB.setDebugLoc(OldDbgLoc);
940 return;
941 }
942
943 const LLT i1Ty = LLT::integer(1);
944 // Build the compare.
945 if (!CB.CmpMHS) {
946 const auto *CI = dyn_cast<ConstantInt>(CB.CmpRHS);
947 // For conditional branch lowering, we might try to do something silly like
948 // emit an G_ICMP to compare an existing G_ICMP i1 result with true. If so,
949 // just re-use the existing condition vreg.
950 if (MRI->getType(CondLHS).getSizeInBits() == 1 && CI && CI->isOne() &&
952 Cond = CondLHS;
953 } else {
954 Register CondRHS = getOrCreateVReg(*CB.CmpRHS);
956 Cond =
957 MIB.buildFCmp(CB.PredInfo.Pred, i1Ty, CondLHS, CondRHS).getReg(0);
958 else
959 Cond =
960 MIB.buildICmp(CB.PredInfo.Pred, i1Ty, CondLHS, CondRHS).getReg(0);
961 }
962 } else {
964 "Can only handle SLE ranges");
965
966 const APInt& Low = cast<ConstantInt>(CB.CmpLHS)->getValue();
967 const APInt& High = cast<ConstantInt>(CB.CmpRHS)->getValue();
968
969 Register CmpOpReg = getOrCreateVReg(*CB.CmpMHS);
970 if (cast<ConstantInt>(CB.CmpLHS)->isMinValue(true)) {
971 Register CondRHS = getOrCreateVReg(*CB.CmpRHS);
972 Cond =
973 MIB.buildICmp(CmpInst::ICMP_SLE, i1Ty, CmpOpReg, CondRHS).getReg(0);
974 } else {
975 const LLT CmpTy = MRI->getType(CmpOpReg);
976 auto Sub = MIB.buildSub({CmpTy}, CmpOpReg, CondLHS);
977 auto Diff = MIB.buildConstant(CmpTy, High - Low);
978 Cond = MIB.buildICmp(CmpInst::ICMP_ULE, i1Ty, Sub, Diff).getReg(0);
979 }
980 }
981
982 // Update successor info
983 addSuccessorWithProb(CB.ThisBB, CB.TrueBB, CB.TrueProb);
984
985 addMachineCFGPred({SwitchBB->getBasicBlock(), CB.TrueBB->getBasicBlock()},
986 CB.ThisBB);
987
988 // TrueBB and FalseBB are always different unless the incoming IR is
989 // degenerate. This only happens when running llc on weird IR.
990 if (CB.TrueBB != CB.FalseBB)
991 addSuccessorWithProb(CB.ThisBB, CB.FalseBB, CB.FalseProb);
993
994 addMachineCFGPred({SwitchBB->getBasicBlock(), CB.FalseBB->getBasicBlock()},
995 CB.ThisBB);
996
997 MIB.buildBrCond(Cond, *CB.TrueBB);
998 MIB.buildBr(*CB.FalseBB);
999 MIB.setDebugLoc(OldDbgLoc);
1000}
1001
1002bool IRTranslator::lowerJumpTableWorkItem(SwitchCG::SwitchWorkListItem W,
1003 MachineBasicBlock *SwitchMBB,
1004 MachineBasicBlock *CurMBB,
1005 MachineBasicBlock *DefaultMBB,
1006 MachineIRBuilder &MIB,
1008 BranchProbability UnhandledProbs,
1010 MachineBasicBlock *Fallthrough,
1011 bool FallthroughUnreachable) {
1012 using namespace SwitchCG;
1013 MachineFunction *CurMF = SwitchMBB->getParent();
1014 // FIXME: Optimize away range check based on pivot comparisons.
1015 JumpTableHeader *JTH = &SL->JTCases[I->JTCasesIndex].first;
1016 SwitchCG::JumpTable *JT = &SL->JTCases[I->JTCasesIndex].second;
1017 BranchProbability DefaultProb = W.DefaultProb;
1018
1019 // The jump block hasn't been inserted yet; insert it here.
1020 MachineBasicBlock *JumpMBB = JT->MBB;
1021 CurMF->insert(BBI, JumpMBB);
1022
1023 // Since the jump table block is separate from the switch block, we need
1024 // to keep track of it as a machine predecessor to the default block,
1025 // otherwise we lose the phi edges.
1026 addMachineCFGPred({SwitchMBB->getBasicBlock(), DefaultMBB->getBasicBlock()},
1027 CurMBB);
1028 addMachineCFGPred({SwitchMBB->getBasicBlock(), DefaultMBB->getBasicBlock()},
1029 JumpMBB);
1030
1031 auto JumpProb = I->Prob;
1032 auto FallthroughProb = UnhandledProbs;
1033
1034 // If the default statement is a target of the jump table, we evenly
1035 // distribute the default probability to successors of CurMBB. Also
1036 // update the probability on the edge from JumpMBB to Fallthrough.
1037 for (MachineBasicBlock::succ_iterator SI = JumpMBB->succ_begin(),
1038 SE = JumpMBB->succ_end();
1039 SI != SE; ++SI) {
1040 if (*SI == DefaultMBB) {
1041 JumpProb += DefaultProb / 2;
1042 FallthroughProb -= DefaultProb / 2;
1043 JumpMBB->setSuccProbability(SI, DefaultProb / 2);
1044 JumpMBB->normalizeSuccProbs();
1045 } else {
1046 // Also record edges from the jump table block to it's successors.
1047 addMachineCFGPred({SwitchMBB->getBasicBlock(), (*SI)->getBasicBlock()},
1048 JumpMBB);
1049 }
1050 }
1051
1052 if (FallthroughUnreachable)
1053 JTH->FallthroughUnreachable = true;
1054
1055 if (!JTH->FallthroughUnreachable)
1056 addSuccessorWithProb(CurMBB, Fallthrough, FallthroughProb);
1057 addSuccessorWithProb(CurMBB, JumpMBB, JumpProb);
1058 CurMBB->normalizeSuccProbs();
1059
1060 // The jump table header will be inserted in our current block, do the
1061 // range check, and fall through to our fallthrough block.
1062 JTH->HeaderBB = CurMBB;
1063 JT->Default = Fallthrough; // FIXME: Move Default to JumpTableHeader.
1064
1065 // If we're in the right place, emit the jump table header right now.
1066 if (CurMBB == SwitchMBB) {
1067 if (!emitJumpTableHeader(*JT, *JTH, CurMBB))
1068 return false;
1069 JTH->Emitted = true;
1070 }
1071 return true;
1072}
1073bool IRTranslator::lowerSwitchRangeWorkItem(SwitchCG::CaseClusterIt I,
1074 Value *Cond,
1075 MachineBasicBlock *Fallthrough,
1076 bool FallthroughUnreachable,
1077 BranchProbability UnhandledProbs,
1078 MachineBasicBlock *CurMBB,
1079 MachineIRBuilder &MIB,
1080 MachineBasicBlock *SwitchMBB) {
1081 using namespace SwitchCG;
1082 const Value *RHS, *LHS, *MHS;
1083 CmpInst::Predicate Pred;
1084 if (I->Low == I->High) {
1085 // Check Cond == I->Low.
1086 Pred = CmpInst::ICMP_EQ;
1087 LHS = Cond;
1088 RHS = I->Low;
1089 MHS = nullptr;
1090 } else {
1091 // Check I->Low <= Cond <= I->High.
1092 Pred = CmpInst::ICMP_SLE;
1093 LHS = I->Low;
1094 MHS = Cond;
1095 RHS = I->High;
1096 }
1097
1098 // If Fallthrough is unreachable, fold away the comparison.
1099 // The false probability is the sum of all unhandled cases.
1100 CaseBlock CB(Pred, FallthroughUnreachable, LHS, RHS, MHS, I->MBB, Fallthrough,
1101 CurMBB, MIB.getDebugLoc(), I->Prob, UnhandledProbs);
1102
1103 emitSwitchCase(CB, SwitchMBB, MIB);
1104 return true;
1105}
1106
1107void IRTranslator::emitBitTestHeader(SwitchCG::BitTestBlock &B,
1108 MachineBasicBlock *SwitchBB) {
1109 MachineIRBuilder &MIB = *CurBuilder;
1110 MIB.setMBB(*SwitchBB);
1111
1112 // Subtract the minimum value.
1113 Register SwitchOpReg = getOrCreateVReg(*B.SValue);
1114
1115 LLT SwitchOpTy = MRI->getType(SwitchOpReg);
1116 Register MinValReg = MIB.buildConstant(SwitchOpTy, B.First).getReg(0);
1117 auto RangeSub = MIB.buildSub(SwitchOpTy, SwitchOpReg, MinValReg);
1118
1119 Type *PtrIRTy = PointerType::getUnqual(MF->getFunction().getContext());
1120 const LLT PtrTy = getLLTForType(*PtrIRTy, *DL);
1121
1122 LLT MaskTy = SwitchOpTy;
1123 if (MaskTy.getSizeInBits() > PtrTy.getSizeInBits() ||
1125 MaskTy = LLT::integer(PtrTy.getSizeInBits());
1126 else {
1127 // Ensure that the type will fit the mask value.
1128 for (const SwitchCG::BitTestCase &Case : B.Cases) {
1129 if (!isUIntN(SwitchOpTy.getSizeInBits(), Case.Mask)) {
1130 // Switch table case range are encoded into series of masks.
1131 // Just use pointer type, it's guaranteed to fit.
1132 MaskTy = LLT::integer(PtrTy.getSizeInBits());
1133 break;
1134 }
1135 }
1136 }
1137 Register SubReg = RangeSub.getReg(0);
1138 if (SwitchOpTy != MaskTy)
1139 SubReg = MIB.buildZExtOrTrunc(MaskTy, SubReg).getReg(0);
1140
1141 B.RegVT = getMVTForLLT(MaskTy);
1142 B.Reg = SubReg;
1143
1144 MachineBasicBlock *MBB = B.Cases[0].ThisBB;
1145
1146 if (!B.FallthroughUnreachable)
1147 addSuccessorWithProb(SwitchBB, B.Default, B.DefaultProb);
1148 addSuccessorWithProb(SwitchBB, MBB, B.Prob);
1149
1150 SwitchBB->normalizeSuccProbs();
1151
1152 if (!B.FallthroughUnreachable) {
1153 // Conditional branch to the default block.
1154 auto RangeCst = MIB.buildConstant(SwitchOpTy, B.Range);
1155 auto RangeCmp = MIB.buildICmp(CmpInst::Predicate::ICMP_UGT, LLT::integer(1),
1156 RangeSub, RangeCst);
1157 MIB.buildBrCond(RangeCmp, *B.Default);
1158 }
1159
1160 // Avoid emitting unnecessary branches to the next block.
1161 if (MBB != SwitchBB->getNextNode())
1162 MIB.buildBr(*MBB);
1163}
1164
1165void IRTranslator::emitBitTestCase(SwitchCG::BitTestBlock &BB,
1166 MachineBasicBlock *NextMBB,
1167 BranchProbability BranchProbToNext,
1169 MachineBasicBlock *SwitchBB) {
1170 MachineIRBuilder &MIB = *CurBuilder;
1171 MIB.setMBB(*SwitchBB);
1172
1173 LLT SwitchTy = getLLTForMVT(BB.RegVT);
1174 Register Cmp;
1175 unsigned PopCount = llvm::popcount(B.Mask);
1176 if (PopCount == 1) {
1177 // Testing for a single bit; just compare the shift count with what it
1178 // would need to be to shift a 1 bit in that position.
1179 auto MaskTrailingZeros =
1180 MIB.buildConstant(SwitchTy, llvm::countr_zero(B.Mask));
1182 MaskTrailingZeros)
1183 .getReg(0);
1184 } else if (PopCount == BB.Range) {
1185 // There is only one zero bit in the range, test for it directly.
1186 auto MaskTrailingOnes =
1187 MIB.buildConstant(SwitchTy, llvm::countr_one(B.Mask));
1188 Cmp =
1189 MIB.buildICmp(CmpInst::ICMP_NE, LLT::integer(1), Reg, MaskTrailingOnes)
1190 .getReg(0);
1191 } else {
1192 // Make desired shift.
1193 auto CstOne = MIB.buildConstant(SwitchTy, 1);
1194 auto SwitchVal = MIB.buildShl(SwitchTy, CstOne, Reg);
1195
1196 // Emit bit tests and jumps.
1197 auto CstMask = MIB.buildConstant(SwitchTy, B.Mask);
1198 auto AndOp = MIB.buildAnd(SwitchTy, SwitchVal, CstMask);
1199 auto CstZero = MIB.buildConstant(SwitchTy, 0);
1200 Cmp = MIB.buildICmp(CmpInst::ICMP_NE, LLT::integer(1), AndOp, CstZero)
1201 .getReg(0);
1202 }
1203
1204 // The branch probability from SwitchBB to B.TargetBB is B.ExtraProb.
1205 addSuccessorWithProb(SwitchBB, B.TargetBB, B.ExtraProb);
1206 // The branch probability from SwitchBB to NextMBB is BranchProbToNext.
1207 addSuccessorWithProb(SwitchBB, NextMBB, BranchProbToNext);
1208 // It is not guaranteed that the sum of B.ExtraProb and BranchProbToNext is
1209 // one as they are relative probabilities (and thus work more like weights),
1210 // and hence we need to normalize them to let the sum of them become one.
1211 SwitchBB->normalizeSuccProbs();
1212
1213 // Record the fact that the IR edge from the header to the bit test target
1214 // will go through our new block. Neeeded for PHIs to have nodes added.
1215 addMachineCFGPred({BB.Parent->getBasicBlock(), B.TargetBB->getBasicBlock()},
1216 SwitchBB);
1217
1218 MIB.buildBrCond(Cmp, *B.TargetBB);
1219
1220 // Avoid emitting unnecessary branches to the next block.
1221 if (NextMBB != SwitchBB->getNextNode())
1222 MIB.buildBr(*NextMBB);
1223}
1224
1225bool IRTranslator::lowerBitTestWorkItem(
1227 MachineBasicBlock *CurMBB, MachineBasicBlock *DefaultMBB,
1229 BranchProbability DefaultProb, BranchProbability UnhandledProbs,
1231 bool FallthroughUnreachable) {
1232 using namespace SwitchCG;
1233 MachineFunction *CurMF = SwitchMBB->getParent();
1234 // FIXME: Optimize away range check based on pivot comparisons.
1235 BitTestBlock *BTB = &SL->BitTestCases[I->BTCasesIndex];
1236 // The bit test blocks haven't been inserted yet; insert them here.
1237 for (BitTestCase &BTC : BTB->Cases)
1238 CurMF->insert(BBI, BTC.ThisBB);
1239
1240 // Fill in fields of the BitTestBlock.
1241 BTB->Parent = CurMBB;
1242 BTB->Default = Fallthrough;
1243
1244 BTB->DefaultProb = UnhandledProbs;
1245 // If the cases in bit test don't form a contiguous range, we evenly
1246 // distribute the probability on the edge to Fallthrough to two
1247 // successors of CurMBB.
1248 if (!BTB->ContiguousRange) {
1249 BTB->Prob += DefaultProb / 2;
1250 BTB->DefaultProb -= DefaultProb / 2;
1251 }
1252
1253 if (FallthroughUnreachable)
1254 BTB->FallthroughUnreachable = true;
1255
1256 // If we're in the right place, emit the bit test header right now.
1257 if (CurMBB == SwitchMBB) {
1258 emitBitTestHeader(*BTB, SwitchMBB);
1259 BTB->Emitted = true;
1260 }
1261 return true;
1262}
1263
1264bool IRTranslator::lowerSwitchWorkItem(SwitchCG::SwitchWorkListItem W,
1265 Value *Cond,
1266 MachineBasicBlock *SwitchMBB,
1267 MachineBasicBlock *DefaultMBB,
1268 MachineIRBuilder &MIB) {
1269 using namespace SwitchCG;
1270 MachineFunction *CurMF = FuncInfo.MF;
1271 MachineBasicBlock *NextMBB = nullptr;
1273 if (++BBI != FuncInfo.MF->end())
1274 NextMBB = &*BBI;
1275
1276 if (EnableOpts) {
1277 // Here, we order cases by probability so the most likely case will be
1278 // checked first. However, two clusters can have the same probability in
1279 // which case their relative ordering is non-deterministic. So we use Low
1280 // as a tie-breaker as clusters are guaranteed to never overlap.
1281 llvm::sort(W.FirstCluster, W.LastCluster + 1,
1282 [](const CaseCluster &a, const CaseCluster &b) {
1283 return a.Prob != b.Prob
1284 ? a.Prob > b.Prob
1285 : a.Low->getValue().slt(b.Low->getValue());
1286 });
1287
1288 // Rearrange the case blocks so that the last one falls through if possible
1289 // without changing the order of probabilities.
1290 for (CaseClusterIt I = W.LastCluster; I > W.FirstCluster;) {
1291 --I;
1292 if (I->Prob > W.LastCluster->Prob)
1293 break;
1294 if (I->Kind == CC_Range && I->MBB == NextMBB) {
1295 std::swap(*I, *W.LastCluster);
1296 break;
1297 }
1298 }
1299 }
1300
1301 // Compute total probability.
1302 BranchProbability DefaultProb = W.DefaultProb;
1303 BranchProbability UnhandledProbs = DefaultProb;
1304 for (CaseClusterIt I = W.FirstCluster; I <= W.LastCluster; ++I)
1305 UnhandledProbs += I->Prob;
1306
1307 MachineBasicBlock *CurMBB = W.MBB;
1308 for (CaseClusterIt I = W.FirstCluster, E = W.LastCluster; I <= E; ++I) {
1309 bool FallthroughUnreachable = false;
1310 MachineBasicBlock *Fallthrough;
1311 if (I == W.LastCluster) {
1312 // For the last cluster, fall through to the default destination.
1313 Fallthrough = DefaultMBB;
1314 FallthroughUnreachable = isa<UnreachableInst>(
1315 DefaultMBB->getBasicBlock()->getFirstNonPHIOrDbg());
1316 } else {
1317 Fallthrough = CurMF->CreateMachineBasicBlock(CurMBB->getBasicBlock());
1318 CurMF->insert(BBI, Fallthrough);
1319 }
1320 UnhandledProbs -= I->Prob;
1321
1322 switch (I->Kind) {
1323 case CC_BitTests: {
1324 if (!lowerBitTestWorkItem(W, SwitchMBB, CurMBB, DefaultMBB, MIB, BBI,
1325 DefaultProb, UnhandledProbs, I, Fallthrough,
1326 FallthroughUnreachable)) {
1327 LLVM_DEBUG(dbgs() << "Failed to lower bit test for switch");
1328 return false;
1329 }
1330 break;
1331 }
1332
1333 case CC_JumpTable: {
1334 if (!lowerJumpTableWorkItem(W, SwitchMBB, CurMBB, DefaultMBB, MIB, BBI,
1335 UnhandledProbs, I, Fallthrough,
1336 FallthroughUnreachable)) {
1337 LLVM_DEBUG(dbgs() << "Failed to lower jump table");
1338 return false;
1339 }
1340 break;
1341 }
1342 case CC_Range: {
1343 if (!lowerSwitchRangeWorkItem(I, Cond, Fallthrough,
1344 FallthroughUnreachable, UnhandledProbs,
1345 CurMBB, MIB, SwitchMBB)) {
1346 LLVM_DEBUG(dbgs() << "Failed to lower switch range");
1347 return false;
1348 }
1349 break;
1350 }
1351 }
1352 CurMBB = Fallthrough;
1353 }
1354
1355 return true;
1356}
1357
1358bool IRTranslator::translateIndirectBr(const User &U,
1359 MachineIRBuilder &MIRBuilder) {
1360 const IndirectBrInst &BrInst = cast<IndirectBrInst>(U);
1361
1362 const Register Tgt = getOrCreateVReg(*BrInst.getAddress());
1363 MIRBuilder.buildBrIndirect(Tgt);
1364
1365 // Link successors.
1366 SmallPtrSet<const BasicBlock *, 32> AddedSuccessors;
1367 MachineBasicBlock &CurBB = MIRBuilder.getMBB();
1368 for (const BasicBlock *Succ : successors(&BrInst)) {
1369 // It's legal for indirectbr instructions to have duplicate blocks in the
1370 // destination list. We don't allow this in MIR. Skip anything that's
1371 // already a successor.
1372 if (!AddedSuccessors.insert(Succ).second)
1373 continue;
1374 CurBB.addSuccessor(&getMBB(*Succ));
1375 }
1376
1377 return true;
1378}
1379
1380static bool isSwiftError(const Value *V) {
1381 if (auto Arg = dyn_cast<Argument>(V))
1382 return Arg->hasSwiftErrorAttr();
1383 if (auto AI = dyn_cast<AllocaInst>(V))
1384 return AI->isSwiftError();
1385 return false;
1386}
1387
1388bool IRTranslator::translateLoad(const User &U, MachineIRBuilder &MIRBuilder) {
1389 const LoadInst &LI = cast<LoadInst>(U);
1390 TypeSize StoreSize = DL->getTypeStoreSize(LI.getType());
1391 if (StoreSize.isZero())
1392 return true;
1393
1394 ArrayRef<Register> Regs = getOrCreateVRegs(LI);
1395 Register Base = getOrCreateVReg(*LI.getPointerOperand());
1396 AAMDNodes AAInfo = LI.getAAMetadata();
1397
1398 const Value *Ptr = LI.getPointerOperand();
1399
1400 if (CLI->supportSwiftError() && isSwiftError(Ptr)) {
1401 assert(Regs.size() == 1 && "swifterror should be single pointer");
1402 Register VReg =
1403 SwiftError.getOrCreateVRegUseAt(&LI, &MIRBuilder.getMBB(), Ptr);
1404 MIRBuilder.buildCopy(Regs[0], VReg);
1405 return true;
1406 }
1407
1409 TLI->getLoadMemOperandFlags(LI, *DL, AC, LibInfo, OptLevel);
1410 if (AA && !(Flags & MachineMemOperand::MOInvariant)) {
1411 if (AA->pointsToConstantMemory(
1412 MemoryLocation(Ptr, LocationSize::precise(StoreSize), AAInfo))) {
1414 }
1415 }
1416
1417 // Fast-path the common single-register load.
1418 if (Regs.size() == 1) {
1419 auto *MMO = MF->getMachineMemOperand(
1420 MachinePointerInfo(LI.getPointerOperand()), Flags,
1421 MRI->getType(Regs[0]), getMemOpAlign(LI), AAInfo,
1422 LI.getMetadata(LLVMContext::MD_range), LI.getSyncScopeID(),
1423 LI.getOrdering());
1424 MIRBuilder.buildLoad(Regs[0], Base, *MMO);
1425 return true;
1426 }
1427
1428 ArrayRef<uint64_t> Offsets = *VMap.getOffsets(LI);
1429 Type *OffsetIRTy = DL->getIndexType(Ptr->getType());
1430 LLT OffsetTy = getLLTForType(*OffsetIRTy, *DL);
1431 for (unsigned i = 0; i < Regs.size(); ++i) {
1432 Register Addr;
1433 MIRBuilder.materializeObjectPtrOffset(Addr, Base, OffsetTy, Offsets[i]);
1434
1435 MachinePointerInfo Ptr(LI.getPointerOperand(), Offsets[i]);
1436 Align BaseAlign = getMemOpAlign(LI);
1437 auto *MMO = MF->getMachineMemOperand(Ptr, Flags, MRI->getType(Regs[i]),
1438 commonAlignment(BaseAlign, Offsets[i]),
1439 AAInfo, nullptr, LI.getSyncScopeID(),
1440 LI.getOrdering());
1441 MIRBuilder.buildLoad(Regs[i], Addr, *MMO);
1442 }
1443
1444 return true;
1445}
1446
1447bool IRTranslator::translateStore(const User &U, MachineIRBuilder &MIRBuilder) {
1448 const StoreInst &SI = cast<StoreInst>(U);
1449 if (DL->getTypeStoreSize(SI.getValueOperand()->getType()).isZero())
1450 return true;
1451
1452 ArrayRef<Register> Vals = getOrCreateVRegs(*SI.getValueOperand());
1453 Register Base = getOrCreateVReg(*SI.getPointerOperand());
1454
1455 if (CLI->supportSwiftError() && isSwiftError(SI.getPointerOperand())) {
1456 assert(Vals.size() == 1 && "swifterror should be single pointer");
1457
1458 Register VReg = SwiftError.getOrCreateVRegDefAt(&SI, &MIRBuilder.getMBB(),
1459 SI.getPointerOperand());
1460 MIRBuilder.buildCopy(VReg, Vals[0]);
1461 return true;
1462 }
1463
1464 MachineMemOperand::Flags Flags = TLI->getStoreMemOperandFlags(SI, *DL);
1465 // Fast-path the common single-register store.
1466 if (Vals.size() == 1) {
1467 auto *MMO = MF->getMachineMemOperand(
1468 MachinePointerInfo(SI.getPointerOperand()), Flags,
1469 MRI->getType(Vals[0]), getMemOpAlign(SI), SI.getAAMetadata(), nullptr,
1470 SI.getSyncScopeID(), SI.getOrdering());
1471 MIRBuilder.buildStore(Vals[0], Base, *MMO);
1472 return true;
1473 }
1474
1475 ArrayRef<uint64_t> Offsets = *VMap.getOffsets(*SI.getValueOperand());
1476 Type *OffsetIRTy = DL->getIndexType(SI.getPointerOperandType());
1477 LLT OffsetTy = getLLTForType(*OffsetIRTy, *DL);
1478 for (unsigned i = 0; i < Vals.size(); ++i) {
1479 Register Addr;
1480 MIRBuilder.materializeObjectPtrOffset(Addr, Base, OffsetTy, Offsets[i]);
1481
1482 MachinePointerInfo Ptr(SI.getPointerOperand(), Offsets[i]);
1483 Align BaseAlign = getMemOpAlign(SI);
1484 auto *MMO = MF->getMachineMemOperand(Ptr, Flags, MRI->getType(Vals[i]),
1485 commonAlignment(BaseAlign, Offsets[i]),
1486 SI.getAAMetadata(), nullptr,
1487 SI.getSyncScopeID(), SI.getOrdering());
1488 MIRBuilder.buildStore(Vals[i], Addr, *MMO);
1489 }
1490 return true;
1491}
1492
1494 const Value *Src = U.getOperand(0);
1495 Type *Int32Ty = Type::getInt32Ty(U.getContext());
1496
1497 // getIndexedOffsetInType is designed for GEPs, so the first index is the
1498 // usual array element rather than looking into the actual aggregate.
1500 Indices.push_back(ConstantInt::get(Int32Ty, 0));
1501
1502 if (const ExtractValueInst *EVI = dyn_cast<ExtractValueInst>(&U)) {
1503 for (auto Idx : EVI->indices())
1504 Indices.push_back(ConstantInt::get(Int32Ty, Idx));
1505 } else if (const InsertValueInst *IVI = dyn_cast<InsertValueInst>(&U)) {
1506 for (auto Idx : IVI->indices())
1507 Indices.push_back(ConstantInt::get(Int32Ty, Idx));
1508 } else {
1509 llvm::append_range(Indices, drop_begin(U.operands()));
1510 }
1511
1512 return static_cast<uint64_t>(
1513 DL.getIndexedOffsetInType(Src->getType(), Indices));
1514}
1515
1516bool IRTranslator::translateExtractValue(const User &U,
1517 MachineIRBuilder &MIRBuilder) {
1518 const Value *Src = U.getOperand(0);
1519 uint64_t Offset = getOffsetFromIndices(U, *DL);
1520 ArrayRef<Register> SrcRegs = getOrCreateVRegs(*Src);
1521 ArrayRef<uint64_t> Offsets = *VMap.getOffsets(*Src);
1522 unsigned Idx = llvm::lower_bound(Offsets, Offset) - Offsets.begin();
1523 auto &DstRegs = allocateVRegs(U);
1524
1525 for (unsigned i = 0; i < DstRegs.size(); ++i)
1526 DstRegs[i] = SrcRegs[Idx++];
1527
1528 return true;
1529}
1530
1531bool IRTranslator::translateInsertValue(const User &U,
1532 MachineIRBuilder &MIRBuilder) {
1533 const Value *Src = U.getOperand(0);
1534 uint64_t Offset = getOffsetFromIndices(U, *DL);
1535 auto &DstRegs = allocateVRegs(U);
1536 ArrayRef<uint64_t> DstOffsets = *VMap.getOffsets(U);
1537 ArrayRef<Register> SrcRegs = getOrCreateVRegs(*Src);
1538 ArrayRef<Register> InsertedRegs = getOrCreateVRegs(*U.getOperand(1));
1539 auto *InsertedIt = InsertedRegs.begin();
1540
1541 for (unsigned i = 0; i < DstRegs.size(); ++i) {
1542 if (DstOffsets[i] >= Offset && InsertedIt != InsertedRegs.end())
1543 DstRegs[i] = *InsertedIt++;
1544 else
1545 DstRegs[i] = SrcRegs[i];
1546 }
1547
1548 return true;
1549}
1550
1551bool IRTranslator::translateSelect(const User &U,
1552 MachineIRBuilder &MIRBuilder) {
1553 Register Tst = getOrCreateVReg(*U.getOperand(0));
1554 ArrayRef<Register> ResRegs = getOrCreateVRegs(U);
1555 ArrayRef<Register> Op0Regs = getOrCreateVRegs(*U.getOperand(1));
1556 ArrayRef<Register> Op1Regs = getOrCreateVRegs(*U.getOperand(2));
1557
1558 uint32_t Flags = 0;
1559 if (const SelectInst *SI = dyn_cast<SelectInst>(&U))
1561
1562 for (unsigned i = 0; i < ResRegs.size(); ++i) {
1563 MIRBuilder.buildSelect(ResRegs[i], Tst, Op0Regs[i], Op1Regs[i], Flags);
1564 }
1565
1566 return true;
1567}
1568
1569bool IRTranslator::translateCopy(const User &U, const Value &V,
1570 MachineIRBuilder &MIRBuilder) {
1571 Register Src = getOrCreateVReg(V);
1572 auto &Regs = *VMap.getVRegs(U);
1573 if (Regs.empty()) {
1574 Regs.push_back(Src);
1575 VMap.getOffsets(U)->push_back(0);
1576 } else {
1577 // If we already assigned a vreg for this instruction, we can't change that.
1578 // Emit a copy to satisfy the users we already emitted.
1579 MIRBuilder.buildCopy(Regs[0], Src);
1580 }
1581 return true;
1582}
1583
1584bool IRTranslator::translateBitCast(const User &U,
1585 MachineIRBuilder &MIRBuilder) {
1586 // If we're bitcasting to the source type, we can reuse the source vreg.
1587 if (getLLTForType(*U.getOperand(0)->getType(), *DL) ==
1588 getLLTForType(*U.getType(), *DL)) {
1589 // If the source is a ConstantInt then it was probably created by
1590 // ConstantHoisting and we should leave it alone.
1591 if (isa<ConstantInt>(U.getOperand(0)))
1592 return translateCast(TargetOpcode::G_CONSTANT_FOLD_BARRIER, U,
1593 MIRBuilder);
1594 return translateCopy(U, *U.getOperand(0), MIRBuilder);
1595 }
1596
1597 return translateCast(TargetOpcode::G_BITCAST, U, MIRBuilder);
1598}
1599
1600bool IRTranslator::translateCast(unsigned Opcode, const User &U,
1601 MachineIRBuilder &MIRBuilder) {
1602 if (!mayTranslateUserTypes(U))
1603 return false;
1604
1605 uint32_t Flags = 0;
1606 if (const Instruction *I = dyn_cast<Instruction>(&U))
1608
1609 Register Op = getOrCreateVReg(*U.getOperand(0));
1610 Register Res = getOrCreateVReg(U);
1611 MIRBuilder.buildInstr(Opcode, {Res}, {Op}, Flags);
1612 return true;
1613}
1614
1615bool IRTranslator::translateGetElementPtr(const User &U,
1616 MachineIRBuilder &MIRBuilder) {
1617 Value &Op0 = *U.getOperand(0);
1618 Register BaseReg = getOrCreateVReg(Op0);
1619 Type *PtrIRTy = Op0.getType();
1620 LLT PtrTy = getLLTForType(*PtrIRTy, *DL);
1621 Type *OffsetIRTy = DL->getIndexType(PtrIRTy);
1622 LLT OffsetTy = getLLTForType(*OffsetIRTy, *DL);
1623
1624 uint32_t PtrAddFlags = 0;
1625 // Each PtrAdd generated to implement the GEP inherits its nuw, nusw, inbounds
1626 // flags.
1627 if (const Instruction *I = dyn_cast<Instruction>(&U))
1629
1630 auto PtrAddFlagsWithConst = [&](int64_t Offset) {
1631 // For nusw/inbounds GEP with an offset that is nonnegative when interpreted
1632 // as signed, assume there is no unsigned overflow.
1633 if (Offset >= 0 && (PtrAddFlags & MachineInstr::MIFlag::NoUSWrap))
1634 return PtrAddFlags | MachineInstr::MIFlag::NoUWrap;
1635 return PtrAddFlags;
1636 };
1637
1638 // Normalize Vector GEP - all scalar operands should be converted to the
1639 // splat vector.
1640 unsigned VectorWidth = 0;
1641
1642 // True if we should use a splat vector; using VectorWidth alone is not
1643 // sufficient.
1644 bool WantSplatVector = false;
1645 if (auto *VT = dyn_cast<VectorType>(U.getType())) {
1646 VectorWidth = cast<FixedVectorType>(VT)->getNumElements();
1647 // We don't produce 1 x N vectors; those are treated as scalars.
1648 WantSplatVector = VectorWidth > 1;
1649 }
1650
1651 // We might need to splat the base pointer into a vector if the offsets
1652 // are vectors.
1653 if (WantSplatVector && !PtrTy.isVector()) {
1654 BaseReg = MIRBuilder
1655 .buildSplatBuildVector(LLT::fixed_vector(VectorWidth, PtrTy),
1656 BaseReg)
1657 .getReg(0);
1658 PtrIRTy = FixedVectorType::get(PtrIRTy, VectorWidth);
1659 PtrTy = getLLTForType(*PtrIRTy, *DL);
1660 OffsetIRTy = DL->getIndexType(PtrIRTy);
1661 OffsetTy = getLLTForType(*OffsetIRTy, *DL);
1662 }
1663
1664 int64_t Offset = 0;
1665 for (gep_type_iterator GTI = gep_type_begin(&U), E = gep_type_end(&U);
1666 GTI != E; ++GTI) {
1667 const Value *Idx = GTI.getOperand();
1668 if (StructType *StTy = GTI.getStructTypeOrNull()) {
1669 unsigned Field = cast<Constant>(Idx)->getUniqueInteger().getZExtValue();
1670 Offset += DL->getStructLayout(StTy)->getElementOffset(Field);
1671 continue;
1672 } else {
1673 uint64_t ElementSize = GTI.getSequentialElementStride(*DL);
1674
1675 // If this is a scalar constant or a splat vector of constants,
1676 // handle it quickly.
1677 if (const auto *CI = dyn_cast<ConstantInt>(Idx)) {
1678 if (std::optional<int64_t> Val = CI->getValue().trySExtValue()) {
1679 Offset += ElementSize * *Val;
1680 continue;
1681 }
1682 }
1683
1684 if (Offset != 0) {
1685 auto OffsetMIB = MIRBuilder.buildConstant({OffsetTy}, Offset);
1686 BaseReg = MIRBuilder
1687 .buildPtrAdd(PtrTy, BaseReg, OffsetMIB.getReg(0),
1688 PtrAddFlagsWithConst(Offset))
1689 .getReg(0);
1690 Offset = 0;
1691 }
1692
1693 Register IdxReg = getOrCreateVReg(*Idx);
1694 LLT IdxTy = MRI->getType(IdxReg);
1695 if (IdxTy != OffsetTy) {
1696 if (!IdxTy.isVector() && WantSplatVector) {
1697 IdxReg = MIRBuilder
1699 IdxReg)
1700 .getReg(0);
1701 }
1702
1703 IdxReg = MIRBuilder.buildSExtOrTrunc(OffsetTy, IdxReg).getReg(0);
1704 }
1705
1706 // N = N + Idx * ElementSize;
1707 // Avoid doing it for ElementSize of 1.
1708 Register GepOffsetReg;
1709 if (ElementSize != 1) {
1710 auto ElementSizeMIB = MIRBuilder.buildConstant(
1711 getLLTForType(*OffsetIRTy, *DL), ElementSize);
1712
1713 // The multiplication is NUW if the GEP is NUW and NSW if the GEP is
1714 // NUSW.
1715 uint32_t ScaleFlags = PtrAddFlags & MachineInstr::MIFlag::NoUWrap;
1716 if (PtrAddFlags & MachineInstr::MIFlag::NoUSWrap)
1717 ScaleFlags |= MachineInstr::MIFlag::NoSWrap;
1718
1719 GepOffsetReg =
1720 MIRBuilder.buildMul(OffsetTy, IdxReg, ElementSizeMIB, ScaleFlags)
1721 .getReg(0);
1722 } else {
1723 GepOffsetReg = IdxReg;
1724 }
1725
1726 BaseReg =
1727 MIRBuilder.buildPtrAdd(PtrTy, BaseReg, GepOffsetReg, PtrAddFlags)
1728 .getReg(0);
1729 }
1730 }
1731
1732 if (Offset != 0) {
1733 auto OffsetMIB =
1734 MIRBuilder.buildConstant(OffsetTy, Offset);
1735
1736 MIRBuilder.buildPtrAdd(getOrCreateVReg(U), BaseReg, OffsetMIB.getReg(0),
1737 PtrAddFlagsWithConst(Offset));
1738 return true;
1739 }
1740
1741 MIRBuilder.buildCopy(getOrCreateVReg(U), BaseReg);
1742 return true;
1743}
1744
1745bool IRTranslator::translateMemFunc(const CallInst &CI,
1746 MachineIRBuilder &MIRBuilder,
1747 unsigned Opcode) {
1748 const Value *SrcPtr = CI.getArgOperand(1);
1749 // If the source is undef, then just emit a nop.
1750 if (isa<UndefValue>(SrcPtr))
1751 return true;
1752
1754
1755 unsigned MinPtrSize = UINT_MAX;
1756 for (auto AI = CI.arg_begin(), AE = CI.arg_end(); std::next(AI) != AE; ++AI) {
1757 Register SrcReg = getOrCreateVReg(**AI);
1758 LLT SrcTy = MRI->getType(SrcReg);
1759 if (SrcTy.isPointer())
1760 MinPtrSize = std::min<unsigned>(SrcTy.getSizeInBits(), MinPtrSize);
1761 SrcRegs.push_back(SrcReg);
1762 }
1763
1764 LLT SizeTy = LLT::integer(MinPtrSize);
1765
1766 // The size operand should be the minimum of the pointer sizes.
1767 Register &SizeOpReg = SrcRegs[SrcRegs.size() - 1];
1768 if (MRI->getType(SizeOpReg) != SizeTy)
1769 SizeOpReg = MIRBuilder.buildZExtOrTrunc(SizeTy, SizeOpReg).getReg(0);
1770
1771 auto ICall = MIRBuilder.buildInstr(Opcode);
1772 for (Register SrcReg : SrcRegs)
1773 ICall.addUse(SrcReg);
1774
1775 Align DstAlign;
1776 Align SrcAlign;
1777 unsigned IsVol =
1778 cast<ConstantInt>(CI.getArgOperand(CI.arg_size() - 1))->getZExtValue();
1779
1780 ConstantInt *CopySize = nullptr;
1781
1782 if (auto *MCI = dyn_cast<MemCpyInst>(&CI)) {
1783 DstAlign = MCI->getDestAlign().valueOrOne();
1784 SrcAlign = MCI->getSourceAlign().valueOrOne();
1785 CopySize = dyn_cast<ConstantInt>(MCI->getArgOperand(2));
1786 } else if (auto *MMI = dyn_cast<MemMoveInst>(&CI)) {
1787 DstAlign = MMI->getDestAlign().valueOrOne();
1788 SrcAlign = MMI->getSourceAlign().valueOrOne();
1789 CopySize = dyn_cast<ConstantInt>(MMI->getArgOperand(2));
1790 } else {
1791 auto *MSI = cast<MemSetInst>(&CI);
1792 DstAlign = MSI->getDestAlign().valueOrOne();
1793 }
1794
1795 if (Opcode != TargetOpcode::G_MEMCPY_INLINE) {
1796 // We need to propagate the tail call flag from the IR inst as an argument.
1797 // Otherwise, we have to pessimize and assume later that we cannot tail call
1798 // any memory intrinsics.
1799 ICall.addImm(CI.isTailCall() ? 1 : 0);
1800 }
1801
1802 // Create mem operands to store the alignment and volatile info.
1805 if (IsVol) {
1806 LoadFlags |= MachineMemOperand::MOVolatile;
1807 StoreFlags |= MachineMemOperand::MOVolatile;
1808 }
1809
1810 AAMDNodes AAInfo = CI.getAAMetadata();
1811 if (AA && CopySize &&
1812 AA->pointsToConstantMemory(MemoryLocation(
1813 SrcPtr, LocationSize::precise(CopySize->getZExtValue()), AAInfo))) {
1814 LoadFlags |= MachineMemOperand::MOInvariant;
1815
1816 // FIXME: pointsToConstantMemory probably does not imply dereferenceable,
1817 // but the previous usage implied it did. Probably should check
1818 // isDereferenceableAndAlignedPointer.
1820 }
1821
1822 ICall.addMemOperand(
1823 MF->getMachineMemOperand(MachinePointerInfo(CI.getArgOperand(0)),
1824 StoreFlags, 1, DstAlign, AAInfo));
1825 if (Opcode != TargetOpcode::G_MEMSET)
1826 ICall.addMemOperand(MF->getMachineMemOperand(
1827 MachinePointerInfo(SrcPtr), LoadFlags, 1, SrcAlign, AAInfo));
1828
1829 return true;
1830}
1831
1832bool IRTranslator::translateTrap(const CallInst &CI,
1833 MachineIRBuilder &MIRBuilder,
1834 unsigned Opcode) {
1835 StringRef TrapFuncName =
1836 CI.getAttributes().getFnAttr("trap-func-name").getValueAsString();
1837 if (TrapFuncName.empty()) {
1838 if (Opcode == TargetOpcode::G_UBSANTRAP) {
1839 uint64_t Code = cast<ConstantInt>(CI.getOperand(0))->getZExtValue();
1840 MIRBuilder.buildInstr(Opcode, {}, ArrayRef<llvm::SrcOp>{Code});
1841 } else {
1842 MIRBuilder.buildInstr(Opcode);
1843 }
1844 return true;
1845 }
1846
1847 CallLowering::CallLoweringInfo Info;
1848 if (Opcode == TargetOpcode::G_UBSANTRAP)
1849 Info.OrigArgs.push_back({getOrCreateVRegs(*CI.getArgOperand(0)),
1850 CI.getArgOperand(0)->getType(), 0});
1851
1852 Info.Callee = MachineOperand::CreateES(TrapFuncName.data());
1853 Info.CB = &CI;
1854 Info.OrigRet = {Register(), Type::getVoidTy(CI.getContext()), 0};
1855 return CLI->lowerCall(MIRBuilder, Info);
1856}
1857
1858bool IRTranslator::translateVectorInterleave2Intrinsic(
1859 const CallInst &CI, MachineIRBuilder &MIRBuilder) {
1860 assert(CI.getIntrinsicID() == Intrinsic::vector_interleave2 &&
1861 "This function can only be called on the interleave2 intrinsic!");
1862 // Canonicalize interleave2 to G_SHUFFLE_VECTOR (similar to SelectionDAG).
1863 Register Op0 = getOrCreateVReg(*CI.getOperand(0));
1864 Register Op1 = getOrCreateVReg(*CI.getOperand(1));
1865 Register Res = getOrCreateVReg(CI);
1866
1867 LLT OpTy = MRI->getType(Op0);
1868 MIRBuilder.buildShuffleVector(Res, Op0, Op1,
1870
1871 return true;
1872}
1873
1874bool IRTranslator::translateVectorDeinterleave2Intrinsic(
1875 const CallInst &CI, MachineIRBuilder &MIRBuilder) {
1876 assert(CI.getIntrinsicID() == Intrinsic::vector_deinterleave2 &&
1877 "This function can only be called on the deinterleave2 intrinsic!");
1878 // Canonicalize deinterleave2 to shuffles that extract sub-vectors (similar to
1879 // SelectionDAG).
1880 Register Op = getOrCreateVReg(*CI.getOperand(0));
1881 auto Undef = MIRBuilder.buildUndef(MRI->getType(Op));
1882 ArrayRef<Register> Res = getOrCreateVRegs(CI);
1883
1884 LLT ResTy = MRI->getType(Res[0]);
1885 MIRBuilder.buildShuffleVector(Res[0], Op, Undef,
1886 createStrideMask(0, 2, ResTy.getNumElements()));
1887 MIRBuilder.buildShuffleVector(Res[1], Op, Undef,
1888 createStrideMask(1, 2, ResTy.getNumElements()));
1889
1890 return true;
1891}
1892
1893void IRTranslator::getStackGuard(Register DstReg,
1894 MachineIRBuilder &MIRBuilder) {
1895 Value *Global =
1896 TLI->getSDagStackGuard(*MF->getFunction().getParent(), *Libcalls);
1897 if (!Global) {
1898 LLVMContext &Ctx = MIRBuilder.getContext();
1899 Ctx.diagnose(DiagnosticInfoGeneric("unable to lower stackguard"));
1900 MIRBuilder.buildUndef(DstReg);
1901 return;
1902 }
1903
1904 const TargetRegisterInfo *TRI = MF->getSubtarget().getRegisterInfo();
1905 MRI->setRegClass(DstReg, TRI->getPointerRegClass());
1906 auto MIB =
1907 MIRBuilder.buildInstr(TargetOpcode::LOAD_STACK_GUARD, {DstReg}, {});
1908
1909 unsigned AddrSpace = Global->getType()->getPointerAddressSpace();
1910 LLT PtrTy = LLT::pointer(AddrSpace, DL->getPointerSizeInBits(AddrSpace));
1911
1912 MachinePointerInfo MPInfo(Global);
1915 MachineMemOperand *MemRef = MF->getMachineMemOperand(
1916 MPInfo, Flags, PtrTy, DL->getPointerABIAlignment(AddrSpace));
1917 MIB.setMemRefs({MemRef});
1918}
1919
1920bool IRTranslator::translateOverflowIntrinsic(const CallInst &CI, unsigned Op,
1921 MachineIRBuilder &MIRBuilder) {
1922 ArrayRef<Register> ResRegs = getOrCreateVRegs(CI);
1923 MIRBuilder.buildInstr(
1924 Op, {ResRegs[0], ResRegs[1]},
1925 {getOrCreateVReg(*CI.getOperand(0)), getOrCreateVReg(*CI.getOperand(1))});
1926
1927 return true;
1928}
1929
1930bool IRTranslator::translateFixedPointIntrinsic(unsigned Op, const CallInst &CI,
1931 MachineIRBuilder &MIRBuilder) {
1932 Register Dst = getOrCreateVReg(CI);
1933 Register Src0 = getOrCreateVReg(*CI.getOperand(0));
1934 Register Src1 = getOrCreateVReg(*CI.getOperand(1));
1935 uint64_t Scale = cast<ConstantInt>(CI.getOperand(2))->getZExtValue();
1936 MIRBuilder.buildInstr(Op, {Dst}, { Src0, Src1, Scale });
1937 return true;
1938}
1939
1940unsigned IRTranslator::getSimpleIntrinsicOpcode(Intrinsic::ID ID) {
1941 switch (ID) {
1942 default:
1943 break;
1944 case Intrinsic::acos:
1945 return TargetOpcode::G_FACOS;
1946 case Intrinsic::asin:
1947 return TargetOpcode::G_FASIN;
1948 case Intrinsic::atan:
1949 return TargetOpcode::G_FATAN;
1950 case Intrinsic::atan2:
1951 return TargetOpcode::G_FATAN2;
1952 case Intrinsic::bswap:
1953 return TargetOpcode::G_BSWAP;
1954 case Intrinsic::bitreverse:
1955 return TargetOpcode::G_BITREVERSE;
1956 case Intrinsic::fshl:
1957 return TargetOpcode::G_FSHL;
1958 case Intrinsic::fshr:
1959 return TargetOpcode::G_FSHR;
1960 case Intrinsic::ceil:
1961 return TargetOpcode::G_FCEIL;
1962 case Intrinsic::cos:
1963 return TargetOpcode::G_FCOS;
1964 case Intrinsic::cosh:
1965 return TargetOpcode::G_FCOSH;
1966 case Intrinsic::ctpop:
1967 return TargetOpcode::G_CTPOP;
1968 case Intrinsic::exp:
1969 return TargetOpcode::G_FEXP;
1970 case Intrinsic::exp2:
1971 return TargetOpcode::G_FEXP2;
1972 case Intrinsic::exp10:
1973 return TargetOpcode::G_FEXP10;
1974 case Intrinsic::fabs:
1975 return TargetOpcode::G_FABS;
1976 case Intrinsic::copysign:
1977 return TargetOpcode::G_FCOPYSIGN;
1978 case Intrinsic::minnum:
1979 return TargetOpcode::G_FMINNUM;
1980 case Intrinsic::maxnum:
1981 return TargetOpcode::G_FMAXNUM;
1982 case Intrinsic::minimum:
1983 return TargetOpcode::G_FMINIMUM;
1984 case Intrinsic::maximum:
1985 return TargetOpcode::G_FMAXIMUM;
1986 case Intrinsic::minimumnum:
1987 return TargetOpcode::G_FMINIMUMNUM;
1988 case Intrinsic::maximumnum:
1989 return TargetOpcode::G_FMAXIMUMNUM;
1990 case Intrinsic::canonicalize:
1991 return TargetOpcode::G_FCANONICALIZE;
1992 case Intrinsic::floor:
1993 return TargetOpcode::G_FFLOOR;
1994 case Intrinsic::fma:
1995 return TargetOpcode::G_FMA;
1996 case Intrinsic::log:
1997 return TargetOpcode::G_FLOG;
1998 case Intrinsic::log2:
1999 return TargetOpcode::G_FLOG2;
2000 case Intrinsic::log10:
2001 return TargetOpcode::G_FLOG10;
2002 case Intrinsic::ldexp:
2003 return TargetOpcode::G_FLDEXP;
2004 case Intrinsic::nearbyint:
2005 return TargetOpcode::G_FNEARBYINT;
2006 case Intrinsic::pow:
2007 return TargetOpcode::G_FPOW;
2008 case Intrinsic::powi:
2009 return TargetOpcode::G_FPOWI;
2010 case Intrinsic::rint:
2011 return TargetOpcode::G_FRINT;
2012 case Intrinsic::round:
2013 return TargetOpcode::G_INTRINSIC_ROUND;
2014 case Intrinsic::roundeven:
2015 return TargetOpcode::G_INTRINSIC_ROUNDEVEN;
2016 case Intrinsic::sin:
2017 return TargetOpcode::G_FSIN;
2018 case Intrinsic::sinh:
2019 return TargetOpcode::G_FSINH;
2020 case Intrinsic::sqrt:
2021 return TargetOpcode::G_FSQRT;
2022 case Intrinsic::tan:
2023 return TargetOpcode::G_FTAN;
2024 case Intrinsic::tanh:
2025 return TargetOpcode::G_FTANH;
2026 case Intrinsic::trunc:
2027 return TargetOpcode::G_INTRINSIC_TRUNC;
2028 case Intrinsic::readcyclecounter:
2029 return TargetOpcode::G_READCYCLECOUNTER;
2030 case Intrinsic::readsteadycounter:
2031 return TargetOpcode::G_READSTEADYCOUNTER;
2032 case Intrinsic::ptrmask:
2033 return TargetOpcode::G_PTRMASK;
2034 case Intrinsic::lrint:
2035 return TargetOpcode::G_INTRINSIC_LRINT;
2036 case Intrinsic::llrint:
2037 return TargetOpcode::G_INTRINSIC_LLRINT;
2038 // FADD/FMUL require checking the FMF, so are handled elsewhere.
2039 case Intrinsic::vector_reduce_fmin:
2040 return TargetOpcode::G_VECREDUCE_FMIN;
2041 case Intrinsic::vector_reduce_fmax:
2042 return TargetOpcode::G_VECREDUCE_FMAX;
2043 case Intrinsic::vector_reduce_fminimum:
2044 return TargetOpcode::G_VECREDUCE_FMINIMUM;
2045 case Intrinsic::vector_reduce_fmaximum:
2046 return TargetOpcode::G_VECREDUCE_FMAXIMUM;
2047 case Intrinsic::vector_reduce_add:
2048 return TargetOpcode::G_VECREDUCE_ADD;
2049 case Intrinsic::vector_reduce_mul:
2050 return TargetOpcode::G_VECREDUCE_MUL;
2051 case Intrinsic::vector_reduce_and:
2052 return TargetOpcode::G_VECREDUCE_AND;
2053 case Intrinsic::vector_reduce_or:
2054 return TargetOpcode::G_VECREDUCE_OR;
2055 case Intrinsic::vector_reduce_xor:
2056 return TargetOpcode::G_VECREDUCE_XOR;
2057 case Intrinsic::vector_reduce_smax:
2058 return TargetOpcode::G_VECREDUCE_SMAX;
2059 case Intrinsic::vector_reduce_smin:
2060 return TargetOpcode::G_VECREDUCE_SMIN;
2061 case Intrinsic::vector_reduce_umax:
2062 return TargetOpcode::G_VECREDUCE_UMAX;
2063 case Intrinsic::vector_reduce_umin:
2064 return TargetOpcode::G_VECREDUCE_UMIN;
2065 case Intrinsic::experimental_vector_compress:
2066 return TargetOpcode::G_VECTOR_COMPRESS;
2067 case Intrinsic::lround:
2068 return TargetOpcode::G_LROUND;
2069 case Intrinsic::llround:
2070 return TargetOpcode::G_LLROUND;
2071 case Intrinsic::get_fpenv:
2072 return TargetOpcode::G_GET_FPENV;
2073 case Intrinsic::get_fpmode:
2074 return TargetOpcode::G_GET_FPMODE;
2075 }
2077}
2078
2079bool IRTranslator::translateSimpleIntrinsic(const CallInst &CI,
2081 MachineIRBuilder &MIRBuilder) {
2082
2083 unsigned Op = getSimpleIntrinsicOpcode(ID);
2084
2085 // Is this a simple intrinsic?
2087 return false;
2088
2089 // Yes. Let's translate it.
2091 for (const auto &Arg : CI.args())
2092 VRegs.push_back(getOrCreateVReg(*Arg));
2093
2094 MIRBuilder.buildInstr(Op, {getOrCreateVReg(CI)}, VRegs,
2096 return true;
2097}
2098
2099// TODO: Include ConstainedOps.def when all strict instructions are defined.
2101 switch (ID) {
2102 case Intrinsic::experimental_constrained_fadd:
2103 return TargetOpcode::G_STRICT_FADD;
2104 case Intrinsic::experimental_constrained_fsub:
2105 return TargetOpcode::G_STRICT_FSUB;
2106 case Intrinsic::experimental_constrained_fmul:
2107 return TargetOpcode::G_STRICT_FMUL;
2108 case Intrinsic::experimental_constrained_fdiv:
2109 return TargetOpcode::G_STRICT_FDIV;
2110 case Intrinsic::experimental_constrained_frem:
2111 return TargetOpcode::G_STRICT_FREM;
2112 case Intrinsic::experimental_constrained_fma:
2113 return TargetOpcode::G_STRICT_FMA;
2114 case Intrinsic::experimental_constrained_sqrt:
2115 return TargetOpcode::G_STRICT_FSQRT;
2116 case Intrinsic::experimental_constrained_ldexp:
2117 return TargetOpcode::G_STRICT_FLDEXP;
2118 case Intrinsic::experimental_constrained_fcmp:
2119 return TargetOpcode::G_STRICT_FCMP;
2120 case Intrinsic::experimental_constrained_fcmps:
2121 return TargetOpcode::G_STRICT_FCMPS;
2122 default:
2123 return 0;
2124 }
2125}
2126
2127bool IRTranslator::translateConstrainedFPIntrinsic(
2128 const ConstrainedFPIntrinsic &FPI, MachineIRBuilder &MIRBuilder) {
2130
2131 unsigned Opcode = getConstrainedOpcode(FPI.getIntrinsicID());
2132 if (!Opcode)
2133 return false;
2134
2138
2139 if (Opcode == TargetOpcode::G_STRICT_FCMP ||
2140 Opcode == TargetOpcode::G_STRICT_FCMPS) {
2141 auto *FPCmp = cast<ConstrainedFPCmpIntrinsic>(&FPI);
2142 Register Operand0 = getOrCreateVReg(*FPCmp->getArgOperand(0));
2143 Register Operand1 = getOrCreateVReg(*FPCmp->getArgOperand(1));
2144 Register Result = getOrCreateVReg(FPI);
2145 MIRBuilder.buildInstr(Opcode, {Result}, {}, Flags)
2146 .addPredicate(FPCmp->getPredicate())
2147 .addUse(Operand0)
2148 .addUse(Operand1);
2149 return true;
2150 }
2151
2153 for (unsigned I = 0, E = FPI.getNonMetadataArgCount(); I != E; ++I)
2154 VRegs.push_back(getOrCreateVReg(*FPI.getArgOperand(I)));
2155
2156 MIRBuilder.buildInstr(Opcode, {getOrCreateVReg(FPI)}, VRegs, Flags);
2157 return true;
2158}
2159
2160std::optional<MCRegister> IRTranslator::getArgPhysReg(Argument &Arg) {
2161 auto VRegs = getOrCreateVRegs(Arg);
2162 if (VRegs.size() != 1)
2163 return std::nullopt;
2164
2165 // Arguments are lowered as a copy of a livein physical register.
2166 auto *VRegDef = MF->getRegInfo().getVRegDef(VRegs[0]);
2167 if (!VRegDef || !VRegDef->isCopy())
2168 return std::nullopt;
2169 return VRegDef->getOperand(1).getReg().asMCReg();
2170}
2171
2172bool IRTranslator::translateIfEntryValueArgument(bool isDeclare, Value *Val,
2173 const DILocalVariable *Var,
2174 const DIExpression *Expr,
2175 const DebugLoc &DL,
2176 MachineIRBuilder &MIRBuilder) {
2177 auto *Arg = dyn_cast<Argument>(Val);
2178 if (!Arg)
2179 return false;
2180
2181 if (!Expr->isEntryValue())
2182 return false;
2183
2184 std::optional<MCRegister> PhysReg = getArgPhysReg(*Arg);
2185 if (!PhysReg) {
2186 LLVM_DEBUG(dbgs() << "Dropping dbg." << (isDeclare ? "declare" : "value")
2187 << ": expression is entry_value but "
2188 << "couldn't find a physical register\n");
2189 LLVM_DEBUG(dbgs() << *Var << "\n");
2190 return true;
2191 }
2192
2193 if (isDeclare) {
2194 // Append an op deref to account for the fact that this is a dbg_declare.
2195 Expr = DIExpression::append(Expr, dwarf::DW_OP_deref);
2196 MF->setVariableDbgInfo(Var, Expr, *PhysReg, DL);
2197 } else {
2198 MIRBuilder.buildDirectDbgValue(*PhysReg, Var, Expr);
2199 }
2200
2201 return true;
2202}
2203
2205 switch (ID) {
2206 default:
2207 llvm_unreachable("Unexpected intrinsic");
2208 case Intrinsic::experimental_convergence_anchor:
2209 return TargetOpcode::CONVERGENCECTRL_ANCHOR;
2210 case Intrinsic::experimental_convergence_entry:
2211 return TargetOpcode::CONVERGENCECTRL_ENTRY;
2212 case Intrinsic::experimental_convergence_loop:
2213 return TargetOpcode::CONVERGENCECTRL_LOOP;
2214 }
2215}
2216
2217bool IRTranslator::translateConvergenceControlIntrinsic(
2218 const CallInst &CI, Intrinsic::ID ID, MachineIRBuilder &MIRBuilder) {
2219 MachineInstrBuilder MIB = MIRBuilder.buildInstr(getConvOpcode(ID));
2220 Register OutputReg = getOrCreateConvergenceTokenVReg(CI);
2221 MIB.addDef(OutputReg);
2222
2223 if (ID == Intrinsic::experimental_convergence_loop) {
2225 assert(Bundle && "Expected a convergence control token.");
2226 Register InputReg =
2227 getOrCreateConvergenceTokenVReg(*Bundle->Inputs[0].get());
2228 MIB.addUse(InputReg);
2229 }
2230
2231 return true;
2232}
2233
2234bool IRTranslator::translateKnownIntrinsic(const CallInst &CI, Intrinsic::ID ID,
2235 MachineIRBuilder &MIRBuilder) {
2236 if (auto *MI = dyn_cast<AnyMemIntrinsic>(&CI)) {
2237 if (ORE->enabled()) {
2238 if (MemoryOpRemark::canHandle(MI, *LibInfo)) {
2239 MemoryOpRemark R(*ORE, "gisel-irtranslator-memsize", *DL, *LibInfo);
2240 R.visit(MI);
2241 }
2242 }
2243 }
2244
2245 // If this is a simple intrinsic (that is, we just need to add a def of
2246 // a vreg, and uses for each arg operand, then translate it.
2247 if (translateSimpleIntrinsic(CI, ID, MIRBuilder))
2248 return true;
2249
2250 switch (ID) {
2251 default:
2252 break;
2253 case Intrinsic::lifetime_start:
2254 case Intrinsic::lifetime_end: {
2255 // No stack colouring in O0, discard region information.
2256 if (MF->getTarget().getOptLevel() == CodeGenOptLevel::None ||
2257 MF->getFunction().hasOptNone())
2258 return true;
2259
2260 unsigned Op = ID == Intrinsic::lifetime_start ? TargetOpcode::LIFETIME_START
2261 : TargetOpcode::LIFETIME_END;
2262
2263 const AllocaInst *AI = dyn_cast<AllocaInst>(CI.getArgOperand(0));
2264 if (!AI || !AI->isStaticAlloca())
2265 return true;
2266
2267 MIRBuilder.buildInstr(Op).addFrameIndex(getOrCreateFrameIndex(*AI));
2268 return true;
2269 }
2270 case Intrinsic::fake_use: {
2272 for (const auto &Arg : CI.args())
2273 llvm::append_range(VRegs, getOrCreateVRegs(*Arg));
2274 MIRBuilder.buildInstr(TargetOpcode::FAKE_USE, {}, VRegs);
2275 MF->setHasFakeUses(true);
2276 return true;
2277 }
2278 case Intrinsic::dbg_declare: {
2279 const DbgDeclareInst &DI = cast<DbgDeclareInst>(CI);
2280 assert(DI.getVariable() && "Missing variable");
2281 translateDbgDeclareRecord(DI.getAddress(), DI.hasArgList(), DI.getVariable(),
2282 DI.getExpression(), DI.getDebugLoc(), MIRBuilder);
2283 return true;
2284 }
2285 case Intrinsic::dbg_label: {
2286 const DbgLabelInst &DI = cast<DbgLabelInst>(CI);
2287 assert(DI.getLabel() && "Missing label");
2288
2290 MIRBuilder.getDebugLoc()) &&
2291 "Expected inlined-at fields to agree");
2292
2293 MIRBuilder.buildDbgLabel(DI.getLabel());
2294 return true;
2295 }
2296 case Intrinsic::vaend:
2297 // No target I know of cares about va_end. Certainly no in-tree target
2298 // does. Simplest intrinsic ever!
2299 return true;
2300 case Intrinsic::vastart: {
2301 Value *Ptr = CI.getArgOperand(0);
2302 unsigned ListSize = TLI->getVaListSizeInBits(*DL) / 8;
2303 Align Alignment = getKnownAlignment(Ptr, *DL);
2304
2305 MIRBuilder.buildInstr(TargetOpcode::G_VASTART, {}, {getOrCreateVReg(*Ptr)})
2306 .addMemOperand(MF->getMachineMemOperand(MachinePointerInfo(Ptr),
2308 ListSize, Alignment));
2309 return true;
2310 }
2311 case Intrinsic::dbg_assign:
2312 // A dbg.assign is a dbg.value with more information about stack locations,
2313 // typically produced during optimisation of variables with leaked
2314 // addresses. We can treat it like a normal dbg_value intrinsic here; to
2315 // benefit from the full analysis of stack/SSA locations, GlobalISel would
2316 // need to register for and use the AssignmentTrackingAnalysis pass.
2317 [[fallthrough]];
2318 case Intrinsic::dbg_value: {
2319 // This form of DBG_VALUE is target-independent.
2320 const DbgValueInst &DI = cast<DbgValueInst>(CI);
2321 translateDbgValueRecord(DI.getValue(), DI.hasArgList(), DI.getVariable(),
2322 DI.getExpression(), DI.getDebugLoc(), MIRBuilder);
2323 return true;
2324 }
2325 case Intrinsic::uadd_with_overflow:
2326 return translateOverflowIntrinsic(CI, TargetOpcode::G_UADDO, MIRBuilder);
2327 case Intrinsic::sadd_with_overflow:
2328 return translateOverflowIntrinsic(CI, TargetOpcode::G_SADDO, MIRBuilder);
2329 case Intrinsic::usub_with_overflow:
2330 return translateOverflowIntrinsic(CI, TargetOpcode::G_USUBO, MIRBuilder);
2331 case Intrinsic::ssub_with_overflow:
2332 return translateOverflowIntrinsic(CI, TargetOpcode::G_SSUBO, MIRBuilder);
2333 case Intrinsic::umul_with_overflow:
2334 return translateOverflowIntrinsic(CI, TargetOpcode::G_UMULO, MIRBuilder);
2335 case Intrinsic::smul_with_overflow:
2336 return translateOverflowIntrinsic(CI, TargetOpcode::G_SMULO, MIRBuilder);
2337 case Intrinsic::uadd_sat:
2338 return translateBinaryOp(TargetOpcode::G_UADDSAT, CI, MIRBuilder);
2339 case Intrinsic::sadd_sat:
2340 return translateBinaryOp(TargetOpcode::G_SADDSAT, CI, MIRBuilder);
2341 case Intrinsic::usub_sat:
2342 return translateBinaryOp(TargetOpcode::G_USUBSAT, CI, MIRBuilder);
2343 case Intrinsic::ssub_sat:
2344 return translateBinaryOp(TargetOpcode::G_SSUBSAT, CI, MIRBuilder);
2345 case Intrinsic::ushl_sat:
2346 return translateBinaryOp(TargetOpcode::G_USHLSAT, CI, MIRBuilder);
2347 case Intrinsic::sshl_sat:
2348 return translateBinaryOp(TargetOpcode::G_SSHLSAT, CI, MIRBuilder);
2349 case Intrinsic::umin:
2350 return translateBinaryOp(TargetOpcode::G_UMIN, CI, MIRBuilder);
2351 case Intrinsic::umax:
2352 return translateBinaryOp(TargetOpcode::G_UMAX, CI, MIRBuilder);
2353 case Intrinsic::smin:
2354 return translateBinaryOp(TargetOpcode::G_SMIN, CI, MIRBuilder);
2355 case Intrinsic::smax:
2356 return translateBinaryOp(TargetOpcode::G_SMAX, CI, MIRBuilder);
2357 case Intrinsic::abs:
2358 // TODO: Preserve "int min is poison" arg in GMIR?
2359 return translateUnaryOp(TargetOpcode::G_ABS, CI, MIRBuilder);
2360 case Intrinsic::smul_fix:
2361 return translateFixedPointIntrinsic(TargetOpcode::G_SMULFIX, CI, MIRBuilder);
2362 case Intrinsic::umul_fix:
2363 return translateFixedPointIntrinsic(TargetOpcode::G_UMULFIX, CI, MIRBuilder);
2364 case Intrinsic::smul_fix_sat:
2365 return translateFixedPointIntrinsic(TargetOpcode::G_SMULFIXSAT, CI, MIRBuilder);
2366 case Intrinsic::umul_fix_sat:
2367 return translateFixedPointIntrinsic(TargetOpcode::G_UMULFIXSAT, CI, MIRBuilder);
2368 case Intrinsic::sdiv_fix:
2369 return translateFixedPointIntrinsic(TargetOpcode::G_SDIVFIX, CI, MIRBuilder);
2370 case Intrinsic::udiv_fix:
2371 return translateFixedPointIntrinsic(TargetOpcode::G_UDIVFIX, CI, MIRBuilder);
2372 case Intrinsic::sdiv_fix_sat:
2373 return translateFixedPointIntrinsic(TargetOpcode::G_SDIVFIXSAT, CI, MIRBuilder);
2374 case Intrinsic::udiv_fix_sat:
2375 return translateFixedPointIntrinsic(TargetOpcode::G_UDIVFIXSAT, CI, MIRBuilder);
2376 case Intrinsic::fmuladd: {
2377 const TargetMachine &TM = MF->getTarget();
2378 Register Dst = getOrCreateVReg(CI);
2379 Register Op0 = getOrCreateVReg(*CI.getArgOperand(0));
2380 Register Op1 = getOrCreateVReg(*CI.getArgOperand(1));
2381 Register Op2 = getOrCreateVReg(*CI.getArgOperand(2));
2383 TLI->isFMAFasterThanFMulAndFAdd(*MF,
2384 TLI->getValueType(*DL, CI.getType()))) {
2385 // TODO: Revisit this to see if we should move this part of the
2386 // lowering to the combiner.
2387 MIRBuilder.buildFMA(Dst, Op0, Op1, Op2,
2389 } else {
2390 LLT Ty = getLLTForType(*CI.getType(), *DL);
2391 auto FMul = MIRBuilder.buildFMul(
2392 Ty, Op0, Op1, MachineInstr::copyFlagsFromInstruction(CI));
2393 MIRBuilder.buildFAdd(Dst, FMul, Op2,
2395 }
2396 return true;
2397 }
2398 case Intrinsic::frexp: {
2399 ArrayRef<Register> VRegs = getOrCreateVRegs(CI);
2400 MIRBuilder.buildFFrexp(VRegs[0], VRegs[1],
2401 getOrCreateVReg(*CI.getArgOperand(0)),
2403 return true;
2404 }
2405 case Intrinsic::modf: {
2406 ArrayRef<Register> VRegs = getOrCreateVRegs(CI);
2407 MIRBuilder.buildModf(VRegs[0], VRegs[1],
2408 getOrCreateVReg(*CI.getArgOperand(0)),
2410 return true;
2411 }
2412 case Intrinsic::sincos: {
2413 ArrayRef<Register> VRegs = getOrCreateVRegs(CI);
2414 MIRBuilder.buildFSincos(VRegs[0], VRegs[1],
2415 getOrCreateVReg(*CI.getArgOperand(0)),
2417 return true;
2418 }
2419 case Intrinsic::fptosi_sat:
2420 MIRBuilder.buildFPTOSI_SAT(getOrCreateVReg(CI),
2421 getOrCreateVReg(*CI.getArgOperand(0)));
2422 return true;
2423 case Intrinsic::fptoui_sat:
2424 MIRBuilder.buildFPTOUI_SAT(getOrCreateVReg(CI),
2425 getOrCreateVReg(*CI.getArgOperand(0)));
2426 return true;
2427 case Intrinsic::memcpy_inline:
2428 return translateMemFunc(CI, MIRBuilder, TargetOpcode::G_MEMCPY_INLINE);
2429 case Intrinsic::memcpy:
2430 return translateMemFunc(CI, MIRBuilder, TargetOpcode::G_MEMCPY);
2431 case Intrinsic::memmove:
2432 return translateMemFunc(CI, MIRBuilder, TargetOpcode::G_MEMMOVE);
2433 case Intrinsic::memset:
2434 return translateMemFunc(CI, MIRBuilder, TargetOpcode::G_MEMSET);
2435 case Intrinsic::eh_typeid_for: {
2436 GlobalValue *GV = ExtractTypeInfo(CI.getArgOperand(0));
2437 Register Reg = getOrCreateVReg(CI);
2438 unsigned TypeID = MF->getTypeIDFor(GV);
2439 MIRBuilder.buildConstant(Reg, TypeID);
2440 return true;
2441 }
2442 case Intrinsic::objectsize:
2443 llvm_unreachable("llvm.objectsize.* should have been lowered already");
2444
2445 case Intrinsic::is_constant:
2446 llvm_unreachable("llvm.is.constant.* should have been lowered already");
2447
2448 case Intrinsic::stackguard:
2449 getStackGuard(getOrCreateVReg(CI), MIRBuilder);
2450 return true;
2451 case Intrinsic::stackprotector: {
2452 LLT PtrTy = getLLTForType(*CI.getArgOperand(0)->getType(), *DL);
2453 Register GuardVal;
2454 if (TLI->useLoadStackGuardNode(*CI.getModule())) {
2455 GuardVal = MRI->createGenericVirtualRegister(PtrTy);
2456 getStackGuard(GuardVal, MIRBuilder);
2457 } else
2458 GuardVal = getOrCreateVReg(*CI.getArgOperand(0)); // The guard's value.
2459
2460 AllocaInst *Slot = cast<AllocaInst>(CI.getArgOperand(1));
2461 int FI = getOrCreateFrameIndex(*Slot);
2462 MF->getFrameInfo().setStackProtectorIndex(FI);
2463
2464 MIRBuilder.buildStore(
2465 GuardVal, getOrCreateVReg(*Slot),
2466 *MF->getMachineMemOperand(MachinePointerInfo::getFixedStack(*MF, FI),
2469 PtrTy, Align(8)));
2470 return true;
2471 }
2472 case Intrinsic::stacksave: {
2473 MIRBuilder.buildInstr(TargetOpcode::G_STACKSAVE, {getOrCreateVReg(CI)}, {});
2474 return true;
2475 }
2476 case Intrinsic::stackrestore: {
2477 MIRBuilder.buildInstr(TargetOpcode::G_STACKRESTORE, {},
2478 {getOrCreateVReg(*CI.getArgOperand(0))});
2479 return true;
2480 }
2481 case Intrinsic::cttz:
2482 case Intrinsic::ctlz: {
2483 ConstantInt *Cst = cast<ConstantInt>(CI.getArgOperand(1));
2484 bool isTrailing = ID == Intrinsic::cttz;
2485 unsigned Opcode = isTrailing
2486 ? Cst->isZero() ? TargetOpcode::G_CTTZ
2487 : TargetOpcode::G_CTTZ_ZERO_UNDEF
2488 : Cst->isZero() ? TargetOpcode::G_CTLZ
2489 : TargetOpcode::G_CTLZ_ZERO_UNDEF;
2490 MIRBuilder.buildInstr(Opcode, {getOrCreateVReg(CI)},
2491 {getOrCreateVReg(*CI.getArgOperand(0))});
2492 return true;
2493 }
2494 case Intrinsic::invariant_start: {
2495 MIRBuilder.buildUndef(getOrCreateVReg(CI));
2496 return true;
2497 }
2498 case Intrinsic::invariant_end:
2499 return true;
2500 case Intrinsic::expect:
2501 case Intrinsic::expect_with_probability:
2502 case Intrinsic::annotation:
2503 case Intrinsic::ptr_annotation:
2504 case Intrinsic::launder_invariant_group:
2505 case Intrinsic::strip_invariant_group: {
2506 // Drop the intrinsic, but forward the value.
2507 MIRBuilder.buildCopy(getOrCreateVReg(CI),
2508 getOrCreateVReg(*CI.getArgOperand(0)));
2509 return true;
2510 }
2511 case Intrinsic::assume:
2512 case Intrinsic::experimental_noalias_scope_decl:
2513 case Intrinsic::var_annotation:
2514 case Intrinsic::sideeffect:
2515 // Discard annotate attributes, assumptions, and artificial side-effects.
2516 return true;
2517 case Intrinsic::read_volatile_register:
2518 case Intrinsic::read_register: {
2519 Value *Arg = CI.getArgOperand(0);
2520 MIRBuilder
2521 .buildInstr(TargetOpcode::G_READ_REGISTER, {getOrCreateVReg(CI)}, {})
2522 .addMetadata(cast<MDNode>(cast<MetadataAsValue>(Arg)->getMetadata()));
2523 return true;
2524 }
2525 case Intrinsic::write_register: {
2526 Value *Arg = CI.getArgOperand(0);
2527 MIRBuilder.buildInstr(TargetOpcode::G_WRITE_REGISTER)
2528 .addMetadata(cast<MDNode>(cast<MetadataAsValue>(Arg)->getMetadata()))
2529 .addUse(getOrCreateVReg(*CI.getArgOperand(1)));
2530 return true;
2531 }
2532 case Intrinsic::localescape: {
2533 MachineBasicBlock &EntryMBB = MF->front();
2534 StringRef EscapedName = GlobalValue::dropLLVMManglingEscape(MF->getName());
2535
2536 // Directly emit some LOCAL_ESCAPE machine instrs. Label assignment emission
2537 // is the same on all targets.
2538 for (unsigned Idx = 0, E = CI.arg_size(); Idx < E; ++Idx) {
2539 Value *Arg = CI.getArgOperand(Idx)->stripPointerCasts();
2540 if (isa<ConstantPointerNull>(Arg))
2541 continue; // Skip null pointers. They represent a hole in index space.
2542
2543 int FI = getOrCreateFrameIndex(*cast<AllocaInst>(Arg));
2544 MCSymbol *FrameAllocSym =
2545 MF->getContext().getOrCreateFrameAllocSymbol(EscapedName, Idx);
2546
2547 // This should be inserted at the start of the entry block.
2548 auto LocalEscape =
2549 MIRBuilder.buildInstrNoInsert(TargetOpcode::LOCAL_ESCAPE)
2550 .addSym(FrameAllocSym)
2551 .addFrameIndex(FI);
2552
2553 EntryMBB.insert(EntryMBB.begin(), LocalEscape);
2554 }
2555
2556 return true;
2557 }
2558 case Intrinsic::vector_reduce_fadd:
2559 case Intrinsic::vector_reduce_fmul: {
2560 // Need to check for the reassoc flag to decide whether we want a
2561 // sequential reduction opcode or not.
2562 Register Dst = getOrCreateVReg(CI);
2563 Register ScalarSrc = getOrCreateVReg(*CI.getArgOperand(0));
2564 Register VecSrc = getOrCreateVReg(*CI.getArgOperand(1));
2565 unsigned Opc = 0;
2566 if (!CI.hasAllowReassoc()) {
2567 // The sequential ordering case.
2568 Opc = ID == Intrinsic::vector_reduce_fadd
2569 ? TargetOpcode::G_VECREDUCE_SEQ_FADD
2570 : TargetOpcode::G_VECREDUCE_SEQ_FMUL;
2571 if (!MRI->getType(VecSrc).isVector())
2572 Opc = ID == Intrinsic::vector_reduce_fadd ? TargetOpcode::G_FADD
2573 : TargetOpcode::G_FMUL;
2574 MIRBuilder.buildInstr(Opc, {Dst}, {ScalarSrc, VecSrc},
2576 return true;
2577 }
2578 // We split the operation into a separate G_FADD/G_FMUL + the reduce,
2579 // since the associativity doesn't matter.
2580 unsigned ScalarOpc;
2581 if (ID == Intrinsic::vector_reduce_fadd) {
2582 Opc = TargetOpcode::G_VECREDUCE_FADD;
2583 ScalarOpc = TargetOpcode::G_FADD;
2584 } else {
2585 Opc = TargetOpcode::G_VECREDUCE_FMUL;
2586 ScalarOpc = TargetOpcode::G_FMUL;
2587 }
2588 LLT DstTy = MRI->getType(Dst);
2589 auto Rdx = MIRBuilder.buildInstr(
2590 Opc, {DstTy}, {VecSrc}, MachineInstr::copyFlagsFromInstruction(CI));
2591 MIRBuilder.buildInstr(ScalarOpc, {Dst}, {ScalarSrc, Rdx},
2593
2594 return true;
2595 }
2596 case Intrinsic::trap:
2597 return translateTrap(CI, MIRBuilder, TargetOpcode::G_TRAP);
2598 case Intrinsic::debugtrap:
2599 return translateTrap(CI, MIRBuilder, TargetOpcode::G_DEBUGTRAP);
2600 case Intrinsic::ubsantrap:
2601 return translateTrap(CI, MIRBuilder, TargetOpcode::G_UBSANTRAP);
2602 case Intrinsic::allow_runtime_check:
2603 case Intrinsic::allow_ubsan_check:
2604 MIRBuilder.buildCopy(getOrCreateVReg(CI),
2605 getOrCreateVReg(*ConstantInt::getTrue(CI.getType())));
2606 return true;
2607 case Intrinsic::amdgcn_cs_chain:
2608 case Intrinsic::amdgcn_call_whole_wave:
2609 return translateCallBase(CI, MIRBuilder);
2610 case Intrinsic::fptrunc_round: {
2612
2613 // Convert the metadata argument to a constant integer
2614 Metadata *MD = cast<MetadataAsValue>(CI.getArgOperand(1))->getMetadata();
2615 std::optional<RoundingMode> RoundMode =
2616 convertStrToRoundingMode(cast<MDString>(MD)->getString());
2617
2618 // Add the Rounding mode as an integer
2619 MIRBuilder
2620 .buildInstr(TargetOpcode::G_INTRINSIC_FPTRUNC_ROUND,
2621 {getOrCreateVReg(CI)},
2622 {getOrCreateVReg(*CI.getArgOperand(0))}, Flags)
2623 .addImm((int)*RoundMode);
2624
2625 return true;
2626 }
2627 case Intrinsic::is_fpclass: {
2628 Value *FpValue = CI.getOperand(0);
2629 ConstantInt *TestMaskValue = cast<ConstantInt>(CI.getOperand(1));
2630
2631 MIRBuilder
2632 .buildInstr(TargetOpcode::G_IS_FPCLASS, {getOrCreateVReg(CI)},
2633 {getOrCreateVReg(*FpValue)})
2634 .addImm(TestMaskValue->getZExtValue());
2635
2636 return true;
2637 }
2638 case Intrinsic::set_fpenv: {
2639 Value *FPEnv = CI.getOperand(0);
2640 MIRBuilder.buildSetFPEnv(getOrCreateVReg(*FPEnv));
2641 return true;
2642 }
2643 case Intrinsic::reset_fpenv:
2644 MIRBuilder.buildResetFPEnv();
2645 return true;
2646 case Intrinsic::set_fpmode: {
2647 Value *FPState = CI.getOperand(0);
2648 MIRBuilder.buildSetFPMode(getOrCreateVReg(*FPState));
2649 return true;
2650 }
2651 case Intrinsic::reset_fpmode:
2652 MIRBuilder.buildResetFPMode();
2653 return true;
2654 case Intrinsic::get_rounding:
2655 MIRBuilder.buildGetRounding(getOrCreateVReg(CI));
2656 return true;
2657 case Intrinsic::set_rounding:
2658 MIRBuilder.buildSetRounding(getOrCreateVReg(*CI.getOperand(0)));
2659 return true;
2660 case Intrinsic::vscale: {
2661 MIRBuilder.buildVScale(getOrCreateVReg(CI), 1);
2662 return true;
2663 }
2664 case Intrinsic::scmp:
2665 MIRBuilder.buildSCmp(getOrCreateVReg(CI),
2666 getOrCreateVReg(*CI.getOperand(0)),
2667 getOrCreateVReg(*CI.getOperand(1)));
2668 return true;
2669 case Intrinsic::ucmp:
2670 MIRBuilder.buildUCmp(getOrCreateVReg(CI),
2671 getOrCreateVReg(*CI.getOperand(0)),
2672 getOrCreateVReg(*CI.getOperand(1)));
2673 return true;
2674 case Intrinsic::vector_extract:
2675 return translateExtractVector(CI, MIRBuilder);
2676 case Intrinsic::vector_insert:
2677 return translateInsertVector(CI, MIRBuilder);
2678 case Intrinsic::stepvector: {
2679 MIRBuilder.buildStepVector(getOrCreateVReg(CI), 1);
2680 return true;
2681 }
2682 case Intrinsic::prefetch: {
2683 Value *Addr = CI.getOperand(0);
2684 unsigned RW = cast<ConstantInt>(CI.getOperand(1))->getZExtValue();
2685 unsigned Locality = cast<ConstantInt>(CI.getOperand(2))->getZExtValue();
2686 unsigned CacheType = cast<ConstantInt>(CI.getOperand(3))->getZExtValue();
2687
2689 auto &MMO = *MF->getMachineMemOperand(MachinePointerInfo(Addr), Flags,
2690 LLT(), Align());
2691
2692 MIRBuilder.buildPrefetch(getOrCreateVReg(*Addr), RW, Locality, CacheType,
2693 MMO);
2694
2695 return true;
2696 }
2697
2698 case Intrinsic::vector_interleave2:
2699 case Intrinsic::vector_deinterleave2: {
2700 // Both intrinsics have at least one operand.
2701 Value *Op0 = CI.getOperand(0);
2702 LLT ResTy = getLLTForType(*Op0->getType(), MIRBuilder.getDataLayout());
2703 if (!ResTy.isFixedVector())
2704 return false;
2705
2706 if (CI.getIntrinsicID() == Intrinsic::vector_interleave2)
2707 return translateVectorInterleave2Intrinsic(CI, MIRBuilder);
2708
2709 return translateVectorDeinterleave2Intrinsic(CI, MIRBuilder);
2710 }
2711
2712#define INSTRUCTION(NAME, NARG, ROUND_MODE, INTRINSIC) \
2713 case Intrinsic::INTRINSIC:
2714#include "llvm/IR/ConstrainedOps.def"
2715 return translateConstrainedFPIntrinsic(cast<ConstrainedFPIntrinsic>(CI),
2716 MIRBuilder);
2717 case Intrinsic::experimental_convergence_anchor:
2718 case Intrinsic::experimental_convergence_entry:
2719 case Intrinsic::experimental_convergence_loop:
2720 return translateConvergenceControlIntrinsic(CI, ID, MIRBuilder);
2721 case Intrinsic::reloc_none: {
2722 Metadata *MD = cast<MetadataAsValue>(CI.getArgOperand(0))->getMetadata();
2723 StringRef SymbolName = cast<MDString>(MD)->getString();
2724 MIRBuilder.buildInstr(TargetOpcode::RELOC_NONE)
2726 return true;
2727 }
2728 }
2729 return false;
2730}
2731
2732bool IRTranslator::translateInlineAsm(const CallBase &CB,
2733 MachineIRBuilder &MIRBuilder) {
2734 if (!mayTranslateUserTypes(CB))
2735 return false;
2736
2737 const InlineAsmLowering *ALI = MF->getSubtarget().getInlineAsmLowering();
2738
2739 if (!ALI) {
2740 LLVM_DEBUG(
2741 dbgs() << "Inline asm lowering is not supported for this target yet\n");
2742 return false;
2743 }
2744
2745 return ALI->lowerInlineAsm(
2746 MIRBuilder, CB, [&](const Value &Val) { return getOrCreateVRegs(Val); });
2747}
2748
2749bool IRTranslator::translateCallBase(const CallBase &CB,
2750 MachineIRBuilder &MIRBuilder) {
2751 ArrayRef<Register> Res = getOrCreateVRegs(CB);
2752
2754 Register SwiftInVReg = 0;
2755 Register SwiftErrorVReg = 0;
2756 for (const auto &Arg : CB.args()) {
2757 if (CLI->supportSwiftError() && isSwiftError(Arg)) {
2758 assert(SwiftInVReg == 0 && "Expected only one swift error argument");
2759 LLT Ty = getLLTForType(*Arg->getType(), *DL);
2760 SwiftInVReg = MRI->createGenericVirtualRegister(Ty);
2761 MIRBuilder.buildCopy(SwiftInVReg, SwiftError.getOrCreateVRegUseAt(
2762 &CB, &MIRBuilder.getMBB(), Arg));
2763 Args.emplace_back(ArrayRef(SwiftInVReg));
2764 SwiftErrorVReg =
2765 SwiftError.getOrCreateVRegDefAt(&CB, &MIRBuilder.getMBB(), Arg);
2766 continue;
2767 }
2768 Args.push_back(getOrCreateVRegs(*Arg));
2769 }
2770
2771 if (auto *CI = dyn_cast<CallInst>(&CB)) {
2772 if (ORE->enabled()) {
2773 if (MemoryOpRemark::canHandle(CI, *LibInfo)) {
2774 MemoryOpRemark R(*ORE, "gisel-irtranslator-memsize", *DL, *LibInfo);
2775 R.visit(CI);
2776 }
2777 }
2778 }
2779
2780 std::optional<CallLowering::PtrAuthInfo> PAI;
2781 if (auto Bundle = CB.getOperandBundle(LLVMContext::OB_ptrauth)) {
2782 // Functions should never be ptrauth-called directly.
2783 assert(!CB.getCalledFunction() && "invalid direct ptrauth call");
2784
2785 const Value *Key = Bundle->Inputs[0];
2786 const Value *Discriminator = Bundle->Inputs[1];
2787
2788 // Look through ptrauth constants to try to eliminate the matching bundle
2789 // and turn this into a direct call with no ptrauth.
2790 // CallLowering will use the raw pointer if it doesn't find the PAI.
2791 const auto *CalleeCPA = dyn_cast<ConstantPtrAuth>(CB.getCalledOperand());
2792 if (!CalleeCPA || !isa<Function>(CalleeCPA->getPointer()) ||
2793 !CalleeCPA->isKnownCompatibleWith(Key, Discriminator, *DL)) {
2794 // If we can't make it direct, package the bundle into PAI.
2795 Register DiscReg = getOrCreateVReg(*Discriminator);
2796 PAI = CallLowering::PtrAuthInfo{cast<ConstantInt>(Key)->getZExtValue(),
2797 DiscReg};
2798 }
2799 }
2800
2801 Register ConvergenceCtrlToken = 0;
2802 if (auto Bundle = CB.getOperandBundle(LLVMContext::OB_convergencectrl)) {
2803 const auto &Token = *Bundle->Inputs[0].get();
2804 ConvergenceCtrlToken = getOrCreateConvergenceTokenVReg(Token);
2805 }
2806
2807 // We don't set HasCalls on MFI here yet because call lowering may decide to
2808 // optimize into tail calls. Instead, we defer that to selection where a final
2809 // scan is done to check if any instructions are calls.
2810 bool Success = CLI->lowerCall(
2811 MIRBuilder, CB, Res, Args, SwiftErrorVReg, PAI, ConvergenceCtrlToken,
2812 [&]() { return getOrCreateVReg(*CB.getCalledOperand()); });
2813
2814 // Check if we just inserted a tail call.
2815 if (Success) {
2816 assert(!HasTailCall && "Can't tail call return twice from block?");
2817 const TargetInstrInfo *TII = MF->getSubtarget().getInstrInfo();
2818 HasTailCall = TII->isTailCall(*std::prev(MIRBuilder.getInsertPt()));
2819 }
2820
2821 return Success;
2822}
2823
2824bool IRTranslator::translateCall(const User &U, MachineIRBuilder &MIRBuilder) {
2825 if (!mayTranslateUserTypes(U))
2826 return false;
2827
2828 const CallInst &CI = cast<CallInst>(U);
2829 const Function *F = CI.getCalledFunction();
2830
2831 // FIXME: support Windows dllimport function calls and calls through
2832 // weak symbols.
2833 if (F && (F->hasDLLImportStorageClass() ||
2834 (MF->getTarget().getTargetTriple().isOSWindows() &&
2835 F->hasExternalWeakLinkage())))
2836 return false;
2837
2838 // FIXME: support control flow guard targets.
2840 return false;
2841
2842 // FIXME: support statepoints and related.
2844 return false;
2845
2846 if (CI.isInlineAsm())
2847 return translateInlineAsm(CI, MIRBuilder);
2848
2849 Intrinsic::ID ID = F ? F->getIntrinsicID() : Intrinsic::not_intrinsic;
2850 if (!F || ID == Intrinsic::not_intrinsic) {
2851 if (translateCallBase(CI, MIRBuilder)) {
2852 diagnoseDontCall(CI);
2853 return true;
2854 }
2855 return false;
2856 }
2857
2858 assert(ID != Intrinsic::not_intrinsic && "unknown intrinsic");
2859
2860 if (translateKnownIntrinsic(CI, ID, MIRBuilder))
2861 return true;
2862
2864 TLI->getTgtMemIntrinsic(Infos, CI, *MF, ID);
2865
2866 return translateIntrinsic(CI, ID, MIRBuilder, Infos);
2867}
2868
2869/// Translate a call or callbr to an intrinsic.
2870bool IRTranslator::translateIntrinsic(
2871 const CallBase &CB, Intrinsic::ID ID, MachineIRBuilder &MIRBuilder,
2872 ArrayRef<TargetLowering::IntrinsicInfo> TgtMemIntrinsicInfos) {
2873 ArrayRef<Register> ResultRegs;
2874 if (!CB.getType()->isVoidTy())
2875 ResultRegs = getOrCreateVRegs(CB);
2876
2877 // Ignore the callsite attributes. Backend code is most likely not expecting
2878 // an intrinsic to sometimes have side effects and sometimes not.
2879 MachineInstrBuilder MIB = MIRBuilder.buildIntrinsic(ID, ResultRegs);
2880 if (isa<FPMathOperator>(CB))
2881 MIB->copyIRFlags(CB);
2882
2883 for (const auto &Arg : enumerate(CB.args())) {
2884 // If this is required to be an immediate, don't materialize it in a
2885 // register.
2886 if (CB.paramHasAttr(Arg.index(), Attribute::ImmArg)) {
2887 if (ConstantInt *CI = dyn_cast<ConstantInt>(Arg.value())) {
2888 // imm arguments are more convenient than cimm (and realistically
2889 // probably sufficient), so use them.
2890 assert(CI->getBitWidth() <= 64 &&
2891 "large intrinsic immediates not handled");
2892 MIB.addImm(CI->getSExtValue());
2893 } else {
2894 MIB.addFPImm(cast<ConstantFP>(Arg.value()));
2895 }
2896 } else if (auto *MDVal = dyn_cast<MetadataAsValue>(Arg.value())) {
2897 auto *MD = MDVal->getMetadata();
2898 auto *MDN = dyn_cast<MDNode>(MD);
2899 if (!MDN) {
2900 if (auto *ConstMD = dyn_cast<ConstantAsMetadata>(MD))
2901 MDN = MDNode::get(MF->getFunction().getContext(), ConstMD);
2902 else // This was probably an MDString.
2903 return false;
2904 }
2905 MIB.addMetadata(MDN);
2906 } else {
2907 ArrayRef<Register> VRegs = getOrCreateVRegs(*Arg.value());
2908 if (VRegs.size() > 1)
2909 return false;
2910 MIB.addUse(VRegs[0]);
2911 }
2912 }
2913
2914 // Add MachineMemOperands for each memory access described by the target.
2915 for (const auto &Info : TgtMemIntrinsicInfos) {
2916 Align Alignment = Info.align.value_or(
2917 DL->getABITypeAlign(Info.memVT.getTypeForEVT(CB.getContext())));
2918 LLT MemTy = Info.memVT.isSimple()
2919 ? getLLTForMVT(Info.memVT.getSimpleVT())
2920 : LLT::scalar(Info.memVT.getStoreSizeInBits());
2921
2922 // TODO: We currently just fallback to address space 0 if
2923 // getTgtMemIntrinsic didn't yield anything useful.
2924 MachinePointerInfo MPI;
2925 if (Info.ptrVal) {
2926 MPI = MachinePointerInfo(Info.ptrVal, Info.offset);
2927 } else if (Info.fallbackAddressSpace) {
2928 MPI = MachinePointerInfo(*Info.fallbackAddressSpace);
2929 }
2930 MIB.addMemOperand(MF->getMachineMemOperand(
2931 MPI, Info.flags, MemTy, Alignment, CB.getAAMetadata(),
2932 /*Ranges=*/nullptr, Info.ssid, Info.order, Info.failureOrder));
2933 }
2934
2935 if (CB.isConvergent()) {
2936 if (auto Bundle = CB.getOperandBundle(LLVMContext::OB_convergencectrl)) {
2937 auto *Token = Bundle->Inputs[0].get();
2938 Register TokenReg = getOrCreateVReg(*Token);
2939 MIB.addUse(TokenReg, RegState::Implicit);
2940 }
2941 }
2942
2944 MIB->setDeactivationSymbol(*MF, Bundle->Inputs[0].get());
2945
2946 return true;
2947}
2948
2949bool IRTranslator::findUnwindDestinations(
2950 const BasicBlock *EHPadBB,
2951 BranchProbability Prob,
2952 SmallVectorImpl<std::pair<MachineBasicBlock *, BranchProbability>>
2953 &UnwindDests) {
2955 EHPadBB->getParent()->getFunction().getPersonalityFn());
2956 bool IsMSVCCXX = Personality == EHPersonality::MSVC_CXX;
2957 bool IsCoreCLR = Personality == EHPersonality::CoreCLR;
2958 bool IsWasmCXX = Personality == EHPersonality::Wasm_CXX;
2959 bool IsSEH = isAsynchronousEHPersonality(Personality);
2960
2961 if (IsWasmCXX) {
2962 // Ignore this for now.
2963 return false;
2964 }
2965
2966 while (EHPadBB) {
2968 BasicBlock *NewEHPadBB = nullptr;
2969 if (isa<LandingPadInst>(Pad)) {
2970 // Stop on landingpads. They are not funclets.
2971 UnwindDests.emplace_back(&getMBB(*EHPadBB), Prob);
2972 break;
2973 }
2974 if (isa<CleanupPadInst>(Pad)) {
2975 // Stop on cleanup pads. Cleanups are always funclet entries for all known
2976 // personalities.
2977 UnwindDests.emplace_back(&getMBB(*EHPadBB), Prob);
2978 UnwindDests.back().first->setIsEHScopeEntry();
2979 UnwindDests.back().first->setIsEHFuncletEntry();
2980 break;
2981 }
2982 if (auto *CatchSwitch = dyn_cast<CatchSwitchInst>(Pad)) {
2983 // Add the catchpad handlers to the possible destinations.
2984 for (const BasicBlock *CatchPadBB : CatchSwitch->handlers()) {
2985 UnwindDests.emplace_back(&getMBB(*CatchPadBB), Prob);
2986 // For MSVC++ and the CLR, catchblocks are funclets and need prologues.
2987 if (IsMSVCCXX || IsCoreCLR)
2988 UnwindDests.back().first->setIsEHFuncletEntry();
2989 if (!IsSEH)
2990 UnwindDests.back().first->setIsEHScopeEntry();
2991 }
2992 NewEHPadBB = CatchSwitch->getUnwindDest();
2993 } else {
2994 continue;
2995 }
2996
2997 BranchProbabilityInfo *BPI = FuncInfo.BPI;
2998 if (BPI && NewEHPadBB)
2999 Prob *= BPI->getEdgeProbability(EHPadBB, NewEHPadBB);
3000 EHPadBB = NewEHPadBB;
3001 }
3002 return true;
3003}
3004
3005bool IRTranslator::translateInvoke(const User &U,
3006 MachineIRBuilder &MIRBuilder) {
3007 const InvokeInst &I = cast<InvokeInst>(U);
3008 MCContext &Context = MF->getContext();
3009
3010 const BasicBlock *ReturnBB = I.getSuccessor(0);
3011 const BasicBlock *EHPadBB = I.getSuccessor(1);
3012
3013 const Function *Fn = I.getCalledFunction();
3014
3015 // FIXME: support invoking patchpoint and statepoint intrinsics.
3016 if (Fn && Fn->isIntrinsic())
3017 return false;
3018
3019 // FIXME: support whatever these are.
3020 if (I.hasDeoptState())
3021 return false;
3022
3023 // FIXME: support control flow guard targets.
3024 if (I.countOperandBundlesOfType(LLVMContext::OB_cfguardtarget))
3025 return false;
3026
3027 // FIXME: support Windows exception handling.
3028 if (!isa<LandingPadInst>(EHPadBB->getFirstNonPHIIt()))
3029 return false;
3030
3031 // FIXME: support Windows dllimport function calls and calls through
3032 // weak symbols.
3033 if (Fn && (Fn->hasDLLImportStorageClass() ||
3034 (MF->getTarget().getTargetTriple().isOSWindows() &&
3035 Fn->hasExternalWeakLinkage())))
3036 return false;
3037
3038 bool LowerInlineAsm = I.isInlineAsm();
3039 bool NeedEHLabel = true;
3040
3041 // Emit the actual call, bracketed by EH_LABELs so that the MF knows about
3042 // the region covered by the try.
3043 MCSymbol *BeginSymbol = nullptr;
3044 if (NeedEHLabel) {
3045 MIRBuilder.buildInstr(TargetOpcode::G_INVOKE_REGION_START);
3046 BeginSymbol = Context.createTempSymbol();
3047 MIRBuilder.buildInstr(TargetOpcode::EH_LABEL).addSym(BeginSymbol);
3048 }
3049
3050 if (LowerInlineAsm) {
3051 if (!translateInlineAsm(I, MIRBuilder))
3052 return false;
3053 } else if (!translateCallBase(I, MIRBuilder))
3054 return false;
3055
3056 MCSymbol *EndSymbol = nullptr;
3057 if (NeedEHLabel) {
3058 EndSymbol = Context.createTempSymbol();
3059 MIRBuilder.buildInstr(TargetOpcode::EH_LABEL).addSym(EndSymbol);
3060 }
3061
3063 BranchProbabilityInfo *BPI = FuncInfo.BPI;
3064 MachineBasicBlock *InvokeMBB = &MIRBuilder.getMBB();
3065 BranchProbability EHPadBBProb =
3066 BPI ? BPI->getEdgeProbability(InvokeMBB->getBasicBlock(), EHPadBB)
3068
3069 if (!findUnwindDestinations(EHPadBB, EHPadBBProb, UnwindDests))
3070 return false;
3071
3072 MachineBasicBlock &EHPadMBB = getMBB(*EHPadBB),
3073 &ReturnMBB = getMBB(*ReturnBB);
3074 // Update successor info.
3075 addSuccessorWithProb(InvokeMBB, &ReturnMBB);
3076 for (auto &UnwindDest : UnwindDests) {
3077 UnwindDest.first->setIsEHPad();
3078 addSuccessorWithProb(InvokeMBB, UnwindDest.first, UnwindDest.second);
3079 }
3080 InvokeMBB->normalizeSuccProbs();
3081
3082 if (NeedEHLabel) {
3083 assert(BeginSymbol && "Expected a begin symbol!");
3084 assert(EndSymbol && "Expected an end symbol!");
3085 MF->addInvoke(&EHPadMBB, BeginSymbol, EndSymbol);
3086 }
3087
3088 MIRBuilder.buildBr(ReturnMBB);
3089 return true;
3090}
3091
3092/// The intrinsics currently supported by callbr are implicit control flow
3093/// intrinsics such as amdgcn.kill.
3094bool IRTranslator::translateCallBr(const User &U,
3095 MachineIRBuilder &MIRBuilder) {
3096 if (!mayTranslateUserTypes(U))
3097 return false; // see translateCall
3098
3099 const CallBrInst &I = cast<CallBrInst>(U);
3100 MachineBasicBlock *CallBrMBB = &MIRBuilder.getMBB();
3101
3102 Intrinsic::ID IID = I.getIntrinsicID();
3103 if (I.isInlineAsm()) {
3104 // FIXME: inline asm is not yet supported for callbr in GlobalISel. As soon
3105 // as we add support, we need to handle the indirect asm targets, see
3106 // SelectionDAGBuilder::visitCallBr().
3107 return false;
3108 }
3109 if (!translateIntrinsic(I, IID, MIRBuilder))
3110 return false;
3111
3112 // Retrieve successors.
3113 SmallPtrSet<BasicBlock *, 8> Dests = {I.getDefaultDest()};
3114 MachineBasicBlock *Return = &getMBB(*I.getDefaultDest());
3115
3116 // Update successor info.
3117 addSuccessorWithProb(CallBrMBB, Return, BranchProbability::getOne());
3118
3119 // Add indirect targets as successors. For intrinsic callbr, these represent
3120 // implicit control flow (e.g., the "kill" path for amdgcn.kill). We mark them
3121 // with setIsInlineAsmBrIndirectTarget so the machine verifier accepts them as
3122 // valid successors, even though they're not from inline asm.
3123 for (BasicBlock *Dest : I.getIndirectDests()) {
3124 MachineBasicBlock &Target = getMBB(*Dest);
3125 Target.setIsInlineAsmBrIndirectTarget();
3126 Target.setLabelMustBeEmitted();
3127 // Don't add duplicate machine successors.
3128 if (Dests.insert(Dest).second)
3129 addSuccessorWithProb(CallBrMBB, &Target, BranchProbability::getZero());
3130 }
3131
3132 CallBrMBB->normalizeSuccProbs();
3133
3134 // Drop into default successor.
3135 MIRBuilder.buildBr(*Return);
3136
3137 return true;
3138}
3139
3140bool IRTranslator::translateLandingPad(const User &U,
3141 MachineIRBuilder &MIRBuilder) {
3142 const LandingPadInst &LP = cast<LandingPadInst>(U);
3143
3144 MachineBasicBlock &MBB = MIRBuilder.getMBB();
3145
3146 MBB.setIsEHPad();
3147
3148 // If there aren't registers to copy the values into (e.g., during SjLj
3149 // exceptions), then don't bother.
3150 const Constant *PersonalityFn = MF->getFunction().getPersonalityFn();
3151 if (TLI->getExceptionPointerRegister(PersonalityFn) == 0 &&
3152 TLI->getExceptionSelectorRegister(PersonalityFn) == 0)
3153 return true;
3154
3155 // If landingpad's return type is token type, we don't create DAG nodes
3156 // for its exception pointer and selector value. The extraction of exception
3157 // pointer or selector value from token type landingpads is not currently
3158 // supported.
3159 if (LP.getType()->isTokenTy())
3160 return true;
3161
3162 // Add a label to mark the beginning of the landing pad. Deletion of the
3163 // landing pad can thus be detected via the MachineModuleInfo.
3164 MIRBuilder.buildInstr(TargetOpcode::EH_LABEL)
3165 .addSym(MF->addLandingPad(&MBB));
3166
3167 // If the unwinder does not preserve all registers, ensure that the
3168 // function marks the clobbered registers as used.
3169 const TargetRegisterInfo &TRI = *MF->getSubtarget().getRegisterInfo();
3170 if (auto *RegMask = TRI.getCustomEHPadPreservedMask(*MF))
3171 MF->getRegInfo().addPhysRegsUsedFromRegMask(RegMask);
3172
3173 LLT Ty = getLLTForType(*LP.getType(), *DL);
3174 Register Undef = MRI->createGenericVirtualRegister(Ty);
3175 MIRBuilder.buildUndef(Undef);
3176
3178 for (Type *Ty : cast<StructType>(LP.getType())->elements())
3179 Tys.push_back(getLLTForType(*Ty, *DL));
3180 assert(Tys.size() == 2 && "Only two-valued landingpads are supported");
3181
3182 // Mark exception register as live in.
3183 Register ExceptionReg = TLI->getExceptionPointerRegister(PersonalityFn);
3184 if (!ExceptionReg)
3185 return false;
3186
3187 MBB.addLiveIn(ExceptionReg);
3188 ArrayRef<Register> ResRegs = getOrCreateVRegs(LP);
3189 MIRBuilder.buildCopy(ResRegs[0], ExceptionReg);
3190
3191 Register SelectorReg = TLI->getExceptionSelectorRegister(PersonalityFn);
3192 if (!SelectorReg)
3193 return false;
3194
3195 MBB.addLiveIn(SelectorReg);
3196 Register PtrVReg = MRI->createGenericVirtualRegister(Tys[0]);
3197 MIRBuilder.buildCopy(PtrVReg, SelectorReg);
3198 MIRBuilder.buildCast(ResRegs[1], PtrVReg);
3199
3200 return true;
3201}
3202
3203bool IRTranslator::translateAlloca(const User &U,
3204 MachineIRBuilder &MIRBuilder) {
3205 auto &AI = cast<AllocaInst>(U);
3206
3207 if (AI.isSwiftError())
3208 return true;
3209
3210 if (AI.isStaticAlloca()) {
3211 Register Res = getOrCreateVReg(AI);
3212 int FI = getOrCreateFrameIndex(AI);
3213 MIRBuilder.buildFrameIndex(Res, FI);
3214 return true;
3215 }
3216
3217 // FIXME: support stack probing for Windows.
3218 if (MF->getTarget().getTargetTriple().isOSWindows())
3219 return false;
3220
3221 // Now we're in the harder dynamic case.
3222 Register NumElts = getOrCreateVReg(*AI.getArraySize());
3223 Type *IntPtrIRTy = DL->getIntPtrType(AI.getType());
3224 LLT IntPtrTy = getLLTForType(*IntPtrIRTy, *DL);
3225 if (MRI->getType(NumElts) != IntPtrTy) {
3226 Register ExtElts = MRI->createGenericVirtualRegister(IntPtrTy);
3227 MIRBuilder.buildZExtOrTrunc(ExtElts, NumElts);
3228 NumElts = ExtElts;
3229 }
3230
3231 Type *Ty = AI.getAllocatedType();
3232 TypeSize TySize = DL->getTypeAllocSize(Ty);
3233
3234 Register AllocSize = MRI->createGenericVirtualRegister(IntPtrTy);
3235 Register TySizeReg;
3236 if (TySize.isScalable()) {
3237 // For scalable types, use vscale * min_value
3238 TySizeReg = MRI->createGenericVirtualRegister(IntPtrTy);
3239 MIRBuilder.buildVScale(TySizeReg, TySize.getKnownMinValue());
3240 } else {
3241 // For fixed types, use a constant
3242 TySizeReg =
3243 getOrCreateVReg(*ConstantInt::get(IntPtrIRTy, TySize.getFixedValue()));
3244 }
3245 MIRBuilder.buildMul(AllocSize, NumElts, TySizeReg);
3246
3247 // Round the size of the allocation up to the stack alignment size
3248 // by add SA-1 to the size. This doesn't overflow because we're computing
3249 // an address inside an alloca.
3250 Align StackAlign = MF->getSubtarget().getFrameLowering()->getStackAlign();
3251 auto SAMinusOne = MIRBuilder.buildConstant(IntPtrTy, StackAlign.value() - 1);
3252 auto AllocAdd = MIRBuilder.buildAdd(IntPtrTy, AllocSize, SAMinusOne,
3254 auto AlignCst =
3255 MIRBuilder.buildConstant(IntPtrTy, ~(uint64_t)(StackAlign.value() - 1));
3256 auto AlignedAlloc = MIRBuilder.buildAnd(IntPtrTy, AllocAdd, AlignCst);
3257
3258 Align Alignment = AI.getAlign();
3259 if (Alignment <= StackAlign)
3260 Alignment = Align(1);
3261 MIRBuilder.buildDynStackAlloc(getOrCreateVReg(AI), AlignedAlloc, Alignment);
3262
3263 MF->getFrameInfo().CreateVariableSizedObject(Alignment, &AI);
3264 assert(MF->getFrameInfo().hasVarSizedObjects());
3265 return true;
3266}
3267
3268bool IRTranslator::translateVAArg(const User &U, MachineIRBuilder &MIRBuilder) {
3269 // FIXME: We may need more info about the type. Because of how LLT works,
3270 // we're completely discarding the i64/double distinction here (amongst
3271 // others). Fortunately the ABIs I know of where that matters don't use va_arg
3272 // anyway but that's not guaranteed.
3273 MIRBuilder.buildInstr(TargetOpcode::G_VAARG, {getOrCreateVReg(U)},
3274 {getOrCreateVReg(*U.getOperand(0)),
3275 DL->getABITypeAlign(U.getType()).value()});
3276 return true;
3277}
3278
3279bool IRTranslator::translateUnreachable(const User &U,
3280 MachineIRBuilder &MIRBuilder) {
3281 auto &UI = cast<UnreachableInst>(U);
3282 if (!UI.shouldLowerToTrap(MF->getTarget().Options.TrapUnreachable,
3283 MF->getTarget().Options.NoTrapAfterNoreturn))
3284 return true;
3285
3286 MIRBuilder.buildTrap();
3287 return true;
3288}
3289
3290bool IRTranslator::translateInsertElement(const User &U,
3291 MachineIRBuilder &MIRBuilder) {
3292 // If it is a <1 x Ty> vector, use the scalar as it is
3293 // not a legal vector type in LLT.
3294 if (auto *FVT = dyn_cast<FixedVectorType>(U.getType());
3295 FVT && FVT->getNumElements() == 1)
3296 return translateCopy(U, *U.getOperand(1), MIRBuilder);
3297
3298 Register Res = getOrCreateVReg(U);
3299 Register Val = getOrCreateVReg(*U.getOperand(0));
3300 Register Elt = getOrCreateVReg(*U.getOperand(1));
3301 unsigned PreferredVecIdxWidth = TLI->getVectorIdxWidth(*DL);
3302 Register Idx;
3303 if (auto *CI = dyn_cast<ConstantInt>(U.getOperand(2))) {
3304 if (CI->getBitWidth() != PreferredVecIdxWidth) {
3305 APInt NewIdx = CI->getValue().zextOrTrunc(PreferredVecIdxWidth);
3306 auto *NewIdxCI = ConstantInt::get(CI->getContext(), NewIdx);
3307 Idx = getOrCreateVReg(*NewIdxCI);
3308 }
3309 }
3310 if (!Idx)
3311 Idx = getOrCreateVReg(*U.getOperand(2));
3312 if (MRI->getType(Idx).getSizeInBits() != PreferredVecIdxWidth) {
3313 const LLT VecIdxTy =
3314 MRI->getType(Idx).changeElementSize(PreferredVecIdxWidth);
3315 Idx = MIRBuilder.buildZExtOrTrunc(VecIdxTy, Idx).getReg(0);
3316 }
3317 MIRBuilder.buildInsertVectorElement(Res, Val, Elt, Idx);
3318 return true;
3319}
3320
3321bool IRTranslator::translateInsertVector(const User &U,
3322 MachineIRBuilder &MIRBuilder) {
3323 Register Dst = getOrCreateVReg(U);
3324 Register Vec = getOrCreateVReg(*U.getOperand(0));
3325 Register Elt = getOrCreateVReg(*U.getOperand(1));
3326
3327 ConstantInt *CI = cast<ConstantInt>(U.getOperand(2));
3328 unsigned PreferredVecIdxWidth = TLI->getVectorIdxWidth(*DL);
3329
3330 // Resize Index to preferred index width.
3331 if (CI->getBitWidth() != PreferredVecIdxWidth) {
3332 APInt NewIdx = CI->getValue().zextOrTrunc(PreferredVecIdxWidth);
3333 CI = ConstantInt::get(CI->getContext(), NewIdx);
3334 }
3335
3336 // If it is a <1 x Ty> vector, we have to use other means.
3337 if (auto *ResultType = dyn_cast<FixedVectorType>(U.getOperand(1)->getType());
3338 ResultType && ResultType->getNumElements() == 1) {
3339 if (auto *InputType = dyn_cast<FixedVectorType>(U.getOperand(0)->getType());
3340 InputType && InputType->getNumElements() == 1) {
3341 // We are inserting an illegal fixed vector into an illegal
3342 // fixed vector, use the scalar as it is not a legal vector type
3343 // in LLT.
3344 return translateCopy(U, *U.getOperand(0), MIRBuilder);
3345 }
3346 if (isa<FixedVectorType>(U.getOperand(0)->getType())) {
3347 // We are inserting an illegal fixed vector into a legal fixed
3348 // vector, use the scalar as it is not a legal vector type in
3349 // LLT.
3350 Register Idx = getOrCreateVReg(*CI);
3351 MIRBuilder.buildInsertVectorElement(Dst, Vec, Elt, Idx);
3352 return true;
3353 }
3354 if (isa<ScalableVectorType>(U.getOperand(0)->getType())) {
3355 // We are inserting an illegal fixed vector into a scalable
3356 // vector, use a scalar element insert.
3357 LLT VecIdxTy = LLT::scalar(PreferredVecIdxWidth);
3358 Register Idx = getOrCreateVReg(*CI);
3359 auto ScaledIndex = MIRBuilder.buildMul(
3360 VecIdxTy, MIRBuilder.buildVScale(VecIdxTy, 1), Idx);
3361 MIRBuilder.buildInsertVectorElement(Dst, Vec, Elt, ScaledIndex);
3362 return true;
3363 }
3364 }
3365
3366 MIRBuilder.buildInsertSubvector(
3367 getOrCreateVReg(U), getOrCreateVReg(*U.getOperand(0)),
3368 getOrCreateVReg(*U.getOperand(1)), CI->getZExtValue());
3369 return true;
3370}
3371
3372bool IRTranslator::translateExtractElement(const User &U,
3373 MachineIRBuilder &MIRBuilder) {
3374 // If it is a <1 x Ty> vector, use the scalar as it is
3375 // not a legal vector type in LLT.
3376 if (const FixedVectorType *FVT =
3377 dyn_cast<FixedVectorType>(U.getOperand(0)->getType()))
3378 if (FVT->getNumElements() == 1)
3379 return translateCopy(U, *U.getOperand(0), MIRBuilder);
3380
3381 Register Res = getOrCreateVReg(U);
3382 Register Val = getOrCreateVReg(*U.getOperand(0));
3383 unsigned PreferredVecIdxWidth = TLI->getVectorIdxWidth(*DL);
3384 Register Idx;
3385 if (auto *CI = dyn_cast<ConstantInt>(U.getOperand(1))) {
3386 if (CI->getBitWidth() != PreferredVecIdxWidth) {
3387 APInt NewIdx = CI->getValue().zextOrTrunc(PreferredVecIdxWidth);
3388 auto *NewIdxCI = ConstantInt::get(CI->getContext(), NewIdx);
3389 Idx = getOrCreateVReg(*NewIdxCI);
3390 }
3391 }
3392 if (!Idx)
3393 Idx = getOrCreateVReg(*U.getOperand(1));
3394 if (MRI->getType(Idx).getSizeInBits() != PreferredVecIdxWidth) {
3395 const LLT VecIdxTy =
3396 MRI->getType(Idx).changeElementSize(PreferredVecIdxWidth);
3397 Idx = MIRBuilder.buildZExtOrTrunc(VecIdxTy, Idx).getReg(0);
3398 }
3399 MIRBuilder.buildExtractVectorElement(Res, Val, Idx);
3400 return true;
3401}
3402
3403bool IRTranslator::translateExtractVector(const User &U,
3404 MachineIRBuilder &MIRBuilder) {
3405 Register Res = getOrCreateVReg(U);
3406 Register Vec = getOrCreateVReg(*U.getOperand(0));
3407 ConstantInt *CI = cast<ConstantInt>(U.getOperand(1));
3408 unsigned PreferredVecIdxWidth = TLI->getVectorIdxWidth(*DL);
3409
3410 // Resize Index to preferred index width.
3411 if (CI->getBitWidth() != PreferredVecIdxWidth) {
3412 APInt NewIdx = CI->getValue().zextOrTrunc(PreferredVecIdxWidth);
3413 CI = ConstantInt::get(CI->getContext(), NewIdx);
3414 }
3415
3416 // If it is a <1 x Ty> vector, we have to use other means.
3417 if (auto *ResultType = dyn_cast<FixedVectorType>(U.getType());
3418 ResultType && ResultType->getNumElements() == 1) {
3419 if (auto *InputType = dyn_cast<FixedVectorType>(U.getOperand(0)->getType());
3420 InputType && InputType->getNumElements() == 1) {
3421 // We are extracting an illegal fixed vector from an illegal fixed vector,
3422 // use the scalar as it is not a legal vector type in LLT.
3423 return translateCopy(U, *U.getOperand(0), MIRBuilder);
3424 }
3425 if (isa<FixedVectorType>(U.getOperand(0)->getType())) {
3426 // We are extracting an illegal fixed vector from a legal fixed
3427 // vector, use the scalar as it is not a legal vector type in
3428 // LLT.
3429 Register Idx = getOrCreateVReg(*CI);
3430 MIRBuilder.buildExtractVectorElement(Res, Vec, Idx);
3431 return true;
3432 }
3433 if (isa<ScalableVectorType>(U.getOperand(0)->getType())) {
3434 // We are extracting an illegal fixed vector from a scalable
3435 // vector, use a scalar element extract.
3436 LLT VecIdxTy = LLT::scalar(PreferredVecIdxWidth);
3437 Register Idx = getOrCreateVReg(*CI);
3438 auto ScaledIndex = MIRBuilder.buildMul(
3439 VecIdxTy, MIRBuilder.buildVScale(VecIdxTy, 1), Idx);
3440 MIRBuilder.buildExtractVectorElement(Res, Vec, ScaledIndex);
3441 return true;
3442 }
3443 }
3444
3445 MIRBuilder.buildExtractSubvector(getOrCreateVReg(U),
3446 getOrCreateVReg(*U.getOperand(0)),
3447 CI->getZExtValue());
3448 return true;
3449}
3450
3451bool IRTranslator::translateShuffleVector(const User &U,
3452 MachineIRBuilder &MIRBuilder) {
3453 // A ShuffleVector that operates on scalable vectors is a splat vector where
3454 // the value of the splat vector is the 0th element of the first operand,
3455 // since the index mask operand is the zeroinitializer (undef and
3456 // poison are treated as zeroinitializer here).
3457 if (U.getOperand(0)->getType()->isScalableTy()) {
3458 Register Val = getOrCreateVReg(*U.getOperand(0));
3459 auto SplatVal = MIRBuilder.buildExtractVectorElementConstant(
3460 MRI->getType(Val).getElementType(), Val, 0);
3461 MIRBuilder.buildSplatVector(getOrCreateVReg(U), SplatVal);
3462 return true;
3463 }
3464
3465 ArrayRef<int> Mask;
3466 if (auto *SVI = dyn_cast<ShuffleVectorInst>(&U))
3467 Mask = SVI->getShuffleMask();
3468 else
3469 Mask = cast<ConstantExpr>(U).getShuffleMask();
3470
3471 // As GISel does not represent <1 x > vectors as a separate type from scalars,
3472 // we transform shuffle_vector with a scalar output to an
3473 // ExtractVectorElement. If the input type is also scalar it becomes a Copy.
3474 unsigned DstElts = cast<FixedVectorType>(U.getType())->getNumElements();
3475 unsigned SrcElts =
3476 cast<FixedVectorType>(U.getOperand(0)->getType())->getNumElements();
3477 if (DstElts == 1) {
3478 unsigned M = Mask[0];
3479 if (SrcElts == 1) {
3480 if (M == 0 || M == 1)
3481 return translateCopy(U, *U.getOperand(M), MIRBuilder);
3482 MIRBuilder.buildUndef(getOrCreateVReg(U));
3483 } else {
3484 Register Dst = getOrCreateVReg(U);
3485 if (M < SrcElts) {
3487 Dst, getOrCreateVReg(*U.getOperand(0)), M);
3488 } else if (M < SrcElts * 2) {
3490 Dst, getOrCreateVReg(*U.getOperand(1)), M - SrcElts);
3491 } else {
3492 MIRBuilder.buildUndef(Dst);
3493 }
3494 }
3495 return true;
3496 }
3497
3498 // A single element src is transformed to a build_vector.
3499 if (SrcElts == 1) {
3502 for (int M : Mask) {
3503 LLT SrcTy = getLLTForType(*U.getOperand(0)->getType(), *DL);
3504 if (M == 0 || M == 1) {
3505 Ops.push_back(getOrCreateVReg(*U.getOperand(M)));
3506 } else {
3507 if (!Undef.isValid()) {
3508 Undef = MRI->createGenericVirtualRegister(SrcTy);
3509 MIRBuilder.buildUndef(Undef);
3510 }
3511 Ops.push_back(Undef);
3512 }
3513 }
3514 MIRBuilder.buildBuildVector(getOrCreateVReg(U), Ops);
3515 return true;
3516 }
3517
3518 ArrayRef<int> MaskAlloc = MF->allocateShuffleMask(Mask);
3519 MIRBuilder
3520 .buildInstr(TargetOpcode::G_SHUFFLE_VECTOR, {getOrCreateVReg(U)},
3521 {getOrCreateVReg(*U.getOperand(0)),
3522 getOrCreateVReg(*U.getOperand(1))})
3523 .addShuffleMask(MaskAlloc);
3524 return true;
3525}
3526
3527bool IRTranslator::translatePHI(const User &U, MachineIRBuilder &MIRBuilder) {
3528 const PHINode &PI = cast<PHINode>(U);
3529
3530 SmallVector<MachineInstr *, 4> Insts;
3531 for (auto Reg : getOrCreateVRegs(PI)) {
3532 auto MIB = MIRBuilder.buildInstr(TargetOpcode::G_PHI, {Reg}, {});
3533 Insts.push_back(MIB.getInstr());
3534 }
3535
3536 PendingPHIs.emplace_back(&PI, std::move(Insts));
3537 return true;
3538}
3539
3540bool IRTranslator::translateAtomicCmpXchg(const User &U,
3541 MachineIRBuilder &MIRBuilder) {
3542 const AtomicCmpXchgInst &I = cast<AtomicCmpXchgInst>(U);
3543
3544 auto Flags = TLI->getAtomicMemOperandFlags(I, *DL);
3545
3546 auto Res = getOrCreateVRegs(I);
3547 Register OldValRes = Res[0];
3548 Register SuccessRes = Res[1];
3549 Register Addr = getOrCreateVReg(*I.getPointerOperand());
3550 Register Cmp = getOrCreateVReg(*I.getCompareOperand());
3551 Register NewVal = getOrCreateVReg(*I.getNewValOperand());
3552
3554 OldValRes, SuccessRes, Addr, Cmp, NewVal,
3555 *MF->getMachineMemOperand(
3556 MachinePointerInfo(I.getPointerOperand()), Flags, MRI->getType(Cmp),
3557 getMemOpAlign(I), I.getAAMetadata(), nullptr, I.getSyncScopeID(),
3558 I.getSuccessOrdering(), I.getFailureOrdering()));
3559 return true;
3560}
3561
3562bool IRTranslator::translateAtomicRMW(const User &U,
3563 MachineIRBuilder &MIRBuilder) {
3564 if (!mayTranslateUserTypes(U))
3565 return false;
3566
3567 const AtomicRMWInst &I = cast<AtomicRMWInst>(U);
3568 auto Flags = TLI->getAtomicMemOperandFlags(I, *DL);
3569
3570 Register Res = getOrCreateVReg(I);
3571 Register Addr = getOrCreateVReg(*I.getPointerOperand());
3572 Register Val = getOrCreateVReg(*I.getValOperand());
3573
3574 unsigned Opcode = 0;
3575 switch (I.getOperation()) {
3576 default:
3577 return false;
3579 Opcode = TargetOpcode::G_ATOMICRMW_XCHG;
3580 break;
3581 case AtomicRMWInst::Add:
3582 Opcode = TargetOpcode::G_ATOMICRMW_ADD;
3583 break;
3584 case AtomicRMWInst::Sub:
3585 Opcode = TargetOpcode::G_ATOMICRMW_SUB;
3586 break;
3587 case AtomicRMWInst::And:
3588 Opcode = TargetOpcode::G_ATOMICRMW_AND;
3589 break;
3591 Opcode = TargetOpcode::G_ATOMICRMW_NAND;
3592 break;
3593 case AtomicRMWInst::Or:
3594 Opcode = TargetOpcode::G_ATOMICRMW_OR;
3595 break;
3596 case AtomicRMWInst::Xor:
3597 Opcode = TargetOpcode::G_ATOMICRMW_XOR;
3598 break;
3599 case AtomicRMWInst::Max:
3600 Opcode = TargetOpcode::G_ATOMICRMW_MAX;
3601 break;
3602 case AtomicRMWInst::Min:
3603 Opcode = TargetOpcode::G_ATOMICRMW_MIN;
3604 break;
3606 Opcode = TargetOpcode::G_ATOMICRMW_UMAX;
3607 break;
3609 Opcode = TargetOpcode::G_ATOMICRMW_UMIN;
3610 break;
3612 Opcode = TargetOpcode::G_ATOMICRMW_FADD;
3613 break;
3615 Opcode = TargetOpcode::G_ATOMICRMW_FSUB;
3616 break;
3618 Opcode = TargetOpcode::G_ATOMICRMW_FMAX;
3619 break;
3621 Opcode = TargetOpcode::G_ATOMICRMW_FMIN;
3622 break;
3624 Opcode = TargetOpcode::G_ATOMICRMW_FMAXIMUM;
3625 break;
3627 Opcode = TargetOpcode::G_ATOMICRMW_FMINIMUM;
3628 break;
3630 Opcode = TargetOpcode::G_ATOMICRMW_FMAXIMUMNUM;
3631 break;
3633 Opcode = TargetOpcode::G_ATOMICRMW_FMINIMUMNUM;
3634 break;
3636 Opcode = TargetOpcode::G_ATOMICRMW_UINC_WRAP;
3637 break;
3639 Opcode = TargetOpcode::G_ATOMICRMW_UDEC_WRAP;
3640 break;
3642 Opcode = TargetOpcode::G_ATOMICRMW_USUB_COND;
3643 break;
3645 Opcode = TargetOpcode::G_ATOMICRMW_USUB_SAT;
3646 break;
3647 }
3648
3649 MIRBuilder.buildAtomicRMW(
3650 Opcode, Res, Addr, Val,
3651 *MF->getMachineMemOperand(MachinePointerInfo(I.getPointerOperand()),
3652 Flags, MRI->getType(Val), getMemOpAlign(I),
3653 I.getAAMetadata(), nullptr, I.getSyncScopeID(),
3654 I.getOrdering()));
3655 return true;
3656}
3657
3658bool IRTranslator::translateFence(const User &U,
3659 MachineIRBuilder &MIRBuilder) {
3660 const FenceInst &Fence = cast<FenceInst>(U);
3661 MIRBuilder.buildFence(static_cast<unsigned>(Fence.getOrdering()),
3662 Fence.getSyncScopeID());
3663 return true;
3664}
3665
3666bool IRTranslator::translateFreeze(const User &U,
3667 MachineIRBuilder &MIRBuilder) {
3668 const ArrayRef<Register> DstRegs = getOrCreateVRegs(U);
3669 const ArrayRef<Register> SrcRegs = getOrCreateVRegs(*U.getOperand(0));
3670
3671 assert(DstRegs.size() == SrcRegs.size() &&
3672 "Freeze with different source and destination type?");
3673
3674 for (unsigned I = 0; I < DstRegs.size(); ++I) {
3675 MIRBuilder.buildFreeze(DstRegs[I], SrcRegs[I]);
3676 }
3677
3678 return true;
3679}
3680
3681void IRTranslator::finishPendingPhis() {
3682#ifndef NDEBUG
3683 DILocationVerifier Verifier;
3684 GISelObserverWrapper WrapperObserver(&Verifier);
3685 RAIIMFObsDelInstaller ObsInstall(*MF, WrapperObserver);
3686#endif // ifndef NDEBUG
3687 for (auto &Phi : PendingPHIs) {
3688 const PHINode *PI = Phi.first;
3689 if (PI->getType()->isEmptyTy())
3690 continue;
3691 ArrayRef<MachineInstr *> ComponentPHIs = Phi.second;
3692 MachineBasicBlock *PhiMBB = ComponentPHIs[0]->getParent();
3693 EntryBuilder->setDebugLoc(PI->getDebugLoc());
3694#ifndef NDEBUG
3695 Verifier.setCurrentInst(PI);
3696#endif // ifndef NDEBUG
3697
3698 SmallPtrSet<const MachineBasicBlock *, 16> SeenPreds;
3699 for (unsigned i = 0; i < PI->getNumIncomingValues(); ++i) {
3700 auto IRPred = PI->getIncomingBlock(i);
3701 ArrayRef<Register> ValRegs = getOrCreateVRegs(*PI->getIncomingValue(i));
3702 for (auto *Pred : getMachinePredBBs({IRPred, PI->getParent()})) {
3703 if (SeenPreds.count(Pred) || !PhiMBB->isPredecessor(Pred))
3704 continue;
3705 SeenPreds.insert(Pred);
3706 for (unsigned j = 0; j < ValRegs.size(); ++j) {
3707 MachineInstrBuilder MIB(*MF, ComponentPHIs[j]);
3708 MIB.addUse(ValRegs[j]);
3709 MIB.addMBB(Pred);
3710 }
3711 }
3712 }
3713 }
3714}
3715
3716void IRTranslator::translateDbgValueRecord(Value *V, bool HasArgList,
3717 const DILocalVariable *Variable,
3718 const DIExpression *Expression,
3719 const DebugLoc &DL,
3720 MachineIRBuilder &MIRBuilder) {
3721 assert(Variable->isValidLocationForIntrinsic(DL) &&
3722 "Expected inlined-at fields to agree");
3723 // Act as if we're handling a debug intrinsic.
3724 MIRBuilder.setDebugLoc(DL);
3725
3726 if (!V || HasArgList) {
3727 // DI cannot produce a valid DBG_VALUE, so produce an undef DBG_VALUE to
3728 // terminate any prior location.
3729 MIRBuilder.buildIndirectDbgValue(0, Variable, Expression);
3730 return;
3731 }
3732
3733 if (const auto *CI = dyn_cast<Constant>(V)) {
3734 MIRBuilder.buildConstDbgValue(*CI, Variable, Expression);
3735 return;
3736 }
3737
3738 if (auto *AI = dyn_cast<AllocaInst>(V);
3739 AI && AI->isStaticAlloca() && Expression->startsWithDeref()) {
3740 // If the value is an alloca and the expression starts with a
3741 // dereference, track a stack slot instead of a register, as registers
3742 // may be clobbered.
3743 auto ExprOperands = Expression->getElements();
3744 auto *ExprDerefRemoved =
3745 DIExpression::get(AI->getContext(), ExprOperands.drop_front());
3746 MIRBuilder.buildFIDbgValue(getOrCreateFrameIndex(*AI), Variable,
3747 ExprDerefRemoved);
3748 return;
3749 }
3750 if (translateIfEntryValueArgument(false, V, Variable, Expression, DL,
3751 MIRBuilder))
3752 return;
3753 for (Register Reg : getOrCreateVRegs(*V)) {
3754 // FIXME: This does not handle register-indirect values at offset 0. The
3755 // direct/indirect thing shouldn't really be handled by something as
3756 // implicit as reg+noreg vs reg+imm in the first place, but it seems
3757 // pretty baked in right now.
3758 MIRBuilder.buildDirectDbgValue(Reg, Variable, Expression);
3759 }
3760}
3761
3762void IRTranslator::translateDbgDeclareRecord(Value *Address, bool HasArgList,
3763 const DILocalVariable *Variable,
3764 const DIExpression *Expression,
3765 const DebugLoc &DL,
3766 MachineIRBuilder &MIRBuilder) {
3767 if (!Address || isa<UndefValue>(Address)) {
3768 LLVM_DEBUG(dbgs() << "Dropping debug info for " << *Variable << "\n");
3769 return;
3770 }
3771
3772 assert(Variable->isValidLocationForIntrinsic(DL) &&
3773 "Expected inlined-at fields to agree");
3774 auto AI = dyn_cast<AllocaInst>(Address);
3775 if (AI && AI->isStaticAlloca()) {
3776 // Static allocas are tracked at the MF level, no need for DBG_VALUE
3777 // instructions (in fact, they get ignored if they *do* exist).
3778 MF->setVariableDbgInfo(Variable, Expression,
3779 getOrCreateFrameIndex(*AI), DL);
3780 return;
3781 }
3782
3783 if (translateIfEntryValueArgument(true, Address, Variable,
3784 Expression, DL,
3785 MIRBuilder))
3786 return;
3787
3788 // A dbg.declare describes the address of a source variable, so lower it
3789 // into an indirect DBG_VALUE.
3790 MIRBuilder.setDebugLoc(DL);
3791 MIRBuilder.buildIndirectDbgValue(getOrCreateVReg(*Address), Variable,
3792 Expression);
3793}
3794
3795void IRTranslator::translateDbgInfo(const Instruction &Inst,
3796 MachineIRBuilder &MIRBuilder) {
3797 for (DbgRecord &DR : Inst.getDbgRecordRange()) {
3798 if (DbgLabelRecord *DLR = dyn_cast<DbgLabelRecord>(&DR)) {
3799 MIRBuilder.setDebugLoc(DLR->getDebugLoc());
3800 assert(DLR->getLabel() && "Missing label");
3801 assert(DLR->getLabel()->isValidLocationForIntrinsic(
3802 MIRBuilder.getDebugLoc()) &&
3803 "Expected inlined-at fields to agree");
3804 MIRBuilder.buildDbgLabel(DLR->getLabel());
3805 continue;
3806 }
3807 DbgVariableRecord &DVR = cast<DbgVariableRecord>(DR);
3808 const DILocalVariable *Variable = DVR.getVariable();
3809 const DIExpression *Expression = DVR.getExpression();
3810 Value *V = DVR.getVariableLocationOp(0);
3811 if (DVR.isDbgDeclare())
3812 translateDbgDeclareRecord(V, DVR.hasArgList(), Variable, Expression,
3813 DVR.getDebugLoc(), MIRBuilder);
3814 else
3815 translateDbgValueRecord(V, DVR.hasArgList(), Variable, Expression,
3816 DVR.getDebugLoc(), MIRBuilder);
3817 }
3818}
3819
3820bool IRTranslator::translate(const Instruction &Inst) {
3821 CurBuilder->setDebugLoc(Inst.getDebugLoc());
3822 CurBuilder->setPCSections(Inst.getMetadata(LLVMContext::MD_pcsections));
3823 CurBuilder->setMMRAMetadata(Inst.getMetadata(LLVMContext::MD_mmra));
3824
3825 if (TLI->fallBackToDAGISel(Inst))
3826 return false;
3827
3828 switch (Inst.getOpcode()) {
3829#define HANDLE_INST(NUM, OPCODE, CLASS) \
3830 case Instruction::OPCODE: \
3831 return translate##OPCODE(Inst, *CurBuilder.get());
3832#include "llvm/IR/Instruction.def"
3833 default:
3834 return false;
3835 }
3836}
3837
3838bool IRTranslator::translate(const Constant &C, Register Reg) {
3839 // We only emit constants into the entry block from here. To prevent jumpy
3840 // debug behaviour remove debug line.
3841 if (auto CurrInstDL = CurBuilder->getDL())
3842 EntryBuilder->setDebugLoc(DebugLoc());
3843
3844 if (auto CI = dyn_cast<ConstantInt>(&C)) {
3845 // buildConstant expects a to-be-splatted scalar ConstantInt.
3846 if (isa<VectorType>(CI->getType()))
3847 CI = ConstantInt::get(CI->getContext(), CI->getValue());
3848 EntryBuilder->buildConstant(Reg, *CI);
3849 } else if (auto CF = dyn_cast<ConstantFP>(&C)) {
3850 // buildFConstant expects a to-be-splatted scalar ConstantFP.
3851 if (isa<VectorType>(CF->getType()))
3852 CF = ConstantFP::get(CF->getContext(), CF->getValue());
3853 EntryBuilder->buildFConstant(Reg, *CF);
3854 } else if (isa<UndefValue>(C))
3855 EntryBuilder->buildUndef(Reg);
3856 else if (isa<ConstantPointerNull>(C))
3857 EntryBuilder->buildConstant(Reg, 0);
3858 else if (auto GV = dyn_cast<GlobalValue>(&C))
3859 EntryBuilder->buildGlobalValue(Reg, GV);
3860 else if (auto CPA = dyn_cast<ConstantPtrAuth>(&C)) {
3861 Register Addr = getOrCreateVReg(*CPA->getPointer());
3862 Register AddrDisc = getOrCreateVReg(*CPA->getAddrDiscriminator());
3863 EntryBuilder->buildConstantPtrAuth(Reg, CPA, Addr, AddrDisc);
3864 } else if (auto CAZ = dyn_cast<ConstantAggregateZero>(&C)) {
3865 Constant &Elt = *CAZ->getElementValue(0u);
3866 if (isa<ScalableVectorType>(CAZ->getType())) {
3867 EntryBuilder->buildSplatVector(Reg, getOrCreateVReg(Elt));
3868 return true;
3869 }
3870 // Return the scalar if it is a <1 x Ty> vector.
3871 unsigned NumElts = CAZ->getElementCount().getFixedValue();
3872 if (NumElts == 1)
3873 return translateCopy(C, Elt, *EntryBuilder);
3874 // All elements are zero so we can just use the first one.
3875 EntryBuilder->buildSplatBuildVector(Reg, getOrCreateVReg(Elt));
3876 } else if (auto CV = dyn_cast<ConstantDataVector>(&C)) {
3877 // Return the scalar if it is a <1 x Ty> vector.
3878 if (CV->getNumElements() == 1)
3879 return translateCopy(C, *CV->getElementAsConstant(0), *EntryBuilder);
3881 for (unsigned i = 0; i < CV->getNumElements(); ++i) {
3882 Constant &Elt = *CV->getElementAsConstant(i);
3883 Ops.push_back(getOrCreateVReg(Elt));
3884 }
3885 EntryBuilder->buildBuildVector(Reg, Ops);
3886 } else if (auto CE = dyn_cast<ConstantExpr>(&C)) {
3887 switch(CE->getOpcode()) {
3888#define HANDLE_INST(NUM, OPCODE, CLASS) \
3889 case Instruction::OPCODE: \
3890 return translate##OPCODE(*CE, *EntryBuilder.get());
3891#include "llvm/IR/Instruction.def"
3892 default:
3893 return false;
3894 }
3895 } else if (auto CV = dyn_cast<ConstantVector>(&C)) {
3896 if (CV->getNumOperands() == 1)
3897 return translateCopy(C, *CV->getOperand(0), *EntryBuilder);
3899 for (unsigned i = 0; i < CV->getNumOperands(); ++i) {
3900 Ops.push_back(getOrCreateVReg(*CV->getOperand(i)));
3901 }
3902 EntryBuilder->buildBuildVector(Reg, Ops);
3903 } else if (auto *BA = dyn_cast<BlockAddress>(&C)) {
3904 EntryBuilder->buildBlockAddress(Reg, BA);
3905 } else
3906 return false;
3907
3908 return true;
3909}
3910
3911bool IRTranslator::mayTranslateUserTypes(const User &U) const {
3912 const TargetMachine &TM = TLI->getTargetMachine();
3913 if (LLT::getUseExtended())
3914 return true;
3915
3916 // BF16 cannot currently be represented by default LLT. To avoid miscompiles
3917 // we prevent any instructions using them by default in all targets that do
3918 // not explicitly enable it via LLT::setUseExtended(true).
3919 // SPIRV target is exception.
3920 return TM.getTargetTriple().isSPIRV() ||
3921 (!U.getType()->getScalarType()->isBFloatTy() &&
3922 !any_of(U.operands(), [](Value *V) {
3923 return V->getType()->getScalarType()->isBFloatTy();
3924 }));
3925}
3926
3927bool IRTranslator::finalizeBasicBlock(const BasicBlock &BB,
3929 for (auto &BTB : SL->BitTestCases) {
3930 // Emit header first, if it wasn't already emitted.
3931 if (!BTB.Emitted)
3932 emitBitTestHeader(BTB, BTB.Parent);
3933
3934 BranchProbability UnhandledProb = BTB.Prob;
3935 for (unsigned j = 0, ej = BTB.Cases.size(); j != ej; ++j) {
3936 UnhandledProb -= BTB.Cases[j].ExtraProb;
3937 // Set the current basic block to the mbb we wish to insert the code into
3938 MachineBasicBlock *MBB = BTB.Cases[j].ThisBB;
3939 // If all cases cover a contiguous range, it is not necessary to jump to
3940 // the default block after the last bit test fails. This is because the
3941 // range check during bit test header creation has guaranteed that every
3942 // case here doesn't go outside the range. In this case, there is no need
3943 // to perform the last bit test, as it will always be true. Instead, make
3944 // the second-to-last bit-test fall through to the target of the last bit
3945 // test, and delete the last bit test.
3946
3947 MachineBasicBlock *NextMBB;
3948 if ((BTB.ContiguousRange || BTB.FallthroughUnreachable) && j + 2 == ej) {
3949 // Second-to-last bit-test with contiguous range: fall through to the
3950 // target of the final bit test.
3951 NextMBB = BTB.Cases[j + 1].TargetBB;
3952 } else if (j + 1 == ej) {
3953 // For the last bit test, fall through to Default.
3954 NextMBB = BTB.Default;
3955 } else {
3956 // Otherwise, fall through to the next bit test.
3957 NextMBB = BTB.Cases[j + 1].ThisBB;
3958 }
3959
3960 emitBitTestCase(BTB, NextMBB, UnhandledProb, BTB.Reg, BTB.Cases[j], MBB);
3961
3962 if ((BTB.ContiguousRange || BTB.FallthroughUnreachable) && j + 2 == ej) {
3963 // We need to record the replacement phi edge here that normally
3964 // happens in emitBitTestCase before we delete the case, otherwise the
3965 // phi edge will be lost.
3966 addMachineCFGPred({BTB.Parent->getBasicBlock(),
3967 BTB.Cases[ej - 1].TargetBB->getBasicBlock()},
3968 MBB);
3969 // Since we're not going to use the final bit test, remove it.
3970 BTB.Cases.pop_back();
3971 break;
3972 }
3973 }
3974 // This is "default" BB. We have two jumps to it. From "header" BB and from
3975 // last "case" BB, unless the latter was skipped.
3976 CFGEdge HeaderToDefaultEdge = {BTB.Parent->getBasicBlock(),
3977 BTB.Default->getBasicBlock()};
3978 addMachineCFGPred(HeaderToDefaultEdge, BTB.Parent);
3979 if (!BTB.ContiguousRange) {
3980 addMachineCFGPred(HeaderToDefaultEdge, BTB.Cases.back().ThisBB);
3981 }
3982 }
3983 SL->BitTestCases.clear();
3984
3985 for (auto &JTCase : SL->JTCases) {
3986 // Emit header first, if it wasn't already emitted.
3987 if (!JTCase.first.Emitted)
3988 emitJumpTableHeader(JTCase.second, JTCase.first, JTCase.first.HeaderBB);
3989
3990 emitJumpTable(JTCase.second, JTCase.second.MBB);
3991 }
3992 SL->JTCases.clear();
3993
3994 for (auto &SwCase : SL->SwitchCases)
3995 emitSwitchCase(SwCase, &CurBuilder->getMBB(), *CurBuilder);
3996 SL->SwitchCases.clear();
3997
3998 // Check if we need to generate stack-protector guard checks.
3999 StackProtector &SP = getAnalysis<StackProtector>();
4000 if (SP.shouldEmitSDCheck(BB)) {
4001 bool FunctionBasedInstrumentation =
4002 TLI->getSSPStackGuardCheck(*MF->getFunction().getParent(), *Libcalls);
4003 SPDescriptor.initialize(&BB, &MBB, FunctionBasedInstrumentation);
4004 }
4005 // Handle stack protector.
4006 if (SPDescriptor.shouldEmitFunctionBasedCheckStackProtector()) {
4007 LLVM_DEBUG(dbgs() << "Unimplemented stack protector case\n");
4008 return false;
4009 } else if (SPDescriptor.shouldEmitStackProtector()) {
4010 MachineBasicBlock *ParentMBB = SPDescriptor.getParentMBB();
4011 MachineBasicBlock *SuccessMBB = SPDescriptor.getSuccessMBB();
4012
4013 // Find the split point to split the parent mbb. At the same time copy all
4014 // physical registers used in the tail of parent mbb into virtual registers
4015 // before the split point and back into physical registers after the split
4016 // point. This prevents us needing to deal with Live-ins and many other
4017 // register allocation issues caused by us splitting the parent mbb. The
4018 // register allocator will clean up said virtual copies later on.
4020 ParentMBB, *MF->getSubtarget().getInstrInfo());
4021
4022 // Splice the terminator of ParentMBB into SuccessMBB.
4023 SuccessMBB->splice(SuccessMBB->end(), ParentMBB, SplitPoint,
4024 ParentMBB->end());
4025
4026 // Add compare/jump on neq/jump to the parent BB.
4027 if (!emitSPDescriptorParent(SPDescriptor, ParentMBB))
4028 return false;
4029
4030 // CodeGen Failure MBB if we have not codegened it yet.
4031 MachineBasicBlock *FailureMBB = SPDescriptor.getFailureMBB();
4032 if (FailureMBB->empty()) {
4033 if (!emitSPDescriptorFailure(SPDescriptor, FailureMBB))
4034 return false;
4035 }
4036
4037 // Clear the Per-BB State.
4038 SPDescriptor.resetPerBBState();
4039 }
4040 return true;
4041}
4042
4043bool IRTranslator::emitSPDescriptorParent(StackProtectorDescriptor &SPD,
4044 MachineBasicBlock *ParentBB) {
4045 CurBuilder->setInsertPt(*ParentBB, ParentBB->end());
4046 // First create the loads to the guard/stack slot for the comparison.
4047 Type *PtrIRTy = PointerType::getUnqual(MF->getFunction().getContext());
4048 const LLT PtrTy = getLLTForType(*PtrIRTy, *DL);
4049 LLT PtrMemTy = getLLTForMVT(TLI->getPointerMemTy(*DL));
4050
4051 MachineFrameInfo &MFI = ParentBB->getParent()->getFrameInfo();
4052 int FI = MFI.getStackProtectorIndex();
4053
4054 Register Guard;
4055 Register StackSlotPtr = CurBuilder->buildFrameIndex(PtrTy, FI).getReg(0);
4056 const Module &M = *ParentBB->getParent()->getFunction().getParent();
4057 Align Align = DL->getPrefTypeAlign(PointerType::getUnqual(M.getContext()));
4058
4059 // Generate code to load the content of the guard slot.
4060 Register GuardVal =
4061 CurBuilder
4062 ->buildLoad(PtrMemTy, StackSlotPtr,
4063 MachinePointerInfo::getFixedStack(*MF, FI), Align,
4065 .getReg(0);
4066
4067 if (TLI->useStackGuardXorFP()) {
4068 LLVM_DEBUG(dbgs() << "Stack protector xor'ing with FP not yet implemented");
4069 return false;
4070 }
4071
4072 // Retrieve guard check function, nullptr if instrumentation is inlined.
4073 if (const Function *GuardCheckFn = TLI->getSSPStackGuardCheck(M, *Libcalls)) {
4074 // This path is currently untestable on GlobalISel, since the only platform
4075 // that needs this seems to be Windows, and we fall back on that currently.
4076 // The code still lives here in case that changes.
4077 // Silence warning about unused variable until the code below that uses
4078 // 'GuardCheckFn' is enabled.
4079 (void)GuardCheckFn;
4080 return false;
4081#if 0
4082 // The target provides a guard check function to validate the guard value.
4083 // Generate a call to that function with the content of the guard slot as
4084 // argument.
4085 FunctionType *FnTy = GuardCheckFn->getFunctionType();
4086 assert(FnTy->getNumParams() == 1 && "Invalid function signature");
4087 ISD::ArgFlagsTy Flags;
4088 if (GuardCheckFn->hasAttribute(1, Attribute::AttrKind::InReg))
4089 Flags.setInReg();
4090 CallLowering::ArgInfo GuardArgInfo(
4091 {GuardVal, FnTy->getParamType(0), {Flags}});
4092
4093 CallLowering::CallLoweringInfo Info;
4094 Info.OrigArgs.push_back(GuardArgInfo);
4095 Info.CallConv = GuardCheckFn->getCallingConv();
4096 Info.Callee = MachineOperand::CreateGA(GuardCheckFn, 0);
4097 Info.OrigRet = {Register(), FnTy->getReturnType()};
4098 if (!CLI->lowerCall(MIRBuilder, Info)) {
4099 LLVM_DEBUG(dbgs() << "Failed to lower call to stack protector check\n");
4100 return false;
4101 }
4102 return true;
4103#endif
4104 }
4105
4106 // If useLoadStackGuardNode returns true, generate LOAD_STACK_GUARD.
4107 // Otherwise, emit a volatile load to retrieve the stack guard value.
4108 if (TLI->useLoadStackGuardNode(*ParentBB->getBasicBlock()->getModule())) {
4109 Guard =
4110 MRI->createGenericVirtualRegister(LLT::scalar(PtrTy.getSizeInBits()));
4111 getStackGuard(Guard, *CurBuilder);
4112 } else {
4113 // TODO: test using android subtarget when we support @llvm.thread.pointer.
4114 const Value *IRGuard = TLI->getSDagStackGuard(M, *Libcalls);
4115 Register GuardPtr = getOrCreateVReg(*IRGuard);
4116
4117 Guard = CurBuilder
4118 ->buildLoad(PtrMemTy, GuardPtr,
4119 MachinePointerInfo::getFixedStack(*MF, FI), Align,
4122 .getReg(0);
4123 }
4124
4125 // Perform the comparison.
4126 auto Cmp =
4127 CurBuilder->buildICmp(CmpInst::ICMP_NE, LLT::integer(1), Guard, GuardVal);
4128 // If the guard/stackslot do not equal, branch to failure MBB.
4129 CurBuilder->buildBrCond(Cmp, *SPD.getFailureMBB());
4130 // Otherwise branch to success MBB.
4131 CurBuilder->buildBr(*SPD.getSuccessMBB());
4132 return true;
4133}
4134
4135bool IRTranslator::emitSPDescriptorFailure(StackProtectorDescriptor &SPD,
4136 MachineBasicBlock *FailureBB) {
4137 const RTLIB::LibcallImpl LibcallImpl =
4138 Libcalls->getLibcallImpl(RTLIB::STACKPROTECTOR_CHECK_FAIL);
4139 if (LibcallImpl == RTLIB::Unsupported)
4140 return false;
4141
4142 CurBuilder->setInsertPt(*FailureBB, FailureBB->end());
4143
4144 CallLowering::CallLoweringInfo Info;
4145 Info.CallConv = Libcalls->getLibcallImplCallingConv(LibcallImpl);
4146
4147 StringRef LibcallName =
4149 Info.Callee = MachineOperand::CreateES(LibcallName.data());
4150 Info.OrigRet = {Register(), Type::getVoidTy(MF->getFunction().getContext()),
4151 0};
4152 if (!CLI->lowerCall(*CurBuilder, Info)) {
4153 LLVM_DEBUG(dbgs() << "Failed to lower call to stack protector fail\n");
4154 return false;
4155 }
4156
4157 // Emit a trap instruction if we are required to do so.
4158 const TargetOptions &TargetOpts = TLI->getTargetMachine().Options;
4159 if (TargetOpts.TrapUnreachable && !TargetOpts.NoTrapAfterNoreturn)
4160 CurBuilder->buildInstr(TargetOpcode::G_TRAP);
4161
4162 return true;
4163}
4164
4165void IRTranslator::finalizeFunction() {
4166 // Release the memory used by the different maps we
4167 // needed during the translation.
4168 PendingPHIs.clear();
4169 VMap.reset();
4170 FrameIndices.clear();
4171 MachinePreds.clear();
4172 // MachineIRBuilder::DebugLoc can outlive the DILocation it holds. Clear it
4173 // to avoid accessing free’d memory (in runOnMachineFunction) and to avoid
4174 // destroying it twice (in ~IRTranslator() and ~LLVMContext())
4175 EntryBuilder.reset();
4176 CurBuilder.reset();
4177 FuncInfo.clear();
4178 SPDescriptor.resetPerFunctionState();
4179}
4180
4181/// Returns true if a BasicBlock \p BB within a variadic function contains a
4182/// variadic musttail call.
4183static bool checkForMustTailInVarArgFn(bool IsVarArg, const BasicBlock &BB) {
4184 if (!IsVarArg)
4185 return false;
4186
4187 // Walk the block backwards, because tail calls usually only appear at the end
4188 // of a block.
4189 return llvm::any_of(llvm::reverse(BB), [](const Instruction &I) {
4190 const auto *CI = dyn_cast<CallInst>(&I);
4191 return CI && CI->isMustTailCall();
4192 });
4193}
4194
4196 MF = &CurMF;
4197 const Function &F = MF->getFunction();
4198 ORE = std::make_unique<OptimizationRemarkEmitter>(&F);
4199 CLI = MF->getSubtarget().getCallLowering();
4200
4201 if (CLI->fallBackToDAGISel(*MF)) {
4202 OptimizationRemarkMissed R("gisel-irtranslator", "GISelFailure",
4203 F.getSubprogram(), &F.getEntryBlock());
4204 R << "unable to lower function: "
4205 << ore::NV("Prototype", F.getFunctionType());
4206
4207 reportTranslationError(*MF, *ORE, R);
4208 return false;
4209 }
4210
4213 // Set the CSEConfig and run the analysis.
4214 GISelCSEInfo *CSEInfo = nullptr;
4216
4217 bool EnableCSE = EnableCSEInIRTranslator.getNumOccurrences()
4219 : TPC->isGISelCSEEnabled();
4220
4221 const TargetSubtargetInfo &Subtarget = MF->getSubtarget();
4222 TLI = Subtarget.getTargetLowering();
4223
4224 if (EnableCSE) {
4225 EntryBuilder = std::make_unique<CSEMIRBuilder>(CurMF);
4226 CSEInfo = &Wrapper.get(TPC->getCSEConfig());
4227 EntryBuilder->setCSEInfo(CSEInfo);
4228 CurBuilder = std::make_unique<CSEMIRBuilder>(CurMF);
4229 CurBuilder->setCSEInfo(CSEInfo);
4230 } else {
4231 EntryBuilder = std::make_unique<MachineIRBuilder>();
4232 CurBuilder = std::make_unique<MachineIRBuilder>();
4233 }
4234 CLI = Subtarget.getCallLowering();
4235 CurBuilder->setMF(*MF);
4236 EntryBuilder->setMF(*MF);
4237 MRI = &MF->getRegInfo();
4238 DL = &F.getDataLayout();
4239 const TargetMachine &TM = MF->getTarget();
4241 EnableOpts = OptLevel != CodeGenOptLevel::None && !skipFunction(F);
4242 FuncInfo.MF = MF;
4243 if (EnableOpts) {
4244 AA = &getAnalysis<AAResultsWrapperPass>().getAAResults();
4245 FuncInfo.BPI = &getAnalysis<BranchProbabilityInfoWrapperPass>().getBPI();
4246 AC = &getAnalysis<AssumptionCacheTracker>().getAssumptionCache(
4247 MF->getFunction());
4248 } else {
4249 AA = nullptr;
4250 FuncInfo.BPI = nullptr;
4251 AC = nullptr;
4252 }
4253 LibInfo = &getAnalysis<TargetLibraryInfoWrapperPass>().getTLI(F);
4254 Libcalls = &getAnalysis<LibcallLoweringInfoWrapper>().getLibcallLowering(
4255 *F.getParent(), Subtarget);
4256
4257 FuncInfo.CanLowerReturn = CLI->checkReturnTypeForCallConv(*MF);
4258
4259 SL = std::make_unique<GISelSwitchLowering>(this, FuncInfo);
4260 SL->init(*TLI, TM, *DL);
4261
4262 assert(PendingPHIs.empty() && "stale PHIs");
4263
4264 // Targets which want to use big endian can enable it using
4265 // enableBigEndian()
4266 if (!DL->isLittleEndian() && !CLI->enableBigEndian()) {
4267 // Currently we don't properly handle big endian code.
4268 OptimizationRemarkMissed R("gisel-irtranslator", "GISelFailure",
4269 F.getSubprogram(), &F.getEntryBlock());
4270 R << "unable to translate in big endian mode";
4271 reportTranslationError(*MF, *ORE, R);
4272 return false;
4273 }
4274
4275 // Release the per-function state when we return, whether we succeeded or not.
4276 llvm::scope_exit FinalizeOnReturn([this]() { finalizeFunction(); });
4277
4278 // Setup a separate basic-block for the arguments and constants
4279 MachineBasicBlock *EntryBB = MF->CreateMachineBasicBlock();
4280 MF->push_back(EntryBB);
4281 EntryBuilder->setMBB(*EntryBB);
4282
4283 DebugLoc DbgLoc = F.getEntryBlock().getFirstNonPHIIt()->getDebugLoc();
4284 SwiftError.setFunction(CurMF);
4285 SwiftError.createEntriesInEntryBlock(DbgLoc);
4286
4287 bool IsVarArg = F.isVarArg();
4288 bool HasMustTailInVarArgFn = false;
4289
4290 // Create all blocks, in IR order, to preserve the layout.
4291 FuncInfo.MBBMap.resize(F.getMaxBlockNumber());
4292 for (const BasicBlock &BB: F) {
4293 auto *&MBB = FuncInfo.MBBMap[BB.getNumber()];
4294
4295 MBB = MF->CreateMachineBasicBlock(&BB);
4296 MF->push_back(MBB);
4297
4298 // Only mark the block if the BlockAddress actually has users. The
4299 // hasAddressTaken flag may be stale if the BlockAddress was optimized away
4300 // but the constant still exists in the uniquing table.
4301 if (BB.hasAddressTaken()) {
4302 if (BlockAddress *BA = BlockAddress::lookup(&BB))
4303 if (!BA->hasZeroLiveUses())
4304 MBB->setAddressTakenIRBlock(const_cast<BasicBlock *>(&BB));
4305 }
4306
4307 if (!HasMustTailInVarArgFn)
4308 HasMustTailInVarArgFn = checkForMustTailInVarArgFn(IsVarArg, BB);
4309 }
4310
4311 MF->getFrameInfo().setHasMustTailInVarArgFunc(HasMustTailInVarArgFn);
4312
4313 // Make our arguments/constants entry block fallthrough to the IR entry block.
4314 EntryBB->addSuccessor(&getMBB(F.front()));
4315
4316 // Lower the actual args into this basic block.
4317 SmallVector<ArrayRef<Register>, 8> VRegArgs;
4318 for (const Argument &Arg: F.args()) {
4319 if (DL->getTypeStoreSize(Arg.getType()).isZero())
4320 continue; // Don't handle zero sized types.
4321 ArrayRef<Register> VRegs = getOrCreateVRegs(Arg);
4322 VRegArgs.push_back(VRegs);
4323
4324 if (CLI->supportSwiftError() && Arg.hasSwiftErrorAttr()) {
4325 assert(VRegs.size() == 1 && "Too many vregs for Swift error");
4326 SwiftError.setCurrentVReg(EntryBB, SwiftError.getFunctionArg(), VRegs[0]);
4327 }
4328 }
4329
4330 if (!CLI->lowerFormalArguments(*EntryBuilder, F, VRegArgs, FuncInfo)) {
4331 OptimizationRemarkMissed R("gisel-irtranslator", "GISelFailure",
4332 F.getSubprogram(), &F.getEntryBlock());
4333 R << "unable to lower arguments: "
4334 << ore::NV("Prototype", F.getFunctionType());
4335 reportTranslationError(*MF, *ORE, R);
4336 return false;
4337 }
4338
4339 // Need to visit defs before uses when translating instructions.
4340 GISelObserverWrapper WrapperObserver;
4341 if (EnableCSE && CSEInfo)
4342 WrapperObserver.addObserver(CSEInfo);
4343 {
4345#ifndef NDEBUG
4346 DILocationVerifier Verifier;
4347 WrapperObserver.addObserver(&Verifier);
4348#endif // ifndef NDEBUG
4349 RAIIMFObsDelInstaller ObsInstall(*MF, WrapperObserver);
4350 for (const BasicBlock *BB : RPOT) {
4351 MachineBasicBlock &MBB = getMBB(*BB);
4352 // Set the insertion point of all the following translations to
4353 // the end of this basic block.
4354 CurBuilder->setMBB(MBB);
4355 HasTailCall = false;
4356 for (const Instruction &Inst : *BB) {
4357 // If we translated a tail call in the last step, then we know
4358 // everything after the call is either a return, or something that is
4359 // handled by the call itself. (E.g. a lifetime marker or assume
4360 // intrinsic.) In this case, we should stop translating the block and
4361 // move on.
4362 if (HasTailCall)
4363 break;
4364#ifndef NDEBUG
4365 Verifier.setCurrentInst(&Inst);
4366#endif // ifndef NDEBUG
4367
4368 // Translate any debug-info attached to the instruction.
4369 translateDbgInfo(Inst, *CurBuilder);
4370
4371 if (translate(Inst))
4372 continue;
4373
4374 OptimizationRemarkMissed R("gisel-irtranslator", "GISelFailure",
4375 Inst.getDebugLoc(), BB);
4376 R << "unable to translate instruction: " << ore::NV("Opcode", &Inst);
4377
4378 if (ORE->allowExtraAnalysis("gisel-irtranslator")) {
4379 std::string InstStrStorage;
4380 raw_string_ostream InstStr(InstStrStorage);
4381 InstStr << Inst;
4382
4383 R << ": '" << InstStrStorage << "'";
4384 }
4385
4386 reportTranslationError(*MF, *ORE, R);
4387 return false;
4388 }
4389
4390 if (!finalizeBasicBlock(*BB, MBB)) {
4391 OptimizationRemarkMissed R("gisel-irtranslator", "GISelFailure",
4392 BB->getTerminator()->getDebugLoc(), BB);
4393 R << "unable to translate basic block";
4394 reportTranslationError(*MF, *ORE, R);
4395 return false;
4396 }
4397 }
4398#ifndef NDEBUG
4399 WrapperObserver.removeObserver(&Verifier);
4400#endif
4401 }
4402
4403 finishPendingPhis();
4404
4405 SwiftError.propagateVRegs();
4406
4407 // Merge the argument lowering and constants block with its single
4408 // successor, the LLVM-IR entry block. We want the basic block to
4409 // be maximal.
4410 assert(EntryBB->succ_size() == 1 &&
4411 "Custom BB used for lowering should have only one successor");
4412 // Get the successor of the current entry block.
4413 MachineBasicBlock &NewEntryBB = **EntryBB->succ_begin();
4414 assert(NewEntryBB.pred_size() == 1 &&
4415 "LLVM-IR entry block has a predecessor!?");
4416 // Move all the instruction from the current entry block to the
4417 // new entry block.
4418 NewEntryBB.splice(NewEntryBB.begin(), EntryBB, EntryBB->begin(),
4419 EntryBB->end());
4420
4421 // Update the live-in information for the new entry block.
4422 for (const MachineBasicBlock::RegisterMaskPair &LiveIn : EntryBB->liveins())
4423 NewEntryBB.addLiveIn(LiveIn);
4424 NewEntryBB.sortUniqueLiveIns();
4425
4426 // Get rid of the now empty basic block.
4427 EntryBB->removeSuccessor(&NewEntryBB);
4428 MF->remove(EntryBB);
4429 MF->deleteMachineBasicBlock(EntryBB);
4430
4431 assert(&MF->front() == &NewEntryBB &&
4432 "New entry wasn't next in the list of basic block!");
4433
4434 // Initialize stack protector information.
4436 SP.copyToMachineFrameInfo(MF->getFrameInfo());
4437
4438 return false;
4439}
#define Success
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
amdgpu aa AMDGPU Address space based Alias Analysis Wrapper
MachineBasicBlock & MBB
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
Provides analysis for continuously CSEing during GISel passes.
This file implements a version of MachineIRBuilder which CSEs insts within a MachineBasicBlock.
This file describes how to lower LLVM calls to machine code calls.
This file contains the declarations for the subclasses of Constant, which represent the different fla...
dxil translate DXIL Translate Metadata
This contains common code to allow clients to notify changes to machine instr.
#define DEBUG_TYPE
const HexagonInstrInfo * TII
static bool checkForMustTailInVarArgFn(bool IsVarArg, const BasicBlock &BB)
Returns true if a BasicBlock BB within a variadic function contains a variadic musttail call.
static unsigned getConvOpcode(Intrinsic::ID ID)
static uint64_t getOffsetFromIndices(const User &U, const DataLayout &DL)
static unsigned getConstrainedOpcode(Intrinsic::ID ID)
IRTranslator LLVM IR MI
IRTranslator LLVM IR static false void reportTranslationError(MachineFunction &MF, OptimizationRemarkEmitter &ORE, OptimizationRemarkMissed &R)
static cl::opt< bool > EnableCSEInIRTranslator("enable-cse-in-irtranslator", cl::desc("Should enable CSE in irtranslator"), cl::Optional, cl::init(false))
static bool isValInBlock(const Value *V, const BasicBlock *BB)
static bool isSwiftError(const Value *V)
This file declares the IRTranslator pass.
This file provides various utilities for inspecting and working with the control flow graph in LLVM I...
This file describes how to lower LLVM inline asm to machine code INLINEASM.
const AbstractManglingParser< Derived, Alloc >::OperatorInfo AbstractManglingParser< Derived, Alloc >::Ops[]
static LVOptions Options
Definition LVOptions.cpp:25
Implement a low-level type suitable for MachineInstr level instruction selection.
Implement a low-level type suitable for MachineInstr level instruction selection.
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
Machine Check Debug Module
This file declares the MachineIRBuilder class.
Register Reg
Register const TargetRegisterInfo * TRI
Promote Memory to Register
Definition Mem2Reg.cpp:110
This file contains the declarations for metadata subclasses.
Type::TypeID TypeID
uint64_t High
OptimizedStructLayoutField Field
if(PassOpts->AAPipeline)
#define INITIALIZE_PASS_DEPENDENCY(depName)
Definition PassSupport.h:42
#define INITIALIZE_PASS_END(passName, arg, name, cfg, analysis)
Definition PassSupport.h:44
#define INITIALIZE_PASS_BEGIN(passName, arg, name, cfg, analysis)
Definition PassSupport.h:39
This file builds on the ADT/GraphTraits.h file to build a generic graph post order iterator.
const SmallVectorImpl< MachineOperand > MachineBasicBlock * TBB
const SmallVectorImpl< MachineOperand > & Cond
std::pair< BasicBlock *, BasicBlock * > Edge
This file contains some templates that are useful if you are working with the STL at all.
verify safepoint Safepoint IR Verifier
This file defines the make_scope_exit function, which executes user-defined cleanup logic at scope ex...
This file defines the SmallVector class.
#define LLVM_DEBUG(...)
Definition Debug.h:114
This file describes how to lower LLVM code to machine code.
Target-Independent Code Generator Pass Configuration Options pass.
Value * RHS
Value * LHS
A wrapper pass to provide the legacy pass manager access to a suitably prepared AAResults object.
LLVM_ABI APInt zextOrTrunc(unsigned width) const
Zero extend or truncate to width.
Definition APInt.cpp:1075
an instruction to allocate memory on the stack
bool isSwiftError() const
Return true if this alloca is used as a swifterror argument to a call.
LLVM_ABI bool isStaticAlloca() const
Return true if this alloca is in the entry block of the function and is a constant size.
Align getAlign() const
Return the alignment of the memory that is being allocated by the instruction.
PointerType * getType() const
Overload to return most specific pointer type.
Type * getAllocatedType() const
Return the type that is being allocated by the instruction.
LLVM_ABI std::optional< TypeSize > getAllocationSize(const DataLayout &DL) const
Get allocation size in bytes.
const Value * getArraySize() const
Get the number of elements allocated.
Represent the analysis usage information of a pass.
AnalysisUsage & addRequired()
AnalysisUsage & addPreserved()
Add the specified Pass class to the set of analyses preserved by this pass.
This class represents an incoming formal argument to a Function.
Definition Argument.h:32
Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:39
iterator end() const
Definition ArrayRef.h:129
size_t size() const
Get the array size.
Definition ArrayRef.h:140
iterator begin() const
Definition ArrayRef.h:128
bool empty() const
Check if the array is empty.
Definition ArrayRef.h:135
An immutable pass that tracks lazily created AssumptionCache objects.
@ Add
*p = old + v
@ FAdd
*p = old + v
@ USubCond
Subtract only if no unsigned overflow.
@ FMinimum
*p = minimum(old, v) minimum matches the behavior of llvm.minimum.
@ Min
*p = old <signed v ? old : v
@ Sub
*p = old - v
@ And
*p = old & v
@ Xor
*p = old ^ v
@ USubSat
*p = usub.sat(old, v) usub.sat matches the behavior of llvm.usub.sat.
@ FMaximum
*p = maximum(old, v) maximum matches the behavior of llvm.maximum.
@ FSub
*p = old - v
@ UIncWrap
Increment one up to a maximum value.
@ Max
*p = old >signed v ? old : v
@ UMin
*p = old <unsigned v ? old : v
@ FMin
*p = minnum(old, v) minnum matches the behavior of llvm.minnum.
@ UMax
*p = old >unsigned v ? old : v
@ FMaximumNum
*p = maximumnum(old, v) maximumnum matches the behavior of llvm.maximumnum.
@ FMax
*p = maxnum(old, v) maxnum matches the behavior of llvm.maxnum.
@ UDecWrap
Decrement one until a minimum value or zero.
@ FMinimumNum
*p = minimumnum(old, v) minimumnum matches the behavior of llvm.minimumnum.
@ Nand
*p = ~(old & v)
LLVM Basic Block Representation.
Definition BasicBlock.h:62
unsigned getNumber() const
Definition BasicBlock.h:95
const Function * getParent() const
Return the enclosing method, or null if none.
Definition BasicBlock.h:213
bool hasAddressTaken() const
Returns true if there are any uses of this basic block other than direct branches,...
Definition BasicBlock.h:687
LLVM_ABI InstListType::const_iterator getFirstNonPHIIt() const
Returns an iterator to the first instruction in this block that is not a PHINode instruction.
InstListType::const_iterator const_iterator
Definition BasicBlock.h:171
LLVM_ABI InstListType::const_iterator getFirstNonPHIOrDbg(bool SkipPseudoOp=true) const
Returns a pointer to the first instruction in this block that is not a PHINode or a debug intrinsic,...
LLVM_ABI const Module * getModule() const
Return the module owning the function this basic block belongs to, or nullptr if the function does no...
The address of a basic block.
Definition Constants.h:1071
static LLVM_ABI BlockAddress * lookup(const BasicBlock *BB)
Lookup an existing BlockAddress constant for the given BasicBlock.
Legacy analysis pass which computes BlockFrequencyInfo.
Legacy analysis pass which computes BranchProbabilityInfo.
LLVM_ABI BranchProbability getEdgeProbability(const BasicBlock *Src, unsigned IndexInSuccessors) const
Get an edge's probability, relative to other out-edges of the Src.
static BranchProbability getOne()
static BranchProbability getZero()
static void normalizeProbabilities(ProbabilityIter Begin, ProbabilityIter End)
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
bool isInlineAsm() const
Check if this call is an inline asm statement.
std::optional< OperandBundleUse > getOperandBundle(StringRef Name) const
Return an operand bundle by name, if present.
Function * getCalledFunction() const
Returns the function called, or null if this is an indirect function invocation or the function signa...
LLVM_ABI bool paramHasAttr(unsigned ArgNo, Attribute::AttrKind Kind) const
Determine whether the argument or parameter has the given attribute.
User::op_iterator arg_begin()
Return the iterator pointing to the beginning of the argument list.
unsigned countOperandBundlesOfType(StringRef Name) const
Return the number of operand bundles with the tag Name attached to this instruction.
Value * getCalledOperand() const
Value * getArgOperand(unsigned i) const
User::op_iterator arg_end()
Return the iterator pointing to the end of the argument list.
bool isConvergent() const
Determine if the invoke is convergent.
LLVM_ABI Intrinsic::ID getIntrinsicID() const
Returns the intrinsic ID of the intrinsic called or Intrinsic::not_intrinsic if the called function i...
iterator_range< User::op_iterator > args()
Iteration adapter for range-for loops.
unsigned arg_size() const
AttributeList getAttributes() const
Return the attributes for this call.
This class represents a function call, abstracting a target machine's calling convention.
bool isTailCall() const
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition InstrTypes.h:676
@ FCMP_TRUE
1 1 1 1 Always true (always folded)
Definition InstrTypes.h:693
@ ICMP_SLT
signed less than
Definition InstrTypes.h:705
@ ICMP_SLE
signed less or equal
Definition InstrTypes.h:706
@ ICMP_UGT
unsigned greater than
Definition InstrTypes.h:699
@ ICMP_NE
not equal
Definition InstrTypes.h:698
@ ICMP_ULE
unsigned less or equal
Definition InstrTypes.h:702
@ FCMP_FALSE
0 0 0 0 Always false (always folded)
Definition InstrTypes.h:678
bool isFPPredicate() const
Definition InstrTypes.h:782
bool isIntPredicate() const
Definition InstrTypes.h:783
Value * getCondition() const
BasicBlock * getSuccessor(unsigned i) const
static LLVM_ABI ConstantInt * getTrue(LLVMContext &Context)
bool isZero() const
This is just a convenience method to make client code smaller for a common code.
Definition Constants.h:219
unsigned getBitWidth() const
getBitWidth - Return the scalar bitwidth of this constant.
Definition Constants.h:162
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
Definition Constants.h:168
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition Constants.h:159
This is an important base class in LLVM.
Definition Constant.h:43
static LLVM_ABI Constant * getAllOnesValue(Type *Ty)
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
This is the common base class for constrained floating point intrinsics.
LLVM_ABI std::optional< fp::ExceptionBehavior > getExceptionBehavior() const
LLVM_ABI unsigned getNonMetadataArgCount() const
DWARF expression.
LLVM_ABI bool isEntryValue() const
Check if the expression consists of exactly one entry value operand.
static LLVM_ABI DIExpression * append(const DIExpression *Expr, ArrayRef< uint64_t > Ops)
Append the opcodes Ops to DIExpr.
LLVM_ABI bool startsWithDeref() const
Return whether the first element a DW_OP_deref.
ArrayRef< uint64_t > getElements() const
bool isValidLocationForIntrinsic(const DILocation *DL) const
Check that a location is valid for this label.
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:64
Value * getAddress() const
DILabel * getLabel() const
DebugLoc getDebugLoc() const
Value * getValue(unsigned OpIdx=0) const
DILocalVariable * getVariable() const
DIExpression * getExpression() const
LLVM_ABI Value * getVariableLocationOp(unsigned OpIdx) const
DIExpression * getExpression() const
DILocalVariable * getVariable() const
A debug info location.
Definition DebugLoc.h:123
Class representing an expression and its matching format.
This instruction extracts a struct member or array element value from an aggregate value.
static LLVM_ABI FixedVectorType * get(Type *ElementType, unsigned NumElts)
Definition Type.cpp:873
bool skipFunction(const Function &F) const
Optional passes call this function to check whether the pass should be skipped.
Definition Pass.cpp:188
const BasicBlock & getEntryBlock() const
Definition Function.h:809
DISubprogram * getSubprogram() const
Get the attached subprogram.
bool hasMinSize() const
Optimize this function for minimum size (-Oz).
Definition Function.h:711
Constant * getPersonalityFn() const
Get the personality function associated with this function.
const Function & getFunction() const
Definition Function.h:166
bool isIntrinsic() const
isIntrinsic - Returns true if the function's name starts with "llvm.".
Definition Function.h:251
The actual analysis pass wrapper.
Definition CSEInfo.h:242
Simple wrapper that does the following.
Definition CSEInfo.h:212
The CSE Analysis object.
Definition CSEInfo.h:72
Abstract class that contains various methods for clients to notify about changes.
Simple wrapper observer that takes several observers, and calls each one for each event.
void removeObserver(GISelChangeObserver *O)
void addObserver(GISelChangeObserver *O)
static StringRef dropLLVMManglingEscape(StringRef Name)
If the given string begins with the GlobalValue name mangling escape character '\1',...
bool hasExternalWeakLinkage() const
bool hasDLLImportStorageClass() const
Module * getParent()
Get the module that this global value is contained inside of...
bool isTailCall(const MachineInstr &MI) const override
bool runOnMachineFunction(MachineFunction &MF) override
runOnMachineFunction - This method must be overloaded to perform the desired machine code transformat...
IRTranslator(CodeGenOptLevel OptLevel=CodeGenOptLevel::None)
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - This function should be overriden by passes that need analysis information to do t...
bool lowerInlineAsm(MachineIRBuilder &MIRBuilder, const CallBase &CB, std::function< ArrayRef< Register >(const Value &Val)> GetOrCreateVRegs) const
Lower the given inline asm call instruction GetOrCreateVRegs is a callback to materialize a register ...
This instruction inserts a struct field of array element value into an aggregate value.
iterator_range< simple_ilist< DbgRecord >::iterator > getDbgRecordRange() const
Return a range over the DbgRecords attached to this instruction.
const DebugLoc & getDebugLoc() const
Return the debug location for this node as a DebugLoc.
LLVM_ABI const Module * getModule() const
Return the module owning the function this instruction belongs to or nullptr it the function does not...
MDNode * getMetadata(unsigned KindID) const
Get the metadata of given kind attached to this Instruction.
LLVM_ABI AAMDNodes getAAMetadata() const
Returns the AA metadata for this instruction.
unsigned getOpcode() const
Returns a member of one of the enums like Instruction::Add.
LLVM_ABI bool hasAllowReassoc() const LLVM_READONLY
Determine whether the allow-reassociation flag is set.
Intrinsic::ID getIntrinsicID() const
Return the intrinsic ID of this intrinsic.
static bool getUseExtended()
constexpr LLT changeElementType(LLT NewEltTy) const
If this type is a vector, return a vector with the same number of elements but the new element type.
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr uint16_t getNumElements() const
Returns the number of elements in a vector LLT.
constexpr bool isVector() const
static constexpr LLT pointer(unsigned AddressSpace, unsigned SizeInBits)
Get a low-level pointer in the given address space.
constexpr TypeSize getSizeInBits() const
Returns the total size of the type. Must only be called on sized types.
constexpr bool isPointer() const
static constexpr LLT fixed_vector(unsigned NumElements, unsigned ScalarSizeInBits)
Get a low-level fixed-width vector of some number of elements and element width.
constexpr bool isFixedVector() const
Returns true if the LLT is a fixed vector.
static LLT integer(unsigned SizeInBits)
LLT changeElementSize(unsigned NewEltSize) const
If this type is a vector, return a vector with the same number of elements but the new element size.
LLVM_ABI void diagnose(const DiagnosticInfo &DI)
Report a message to the currently installed diagnostic handler.
Value * getPointerOperand()
AtomicOrdering getOrdering() const
Returns the ordering constraint of this load instruction.
SyncScope::ID getSyncScopeID() const
Returns the synchronization scope ID of this load instruction.
static LocationSize precise(uint64_t Value)
static MDTuple * get(LLVMContext &Context, ArrayRef< Metadata * > MDs)
Definition Metadata.h:1572
void normalizeSuccProbs()
Normalize probabilities of all successors so that the sum of them becomes one.
LLVM_ABI instr_iterator insert(instr_iterator I, MachineInstr *M)
Insert MI into the instruction list before I, possibly inside a bundle.
void push_back(MachineInstr *MI)
const BasicBlock * getBasicBlock() const
Return the LLVM basic block that this instance corresponded to originally.
LLVM_ABI void setSuccProbability(succ_iterator I, BranchProbability Prob)
Set successor probability of a given iterator.
LLVM_ABI void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
SmallVectorImpl< MachineBasicBlock * >::iterator succ_iterator
LLVM_ABI void sortUniqueLiveIns()
Sorts and uniques the LiveIns vector.
LLVM_ABI bool isPredecessor(const MachineBasicBlock *MBB) const
Return true if the specified MBB is a predecessor of this block.
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineInstrBundleIterator< MachineInstr > iterator
void setIsEHPad(bool V=true)
Indicates the block is a landing pad.
int getStackProtectorIndex() const
Return the index for the stack protector object.
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
Function & getFunction()
Return the LLVM function that this machine code represents.
BasicBlockListType::iterator iterator
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineInstr - Allocate a new MachineInstr.
void insert(iterator MBBI, MachineBasicBlock *MBB)
Helper class to build MachineInstr.
MachineInstrBuilder buildFPTOUI_SAT(const DstOp &Dst, const SrcOp &Src0)
Build and insert Res = G_FPTOUI_SAT Src0.
MachineInstrBuilder buildFMul(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
MachineInstrBuilder buildFreeze(const DstOp &Dst, const SrcOp &Src)
Build and insert Dst = G_FREEZE Src.
MachineInstrBuilder buildBr(MachineBasicBlock &Dest)
Build and insert G_BR Dest.
MachineInstrBuilder buildModf(const DstOp &Fract, const DstOp &Int, const SrcOp &Src, std::optional< unsigned > Flags=std::nullopt)
Build and insert Fract, Int = G_FMODF Src.
LLVMContext & getContext() const
MachineInstrBuilder buildAdd(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_ADD Op0, Op1.
MachineInstrBuilder buildUndef(const DstOp &Res)
Build and insert Res = IMPLICIT_DEF.
MachineInstrBuilder buildResetFPMode()
Build and insert G_RESET_FPMODE.
MachineInstrBuilder buildFPTOSI_SAT(const DstOp &Dst, const SrcOp &Src0)
Build and insert Res = G_FPTOSI_SAT Src0.
MachineInstrBuilder buildUCmp(const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1)
Build and insert a Res = G_UCMP Op0, Op1.
MachineInstrBuilder buildJumpTable(const LLT PtrTy, unsigned JTI)
Build and insert Res = G_JUMP_TABLE JTI.
MachineInstrBuilder buildGetRounding(const DstOp &Dst)
Build and insert Dst = G_GET_ROUNDING.
MachineInstrBuilder buildSCmp(const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1)
Build and insert a Res = G_SCMP Op0, Op1.
MachineInstrBuilder buildFence(unsigned Ordering, unsigned Scope)
Build and insert G_FENCE Ordering, Scope.
MachineInstrBuilder buildSelect(const DstOp &Res, const SrcOp &Tst, const SrcOp &Op0, const SrcOp &Op1, std::optional< unsigned > Flags=std::nullopt)
Build and insert a Res = G_SELECT Tst, Op0, Op1.
MachineInstrBuilder buildFMA(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, const SrcOp &Src2, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_FMA Op0, Op1, Op2.
MachineInstrBuilder buildMul(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_MUL Op0, Op1.
MachineInstrBuilder buildInsertSubvector(const DstOp &Res, const SrcOp &Src0, const SrcOp &Src1, unsigned Index)
Build and insert Res = G_INSERT_SUBVECTOR Src0, Src1, Idx.
MachineInstrBuilder buildAnd(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1)
Build and insert Res = G_AND Op0, Op1.
MachineInstrBuilder buildCast(const DstOp &Dst, const SrcOp &Src)
Build and insert an appropriate cast between two registers of equal size.
MachineInstrBuilder buildICmp(CmpInst::Predicate Pred, const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1, std::optional< unsigned > Flags=std::nullopt)
Build and insert a Res = G_ICMP Pred, Op0, Op1.
MachineBasicBlock::iterator getInsertPt()
Current insertion point for new instructions.
MachineInstrBuilder buildSExtOrTrunc(const DstOp &Res, const SrcOp &Op)
Build and insert Res = G_SEXT Op, Res = G_TRUNC Op, or Res = COPY Op depending on the differing sizes...
MachineInstrBuilder buildAtomicRMW(unsigned Opcode, const DstOp &OldValRes, const SrcOp &Addr, const SrcOp &Val, MachineMemOperand &MMO)
Build and insert OldValRes<def> = G_ATOMICRMW_<Opcode> Addr, Val, MMO.
MachineInstrBuilder buildSub(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_SUB Op0, Op1.
MachineInstrBuilder buildIntrinsic(Intrinsic::ID ID, ArrayRef< Register > Res, bool HasSideEffects, bool isConvergent)
Build and insert a G_INTRINSIC instruction.
MachineInstrBuilder buildVScale(const DstOp &Res, unsigned MinElts)
Build and insert Res = G_VSCALE MinElts.
MachineInstrBuilder buildSplatBuildVector(const DstOp &Res, const SrcOp &Src)
Build and insert Res = G_BUILD_VECTOR with Src replicated to fill the number of elements.
MachineInstrBuilder buildSetFPMode(const SrcOp &Src)
Build and insert G_SET_FPMODE Src.
MachineInstrBuilder buildIndirectDbgValue(Register Reg, const MDNode *Variable, const MDNode *Expr)
Build and insert a DBG_VALUE instruction expressing the fact that the associated Variable lives in me...
MachineInstrBuilder buildBuildVector(const DstOp &Res, ArrayRef< Register > Ops)
Build and insert Res = G_BUILD_VECTOR Op0, ...
MachineInstrBuilder buildConstDbgValue(const Constant &C, const MDNode *Variable, const MDNode *Expr)
Build and insert a DBG_VALUE instructions specifying that Variable is given by C (suitably modified b...
MachineInstrBuilder buildBrCond(const SrcOp &Tst, MachineBasicBlock &Dest)
Build and insert G_BRCOND Tst, Dest.
std::optional< MachineInstrBuilder > materializeObjectPtrOffset(Register &Res, Register Op0, const LLT ValueTy, uint64_t Value)
Materialize and insert an instruction with appropriate flags for addressing some offset of an object,...
MachineInstrBuilder buildSetRounding(const SrcOp &Src)
Build and insert G_SET_ROUNDING.
MachineInstrBuilder buildExtractVectorElement(const DstOp &Res, const SrcOp &Val, const SrcOp &Idx)
Build and insert Res = G_EXTRACT_VECTOR_ELT Val, Idx.
MachineInstrBuilder buildLoad(const DstOp &Res, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert Res = G_LOAD Addr, MMO.
MachineInstrBuilder buildPtrAdd(const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_PTR_ADD Op0, Op1.
MachineInstrBuilder buildZExtOrTrunc(const DstOp &Res, const SrcOp &Op)
Build and insert Res = G_ZEXT Op, Res = G_TRUNC Op, or Res = COPY Op depending on the differing sizes...
MachineInstrBuilder buildExtractVectorElementConstant(const DstOp &Res, const SrcOp &Val, const int Idx)
Build and insert Res = G_EXTRACT_VECTOR_ELT Val, Idx.
MachineInstrBuilder buildShl(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
MachineInstrBuilder buildStore(const SrcOp &Val, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert G_STORE Val, Addr, MMO.
MachineInstrBuilder buildInstr(unsigned Opcode)
Build and insert <empty> = Opcode <empty>.
MachineInstrBuilder buildFrameIndex(const DstOp &Res, int Idx)
Build and insert Res = G_FRAME_INDEX Idx.
MachineInstrBuilder buildDirectDbgValue(Register Reg, const MDNode *Variable, const MDNode *Expr)
Build and insert a DBG_VALUE instruction expressing the fact that the associated Variable lives in Re...
MachineInstrBuilder buildDbgLabel(const MDNode *Label)
Build and insert a DBG_LABEL instructions specifying that Label is given.
MachineInstrBuilder buildBrJT(Register TablePtr, unsigned JTI, Register IndexReg)
Build and insert G_BRJT TablePtr, JTI, IndexReg.
MachineInstrBuilder buildDynStackAlloc(const DstOp &Res, const SrcOp &Size, Align Alignment)
Build and insert Res = G_DYN_STACKALLOC Size, Align.
MachineInstrBuilder buildFIDbgValue(int FI, const MDNode *Variable, const MDNode *Expr)
Build and insert a DBG_VALUE instruction expressing the fact that the associated Variable lives in th...
MachineInstrBuilder buildResetFPEnv()
Build and insert G_RESET_FPENV.
void setDebugLoc(const DebugLoc &DL)
Set the debug location to DL for all the next build instructions.
const MachineBasicBlock & getMBB() const
Getter for the basic block we currently build.
MachineInstrBuilder buildInsertVectorElement(const DstOp &Res, const SrcOp &Val, const SrcOp &Elt, const SrcOp &Idx)
Build and insert Res = G_INSERT_VECTOR_ELT Val, Elt, Idx.
MachineInstrBuilder buildAtomicCmpXchgWithSuccess(const DstOp &OldValRes, const DstOp &SuccessRes, const SrcOp &Addr, const SrcOp &CmpVal, const SrcOp &NewVal, MachineMemOperand &MMO)
Build and insert OldValRes<def>, SuccessRes<def> = / G_ATOMIC_CMPXCHG_WITH_SUCCESS Addr,...
void setMBB(MachineBasicBlock &MBB)
Set the insertion point to the end of MBB.
const DebugLoc & getDebugLoc()
Get the current instruction's debug location.
MachineInstrBuilder buildTrap(bool Debug=false)
Build and insert G_TRAP or G_DEBUGTRAP.
MachineInstrBuilder buildFFrexp(const DstOp &Fract, const DstOp &Exp, const SrcOp &Src, std::optional< unsigned > Flags=std::nullopt)
Build and insert Fract, Exp = G_FFREXP Src.
MachineInstrBuilder buildFSincos(const DstOp &Sin, const DstOp &Cos, const SrcOp &Src, std::optional< unsigned > Flags=std::nullopt)
Build and insert Sin, Cos = G_FSINCOS Src.
MachineInstrBuilder buildShuffleVector(const DstOp &Res, const SrcOp &Src1, const SrcOp &Src2, ArrayRef< int > Mask)
Build and insert Res = G_SHUFFLE_VECTOR Src1, Src2, Mask.
MachineInstrBuilder buildInstrNoInsert(unsigned Opcode)
Build but don't insert <empty> = Opcode <empty>.
MachineInstrBuilder buildCopy(const DstOp &Res, const SrcOp &Op)
Build and insert Res = COPY Op.
MachineInstrBuilder buildPrefetch(const SrcOp &Addr, unsigned RW, unsigned Locality, unsigned CacheType, MachineMemOperand &MMO)
Build and insert G_PREFETCH Addr, RW, Locality, CacheType.
MachineInstrBuilder buildExtractSubvector(const DstOp &Res, const SrcOp &Src, unsigned Index)
Build and insert Res = G_EXTRACT_SUBVECTOR Src, Idx0.
const DataLayout & getDataLayout() const
MachineInstrBuilder buildBrIndirect(Register Tgt)
Build and insert G_BRINDIRECT Tgt.
MachineInstrBuilder buildSplatVector(const DstOp &Res, const SrcOp &Val)
Build and insert Res = G_SPLAT_VECTOR Val.
MachineInstrBuilder buildStepVector(const DstOp &Res, unsigned Step)
Build and insert Res = G_STEP_VECTOR Step.
virtual MachineInstrBuilder buildConstant(const DstOp &Res, const ConstantInt &Val)
Build and insert Res = G_CONSTANT Val.
MachineInstrBuilder buildFCmp(CmpInst::Predicate Pred, const DstOp &Res, const SrcOp &Op0, const SrcOp &Op1, std::optional< unsigned > Flags=std::nullopt)
Build and insert a Res = G_FCMP PredOp0, Op1.
MachineInstrBuilder buildFAdd(const DstOp &Dst, const SrcOp &Src0, const SrcOp &Src1, std::optional< unsigned > Flags=std::nullopt)
Build and insert Res = G_FADD Op0, Op1.
MachineInstrBuilder buildSetFPEnv(const SrcOp &Src)
Build and insert G_SET_FPENV Src.
Register getReg(unsigned Idx) const
Get the register for the operand index.
const MachineInstrBuilder & addExternalSymbol(const char *FnName, unsigned TargetFlags=0) const
const MachineInstrBuilder & addUse(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register use operand.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addMetadata(const MDNode *MD) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addFrameIndex(int Idx) const
const MachineInstrBuilder & addFPImm(const ConstantFP *Val) const
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
const MachineInstrBuilder & addDef(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a virtual register definition operand.
const MachineInstrBuilder & addMemOperand(MachineMemOperand *MMO) const
MachineInstr * getInstr() const
If conversion operators fail, use this method to get the MachineInstr explicitly.
LLVM_ABI void copyIRFlags(const Instruction &I)
Copy all flags to MachineInst MIFlags.
static LLVM_ABI uint32_t copyFlagsFromInstruction(const Instruction &I)
LLVM_ABI void setDeactivationSymbol(MachineFunction &MF, Value *DS)
void setDebugLoc(DebugLoc DL)
Replace current source information with new such.
Flags
Flags values. These may be or'd together.
@ MOVolatile
The memory access is volatile.
@ MODereferenceable
The memory access is dereferenceable (i.e., doesn't trap).
@ MOLoad
The memory access reads data.
@ MOInvariant
The memory access always returns the same value (or traps).
@ MOStore
The memory access writes data.
static MachineOperand CreateES(const char *SymName, unsigned TargetFlags=0)
static MachineOperand CreateGA(const GlobalValue *GV, int64_t Offset, unsigned TargetFlags=0)
LLVM_ABI Register createGenericVirtualRegister(LLT Ty, StringRef Name="")
Create and return a new generic virtual register with low-level type Ty.
The optimization diagnostic interface.
Diagnostic information for missed-optimization remarks.
BasicBlock * getIncomingBlock(unsigned i) const
Return incoming basic block number i.
Value * getIncomingValue(unsigned i) const
Return incoming value number x.
unsigned getNumIncomingValues() const
Return the number of incoming edges.
AnalysisType & getAnalysis() const
getAnalysis<AnalysisType>() - This function is used by subclasses to get to the analysis information ...
static PointerType * getUnqual(Type *ElementType)
This constructs a pointer to an object of the specified type in the default address space (address sp...
Class to install both of the above.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
Value * getReturnValue() const
Convenience accessor. Returns null if there is no return value.
size_type count(ConstPtrType Ptr) const
count - Return 1 if the specified pointer is in the set, 0 otherwise.
std::pair< iterator, bool > insert(PtrType Ptr)
Inserts Ptr if and only if there is no element in the container equal to Ptr.
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Encapsulates all of the information needed to generate a stack protector check, and signals to isel w...
MachineBasicBlock * getSuccessMBB()
MachineBasicBlock * getFailureMBB()
constexpr bool empty() const
Check if the string is empty.
Definition StringRef.h:140
constexpr const char * data() const
Get a pointer to the start of the string (which may not be null terminated).
Definition StringRef.h:137
Primary interface to the complete machine description for the target machine.
const Triple & getTargetTriple() const
TargetOptions Options
const Target & getTarget() const
void resetTargetOptions(const Function &F) const
Reset the target options based on the function's attributes.
unsigned NoTrapAfterNoreturn
Do not emit a trap instruction for 'unreachable' IR instructions behind noreturn calls,...
unsigned TrapUnreachable
Emit target-specific trap instruction for 'unreachable' IR instructions.
FPOpFusion::FPOpFusionMode AllowFPOpFusion
AllowFPOpFusion - This flag is set by the -fp-contract=xxx option.
Target-Independent Code Generator Pass Configuration Options.
TargetSubtargetInfo - Generic base class for all target subtargets.
virtual const CallLowering * getCallLowering() const
virtual const TargetLowering * getTargetLowering() const
bool isSPIRV() const
Tests whether the target is SPIR-V (32/64-bit/Logical).
Definition Triple.h:887
Twine - A lightweight data structure for efficiently representing the concatenation of temporary valu...
Definition Twine.h:82
static constexpr TypeSize getZero()
Definition TypeSize.h:349
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:46
LLVM_ABI bool isEmptyTy() const
Return true if this type is empty, that is, it has no elements or all of its elements are empty.
Definition Type.cpp:184
static LLVM_ABI IntegerType * getInt32Ty(LLVMContext &C)
Definition Type.cpp:313
static LLVM_ABI Type * getVoidTy(LLVMContext &C)
Definition Type.cpp:286
bool isSized(SmallPtrSetImpl< Type * > *Visited=nullptr) const
Return true if it makes sense to take the size of this type.
Definition Type.h:328
bool isAggregateType() const
Return true if the type is an aggregate type.
Definition Type.h:321
bool isTokenTy() const
Return true if this is 'token'.
Definition Type.h:236
bool isVoidTy() const
Return true if this is 'void'.
Definition Type.h:141
BasicBlock * getSuccessor(unsigned i=0) const
Value * getOperand(unsigned i) const
Definition User.h:207
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:255
bool hasOneUse() const
Return true if there is exactly one use of this value.
Definition Value.h:439
LLVMContext & getContext() const
All values hold a context through their type.
Definition Value.h:258
LLVM_ABI const Value * stripPointerCasts() const
Strip off pointer casts, all-zero GEPs and address space casts.
Definition Value.cpp:709
constexpr ScalarTy getFixedValue() const
Definition TypeSize.h:200
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition TypeSize.h:168
constexpr ScalarTy getKnownMinValue() const
Returns the minimum value this quantity can represent.
Definition TypeSize.h:165
constexpr bool isZero() const
Definition TypeSize.h:153
const ParentTy * getParent() const
Definition ilist_node.h:34
NodeTy * getNextNode()
Get the next node, or nullptr for the list tail.
Definition ilist_node.h:348
A raw_ostream that writes to an std::string.
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
constexpr char Args[]
Key for Kernel::Metadata::mArgs.
constexpr char SymbolName[]
Key for Kernel::Metadata::mSymbolName.
constexpr std::underlying_type_t< E > Mask()
Get a bitmask with 1s in all places up to the high-order bit of E's largest value.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
@ C
The default llvm calling convention, compatible with C.
Definition CallingConv.h:34
@ BasicBlock
Various leaf nodes.
Definition ISDOpcodes.h:81
BinaryOp_match< SrcTy, SpecificConstantMatch, TargetOpcode::G_XOR, true > m_Not(const SrcTy &&Src)
Matches a register not-ed by a G_XOR.
OneUse_match< SubPat > m_OneUse(const SubPat &SP)
bool match(Val *V, const Pattern &P)
specificval_ty m_Specific(const Value *V)
Match if we have a specific specified value.
TwoOps_match< Val_t, Idx_t, Instruction::ExtractElement > m_ExtractElt(const Val_t &Val, const Idx_t &Idx)
Matches ExtractElementInst.
auto m_Value()
Match an arbitrary value and ignore it.
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
Offsets
Offsets in bytes from the start of the input buffer.
void sortAndRangeify(CaseClusterVector &Clusters)
Sort Clusters and merge adjacent cases.
std::vector< CaseCluster > CaseClusterVector
@ CC_Range
A cluster of adjacent case labels with the same destination, or just one case.
@ CC_JumpTable
A cluster of cases suitable for jump table lowering.
@ CC_BitTests
A cluster of cases suitable for bit test lowering.
SmallVector< SwitchWorkListItem, 4 > SwitchWorkList
CaseClusterVector::iterator CaseClusterIt
@ CE
Windows NT (Windows on ARM)
Definition MCAsmInfo.h:48
initializer< Ty > init(const Ty &Val)
ExceptionBehavior
Exception behavior used for floating point operations.
Definition FPEnv.h:39
@ ebIgnore
This corresponds to "fpexcept.ignore".
Definition FPEnv.h:40
DiagnosticInfoOptimizationBase::Argument NV
NodeAddr< PhiNode * > Phi
Definition RDFGraph.h:390
NodeAddr< CodeNode * > Code
Definition RDFGraph.h:388
friend class Instruction
Iterator for Instructions in a `BasicBlock.
Definition BasicBlock.h:73
BaseReg
Stack frame base register. Bit 0 of FREInfo.Info.
Definition SFrame.h:77
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
Definition STLExtras.h:315
@ Low
Lower the current thread's priority such that it does not affect foreground tasks significantly.
Definition Threading.h:280
@ Offset
Definition DWP.cpp:557
FunctionAddr VTableAddr Value
Definition InstrProf.h:137
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Undef
Value of the register doesn't matter.
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are tuples (A, B,...
Definition STLExtras.h:2553
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
int countr_one(T Value)
Count the number of ones from the least significant bit to the first zero bit.
Definition bit.h:315
FunctionAddr VTableAddr uintptr_t uintptr_t Int32Ty
Definition InstrProf.h:328
LLVM_ABI void diagnoseDontCall(const CallInst &CI)
auto successors(const MachineBasicBlock *BB)
LLVM_ABI MVT getMVTForLLT(LLT Ty)
Get a rough equivalent of an MVT for a given LLT.
void append_range(Container &C, Range &&R)
Wrapper function to append range R to container C.
Definition STLExtras.h:2207
constexpr bool isUIntN(unsigned N, uint64_t x)
Checks if an unsigned integer fits into the given (dynamic) bit width.
Definition MathExtras.h:243
gep_type_iterator gep_type_end(const User *GEP)
MachineBasicBlock::iterator findSplitPointForStackProtector(MachineBasicBlock *BB, const TargetInstrInfo &TII)
Find the split point at which to splice the end of BB into its success stack protector check machine ...
LLVM_ABI LLT getLLTForMVT(MVT Ty)
Get a rough equivalent of an LLT for a given MVT.
constexpr int popcount(T Value) noexcept
Count the number of set bits in a value.
Definition bit.h:156
int countr_zero(T Val)
Count number of 0's from the least significant bit to the most stopping at the first 1.
Definition bit.h:204
Align getKnownAlignment(Value *V, const DataLayout &DL, const Instruction *CxtI=nullptr, AssumptionCache *AC=nullptr, const DominatorTree *DT=nullptr)
Try to infer an alignment for the specified pointer.
Definition Local.h:252
constexpr bool has_single_bit(T Value) noexcept
Definition bit.h:149
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition STLExtras.h:1745
LLVM_ABI llvm::SmallVector< int, 16 > createStrideMask(unsigned Start, unsigned Stride, unsigned VF)
Create a stride shuffle mask.
auto reverse(ContainerTy &&C)
Definition STLExtras.h:407
void sort(IteratorTy Start, IteratorTy End)
Definition STLExtras.h:1635
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:207
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
Definition Error.cpp:163
generic_gep_type_iterator<> gep_type_iterator
auto succ_size(const MachineBasicBlock *BB)
LLVM_ABI EHPersonality classifyEHPersonality(const Value *Pers)
See if the given exception handling personality function is one that we understand.
CodeGenOptLevel
Code generation optimization level.
Definition CodeGen.h:82
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
@ Success
The lock was released successfully.
LLVM_ATTRIBUTE_VISIBILITY_DEFAULT AnalysisKey InnerAnalysisManagerProxy< AnalysisManagerT, IRUnitT, ExtraArgTs... >::Key
@ Global
Append to llvm.global_dtors.
@ First
Helpers to iterate all locations in the MemoryEffectsBase class.
Definition ModRef.h:74
LLVM_ABI void getSelectionDAGFallbackAnalysisUsage(AnalysisUsage &AU)
Modify analysis usage so it preserves passes required for the SelectionDAG fallback.
Definition Utils.cpp:1142
auto lower_bound(R &&Range, T &&Value)
Provide wrappers to std::lower_bound which take ranges instead of having to pass begin/end explicitly...
Definition STLExtras.h:2051
LLVM_ABI llvm::SmallVector< int, 16 > createInterleaveMask(unsigned VF, unsigned NumVecs)
Create an interleave shuffle mask.
@ FMul
Product of floats.
@ Sub
Subtraction of integers.
DWARFExpression::Operation Op
ArrayRef(const T &OneElt) -> ArrayRef< T >
bool isAsynchronousEHPersonality(EHPersonality Pers)
Returns true if this personality function catches asynchronous exceptions.
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
LLVM_ABI std::optional< RoundingMode > convertStrToRoundingMode(StringRef)
Returns a valid RoundingMode enumerator when given a string that is valid as input in constrained int...
Definition FPEnv.cpp:25
gep_type_iterator gep_type_begin(const User *GEP)
void computeValueLLTs(const DataLayout &DL, Type &Ty, SmallVectorImpl< LLT > &ValueLLTs, SmallVectorImpl< TypeSize > *Offsets=nullptr, TypeSize StartingOffset=TypeSize::getZero())
computeValueLLTs - Given an LLVM IR type, compute a sequence of LLTs that represent all the individua...
Definition Analysis.cpp:153
GlobalValue * ExtractTypeInfo(Value *V)
ExtractTypeInfo - Returns the type info, possibly bitcast, encoded in V.
Definition Analysis.cpp:181
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
Definition Alignment.h:201
LLVM_ABI LLT getLLTForType(Type &Ty, const DataLayout &DL)
Construct a low-level type based on an LLVM type.
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
Definition BitVector.h:876
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
Definition Alignment.h:77
Pair of physical register and lane mask.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
static bool canHandle(const Instruction *I, const TargetLibraryInfo &TLI)
static StringRef getLibcallImplName(RTLIB::LibcallImpl CallImpl)
Get the libcall routine name for the specified libcall implementation.
This structure is used to communicate between SelectionDAGBuilder and SDISel for the code generation ...
Register Reg
The virtual register containing the index of the jump table entry to jump to.
MachineBasicBlock * Default
The MBB of the default bb, which is a successor of the range check MBB.
unsigned JTI
The JumpTableIndex for this jump table in the function.
MachineBasicBlock * MBB
The MBB into which to emit the code for the indirect jump.