30#define DEBUG_TYPE "x86-pseudo"
31#define X86_EXPAND_PSEUDO_NAME "X86 pseudo instruction expansion pass"
56 MachineFunctionProperties::Property::NoVRegs);
60 return "X86 pseudo instruction expansion pass";
79 void expandVastartSaveXmmRegs(
83char X86ExpandPseudo::ID = 0;
90void X86ExpandPseudo::expandICallBranchFunnel(
99 std::vector<std::pair<MachineBasicBlock *, unsigned>> TargetMBBs;
104 auto CmpTarget = [&](
unsigned Target) {
105 if (Selector.
isReg())
119 auto CreateMBB = [&]() {
130 auto *ElseMBB = CreateMBB();
131 MF->
insert(InsPt, ElseMBB);
136 auto EmitCondJumpTarget = [&](
unsigned CC,
unsigned Target) {
137 auto *ThenMBB = CreateMBB();
138 TargetMBBs.push_back({ThenMBB,
Target});
139 EmitCondJump(
CC, ThenMBB);
142 auto EmitTailCall = [&](
unsigned Target) {
147 std::function<void(
unsigned,
unsigned)> EmitBranchFunnel =
149 if (NumTargets == 1) {
154 if (NumTargets == 2) {
161 if (NumTargets < 6) {
169 auto *ThenMBB = CreateMBB();
173 EmitBranchFunnel(
FirstTarget + (NumTargets / 2) + 1,
174 NumTargets - (NumTargets / 2) - 1);
176 MF->
insert(InsPt, ThenMBB);
183 for (
auto P : TargetMBBs) {
188 JTMBB->
erase(JTInst);
198 assert((
MI.getOperand(1).isGlobal() ||
MI.getOperand(1).isReg()) &&
199 "invalid operand for regular call");
201 if (
MI.getOpcode() == X86::CALL64m_RVMARKER)
203 else if (
MI.getOpcode() == X86::CALL64r_RVMARKER)
205 else if (
MI.getOpcode() == X86::CALL64pcrel32_RVMARKER)
206 Opc = X86::CALL64pcrel32;
211 bool RAXImplicitDead =
false;
215 if (
Op.isReg() &&
Op.isImplicit() &&
Op.isDead() &&
216 TRI->regsOverlap(
Op.getReg(), X86::RAX)) {
219 RAXImplicitDead =
true;
229 auto TargetReg = STI->getTargetTriple().isOSWindows() ? X86::RCX : X86::RDI;
234 if (
MI.shouldUpdateCallSiteInfo())
249 MI.eraseFromParent();
254 if (
TM.getTargetTriple().isOSDarwin())
265 unsigned Opcode =
MI.getOpcode();
267#define GET_EGPR_IF_ENABLED(OPC) (STI->hasEGPR() ? OPC##_EVEX : OPC)
271 case X86::TCRETURNdi:
272 case X86::TCRETURNdicc:
273 case X86::TCRETURNri:
274 case X86::TCRETURNmi:
275 case X86::TCRETURNdi64:
276 case X86::TCRETURNdi64cc:
277 case X86::TCRETURNri64:
278 case X86::TCRETURNmi64: {
279 bool isMem = Opcode == X86::TCRETURNmi || Opcode == X86::TCRETURNmi64;
283 assert(StackAdjust.
isImm() &&
"Expecting immediate value.");
286 int StackAdj = StackAdjust.
getImm();
287 int MaxTCDelta = X86FI->getTCReturnAddrDelta();
289 assert(MaxTCDelta <= 0 &&
"MaxTCDelta should never be positive");
292 Offset = StackAdj - MaxTCDelta;
293 assert(
Offset >= 0 &&
"Offset should never be negative");
295 if (Opcode == X86::TCRETURNdicc || Opcode == X86::TCRETURNdi64cc) {
296 assert(
Offset == 0 &&
"Conditional tail call cannot adjust the stack.");
306 bool IsWin64 = STI->isTargetWin64();
307 if (Opcode == X86::TCRETURNdi || Opcode == X86::TCRETURNdicc ||
308 Opcode == X86::TCRETURNdi64 || Opcode == X86::TCRETURNdi64cc) {
311 case X86::TCRETURNdi:
314 case X86::TCRETURNdicc:
315 Op = X86::TAILJMPd_CC;
317 case X86::TCRETURNdi64cc:
319 "Conditional tail calls confuse "
320 "the Win64 unwinder.");
321 Op = X86::TAILJMPd64_CC;
326 Op = X86::TAILJMPd64;
338 if (
Op == X86::TAILJMPd_CC ||
Op == X86::TAILJMPd64_CC) {
342 }
else if (Opcode == X86::TCRETURNmi || Opcode == X86::TCRETURNmi64) {
343 unsigned Op = (Opcode == X86::TCRETURNmi)
345 : (IsWin64 ? X86::TAILJMPm64_REX : X86::TAILJMPm64);
349 }
else if (Opcode == X86::TCRETURNri64) {
352 TII->get(IsWin64 ? X86::TAILJMPr64_REX : X86::TAILJMPr64))
365 if (
MBBI->isCandidateForCallSiteEntry())
374 case X86::EH_RETURN64: {
376 assert(DestAddr.
isReg() &&
"Offset should be in register!");
377 const bool Uses64BitFramePtr =
378 STI->isTarget64BitLP64() || STI->isTargetNaCl64();
381 TII->get(Uses64BitFramePtr ? X86::MOV64rr : X86::MOV32rr), StackPtr)
388 int64_t StackAdj =
MBBI->getOperand(0).getImm();
389 X86FL->emitSPUpdate(
MBB,
MBBI,
DL, StackAdj,
true);
391 unsigned RetOp = STI->is64Bit() ? X86::IRET64 : X86::IRET32;
393 if (STI->is64Bit() && STI->hasUINTR() &&
402 int64_t StackAdj =
MBBI->getOperand(0).getImm();
406 TII->get(STI->is64Bit() ? X86::RET64 : X86::RET32));
407 }
else if (isUInt<16>(StackAdj)) {
409 TII->get(STI->is64Bit() ? X86::RETI64 : X86::RETI32))
413 "shouldn't need to do this for x86_64 targets!");
417 X86FL->emitSPUpdate(
MBB,
MBBI,
DL, StackAdj,
true);
421 for (
unsigned I = 1, E =
MBBI->getNumOperands();
I != E; ++
I)
426 case X86::LCMPXCHG16B_SAVE_RBX: {
462 case X86::MASKPAIR16LOAD: {
464 assert(Disp >= 0 && Disp <= INT32_MAX - 2 &&
"Unexpected displacement");
466 bool DstIsDead =
MBBI->getOperand(0).isDead();
467 Register Reg0 =
TRI->getSubReg(Reg, X86::sub_mask_0);
468 Register Reg1 =
TRI->getSubReg(Reg, X86::sub_mask_1);
478 MIBLo.
add(
MBBI->getOperand(1 + i));
482 MIBHi.
add(
MBBI->getOperand(1 + i));
491 MIBLo.setMemRefs(MMOLo);
492 MIBHi.setMemRefs(MMOHi);
498 case X86::MASKPAIR16STORE: {
500 assert(Disp >= 0 && Disp <= INT32_MAX - 2 &&
"Unexpected displacement");
503 Register Reg0 =
TRI->getSubReg(Reg, X86::sub_mask_0);
504 Register Reg1 =
TRI->getSubReg(Reg, X86::sub_mask_1);
512 MIBLo.add(
MBBI->getOperand(i));
514 MIBHi.addImm(Disp + 2);
516 MIBHi.add(
MBBI->getOperand(i));
527 MIBLo.setMemRefs(MMOLo);
528 MIBHi.setMemRefs(MMOHi);
534 case X86::MWAITX_SAVE_RBX: {
554 case TargetOpcode::ICALL_BRANCH_FUNNEL:
555 expandICallBranchFunnel(&
MBB,
MBBI);
557 case X86::PLDTILECFGV: {
561 case X86::PTILELOADDV:
562 case X86::PTILELOADDT1V: {
563 for (
unsigned i = 2; i > 0; --i)
565 unsigned Opc = Opcode == X86::PTILELOADDV
568 MI.setDesc(
TII->get(Opc));
571 case X86::PTCMMIMFP16PSV:
572 case X86::PTCMMRLFP16PSV:
577 case X86::PTDPBF16PSV:
578 case X86::PTDPFP16PSV: {
579 MI.untieRegOperand(4);
580 for (
unsigned i = 3; i > 0; --i)
584 case X86::PTCMMIMFP16PSV: Opc = X86::TCMMIMFP16PS;
break;
585 case X86::PTCMMRLFP16PSV: Opc = X86::TCMMRLFP16PS;
break;
586 case X86::PTDPBSSDV: Opc = X86::TDPBSSD;
break;
587 case X86::PTDPBSUDV: Opc = X86::TDPBSUD;
break;
588 case X86::PTDPBUSDV: Opc = X86::TDPBUSD;
break;
589 case X86::PTDPBUUDV: Opc = X86::TDPBUUD;
break;
590 case X86::PTDPBF16PSV: Opc = X86::TDPBF16PS;
break;
591 case X86::PTDPFP16PSV: Opc = X86::TDPFP16PS;
break;
594 MI.setDesc(
TII->get(Opc));
595 MI.tieOperands(0, 1);
598 case X86::PTILESTOREDV: {
599 for (
int i = 1; i >= 0; --i)
604#undef GET_EGPR_IF_ENABLED
605 case X86::PTILEZEROV: {
606 for (
int i = 2; i > 0; --i)
608 MI.setDesc(
TII->get(X86::TILEZERO));
611 case X86::CALL64pcrel32_RVMARKER:
612 case X86::CALL64r_RVMARKER:
613 case X86::CALL64m_RVMARKER:
614 expandCALL_RVMARKER(
MBB,
MBBI);
616 case X86::ADD32mi_ND:
617 case X86::ADD64mi32_ND:
618 case X86::SUB32mi_ND:
619 case X86::SUB64mi32_ND:
620 case X86::AND32mi_ND:
621 case X86::AND64mi32_ND:
623 case X86::OR64mi32_ND:
624 case X86::XOR32mi_ND:
625 case X86::XOR64mi32_ND:
626 case X86::ADC32mi_ND:
627 case X86::ADC64mi32_ND:
628 case X86::SBB32mi_ND:
629 case X86::SBB64mi32_ND: {
647 MI.getOperand(
MI.getNumExplicitOperands() - 1);
662 if (X86MCRegisterClasses[X86::GR32RegClassID].
contains(
Base) ||
667 unsigned Opc, LoadOpc;
669#define MI_TO_RI(OP) \
670 case X86::OP##32mi_ND: \
671 Opc = X86::OP##32ri; \
672 LoadOpc = X86::MOV32rm; \
674 case X86::OP##64mi32_ND: \
675 Opc = X86::OP##64ri32; \
676 LoadOpc = X86::MOV64rm; \
696 for (
unsigned I =
MI.getNumImplicitOperands() + 1;
I != 0; --
I)
697 MI.removeOperand(
MI.getNumOperands() - 1);
698 MI.setDesc(
TII->get(LoadOpc));
719void X86ExpandPseudo::expandVastartSaveXmmRegs(
722 assert(VAStartPseudoInstr->getOpcode() == X86::VASTART_SAVE_XMM_REGS);
726 const DebugLoc &
DL = VAStartPseudoInstr->getDebugLoc();
727 Register CountReg = VAStartPseudoInstr->getOperand(0).getReg();
733 LiveRegs.addLiveIns(*EntryBlk);
735 if (
MI.getOpcode() == VAStartPseudoInstr->getOpcode())
738 LiveRegs.stepForward(
MI, Clobbers);
748 Func->insert(EntryBlkIter, GuardedRegsBlk);
749 Func->insert(EntryBlkIter, TailBlk);
757 uint64_t FrameOffset = VAStartPseudoInstr->getOperand(4).getImm();
758 uint64_t VarArgsRegsOffset = VAStartPseudoInstr->getOperand(6).getImm();
761 unsigned MOVOpc = STI->hasAVX() ? X86::VMOVAPSmr : X86::MOVAPSmr;
764 for (int64_t OpndIdx = 7, RegIdx = 0;
765 OpndIdx < VAStartPseudoInstr->getNumOperands() - 1;
766 OpndIdx++, RegIdx++) {
767 auto NewMI =
BuildMI(GuardedRegsBlk,
DL,
TII->get(MOVOpc));
770 NewMI.addImm(FrameOffset + VarArgsRegsOffset + RegIdx * 16);
772 NewMI.add(VAStartPseudoInstr->getOperand(i + 1));
774 NewMI.addReg(VAStartPseudoInstr->getOperand(OpndIdx).getReg());
775 assert(VAStartPseudoInstr->getOperand(OpndIdx).getReg().isPhysical());
783 if (!STI->isCallingConvWin64(
Func->getFunction().getCallingConv())) {
799 VAStartPseudoInstr->eraseFromParent();
818bool X86ExpandPseudo::expandPseudosWhichAffectControlFlow(
MachineFunction &MF) {
823 if (
Instr.getOpcode() == X86::VASTART_SAVE_XMM_REGS) {
824 expandVastartSaveXmmRegs(&(MF.
front()), Instr);
834 TII = STI->getInstrInfo();
835 TRI = STI->getRegisterInfo();
837 X86FL = STI->getFrameLowering();
839 bool Modified = expandPseudosWhichAffectControlFlow(MF);
848 return new X86ExpandPseudo();
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
MachineBasicBlock MachineBasicBlock::iterator MBBI
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
const HexagonInstrInfo * TII
This file implements the LivePhysRegs utility for tracking liveness of physical registers.
unsigned const TargetRegisterInfo * TRI
#define INITIALIZE_PASS(passName, arg, name, cfg, analysis)
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
static Target * FirstTarget
static bool contains(SmallPtrSetImpl< ConstantExpr * > &Cache, ConstantExpr *Expr, Constant *C)
#define GET_EGPR_IF_ENABLED(OPC)
#define X86_EXPAND_PSEUDO_NAME
Represent the analysis usage information of a pass.
AnalysisUsage & addPreservedID(const void *ID)
void setPreservesCFG()
This function should be called by the pass, iff they do not:
LLVM Basic Block Representation.
This class represents an Operation in the Expression.
FunctionPass class - This class is used to implement most global optimizations.
void copyPhysReg(MachineBasicBlock &MBB, MachineBasicBlock::iterator I, const DebugLoc &DL, MCRegister DestReg, MCRegister SrcReg, bool KillSrc) const override
Emit instructions to copy a pair of physical registers.
A set of physical registers with utility functions to track liveness when walking backward/forward th...
void transferSuccessorsAndUpdatePHIs(MachineBasicBlock *FromMBB)
Transfers all the successors, as in transferSuccessors, and update PHI operands in the successor bloc...
const BasicBlock * getBasicBlock() const
Return the LLVM basic block that this instance corresponded to originally.
bool isLiveIn(MCPhysReg Reg, LaneBitmask LaneMask=LaneBitmask::getAll()) const
Return true if the specified register is in the live in set.
void addSuccessor(MachineBasicBlock *Succ, BranchProbability Prob=BranchProbability::getUnknown())
Add Succ as a successor of this MachineBasicBlock.
void eraseFromParent()
This method unlinks 'this' from the containing function and deletes it.
void addLiveIn(MCRegister PhysReg, LaneBitmask LaneMask=LaneBitmask::getAll())
Adds the specified register as a live in.
const MachineFunction * getParent() const
Return the MachineFunction containing this basic block.
instr_iterator erase(instr_iterator I)
Remove an instruction from the instruction list and delete it.
void splice(iterator Where, MachineBasicBlock *Other, iterator From)
Take an instruction from MBB 'Other' at the position From, and insert it into this MBB right before '...
MachineFunctionPass - This class adapts the FunctionPass interface to allow convenient creation of pa...
void getAnalysisUsage(AnalysisUsage &AU) const override
getAnalysisUsage - Subclasses that override getAnalysisUsage must call this.
virtual bool runOnMachineFunction(MachineFunction &MF)=0
runOnMachineFunction - This method must be overloaded to perform the desired machine code transformat...
virtual MachineFunctionProperties getRequiredProperties() const
Properties which a MachineFunction may have at a given point in time.
MachineFunctionProperties & set(Property P)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
const LLVMTargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
BasicBlockListType::iterator iterator
Ty * getInfo()
getInfo - Keep track of various per-function pieces of information for backends that would like to do...
const MachineBasicBlock & front() const
void moveCallSiteInfo(const MachineInstr *Old, const MachineInstr *New)
Move the call site info from Old to \New call site info.
MachineBasicBlock * CreateMachineBasicBlock(const BasicBlock *BB=nullptr, std::optional< UniqueBBID > BBID=std::nullopt)
CreateMachineBasicBlock - Allocate a new MachineBasicBlock.
void insert(iterator MBBI, MachineBasicBlock *MBB)
const MachineInstrBuilder & addExternalSymbol(const char *FnName, unsigned TargetFlags=0) const
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & add(const MachineOperand &MO) const
const MachineInstrBuilder & addRegMask(const uint32_t *Mask) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addReg(Register RegNo, unsigned flags=0, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
MachineInstr * getInstr() const
If conversion operators fail, use this method to get the MachineInstr explicitly.
Representation of each machine instruction.
void setCFIType(MachineFunction &MF, uint32_t Type)
Set the CFI type for the instruction.
unsigned getNumOperands() const
Retuns the total number of operands.
void addOperand(MachineFunction &MF, const MachineOperand &Op)
Add the specified operand to the instruction.
void copyImplicitOps(MachineFunction &MF, const MachineInstr &MI)
Copy implicit register operands from specified instruction to this instruction.
const DebugLoc & getDebugLoc() const
Returns the debug location id of this MachineInstr.
const MachineOperand & getOperand(unsigned i) const
A description of a memory reference used in the backend.
MachineOperand class - Representation of each machine instruction operand.
const GlobalValue * getGlobal() const
bool isReg() const
isReg - Tests if this is a MO_Register operand.
bool isImm() const
isImm - Tests if this is a MO_Immediate operand.
bool isSymbol() const
isSymbol - Tests if this is a MO_ExternalSymbol operand.
void setIsKill(bool Val=true)
unsigned getTargetFlags() const
bool isGlobal() const
isGlobal - Tests if this is a MO_GlobalAddress operand.
const char * getSymbolName() const
Register getReg() const
getReg - Returns the register number.
int64_t getOffset() const
Return the offset from the symbol in this operand.
virtual StringRef getPassName() const
getPassName - Return a nice clean name for a pass.
Wrapper class representing virtual and physical registers.
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
StringRef - Represent a constant reference to a string, i.e.
TargetInstrInfo - Interface to description of machine instruction set.
CodeModel::Model getCodeModel() const
Returns the code model.
Target - Wrapper for Target specific information.
X86MachineFunctionInfo - This class is derived from MachineFunction and contains private X86 target-s...
self_iterator getIterator()
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ C
The default llvm calling convention, compatible with C.
@ Implicit
Not emitted register (e.g. carry, or temporary result).
@ Define
Register definition.
@ X86
Windows x64, Windows Itanium (IA-64)
Reg
All possible values of the reg field in the ModR/M byte.
bool needSIB(unsigned BaseReg, unsigned IndexReg, bool In64BitMode)
int getFirstAddrOperandIdx(const MachineInstr &MI)
Return the index of the instruction's first address operand, if it has a memory reference,...
NodeAddr< InstrNode * > Instr
NodeAddr< FuncNode * > Func
This is an optimization pass for GlobalISel generic memory operations.
auto drop_begin(T &&RangeOrContainer, size_t N=1)
Return a range covering RangeOrContainer with the first N elements excluded.
void finalizeBundle(MachineBasicBlock &MBB, MachineBasicBlock::instr_iterator FirstMI, MachineBasicBlock::instr_iterator LastMI)
finalizeBundle - Finalize a machine instruction bundle which includes a sequence of instructions star...
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
static bool isMem(const MachineInstr &MI, unsigned Op)
char & MachineDominatorsID
MachineDominators - This pass is a machine dominators analysis pass.
unsigned getDeadRegState(bool B)
char & MachineLoopInfoID
MachineLoopInfo - This pass is a loop analysis pass.
FunctionPass * createX86ExpandPseudoPass()
Return a Machine IR pass that expands X86-specific pseudo instructions into a sequence of actual inst...
unsigned getKillRegState(bool B)
DWARFExpression::Operation Op
void addLiveIns(MachineBasicBlock &MBB, const LivePhysRegs &LiveRegs)
Adds registers contained in LiveRegs to the block live-in list of MBB.