36#include "llvm/IR/IntrinsicsX86.h"
47class X86FastISel final :
public FastISel {
50 const X86Subtarget *Subtarget;
53 explicit X86FastISel(FunctionLoweringInfo &funcInfo,
54 const TargetLibraryInfo *libInfo,
55 const LibcallLoweringInfo *libcallLowering)
56 : FastISel(funcInfo, libInfo, libcallLowering) {
60 bool fastSelectInstruction(
const Instruction *
I)
override;
66 bool tryToFoldLoadIntoMI(MachineInstr *
MI,
unsigned OpNo,
67 const LoadInst *LI)
override;
69 bool fastLowerArguments()
override;
70 bool fastLowerCall(CallLoweringInfo &CLI)
override;
71 bool fastLowerIntrinsicCall(
const IntrinsicInst *
II)
override;
73#include "X86GenFastISel.inc"
79 bool X86FastEmitLoad(MVT VT, X86AddressMode &AM, MachineMemOperand *MMO,
80 Register &ResultReg,
unsigned Alignment = 1);
82 bool X86FastEmitStore(EVT VT,
const Value *Val, X86AddressMode &AM,
83 MachineMemOperand *MMO =
nullptr,
bool Aligned =
false);
84 bool X86FastEmitStore(EVT VT,
Register ValReg, X86AddressMode &AM,
85 MachineMemOperand *MMO =
nullptr,
bool Aligned =
false);
91 bool X86SelectCallAddress(
const Value *V, X86AddressMode &AM);
93 bool X86SelectLoad(
const Instruction *
I);
95 bool X86SelectStore(
const Instruction *
I);
97 bool X86SelectRet(
const Instruction *
I);
99 bool X86SelectCmp(
const Instruction *
I);
101 bool X86SelectZExt(
const Instruction *
I);
103 bool X86SelectSExt(
const Instruction *
I);
105 bool X86SelectBranch(
const Instruction *
I);
107 bool X86SelectShift(
const Instruction *
I);
109 bool X86SelectDivRem(
const Instruction *
I);
111 bool X86FastEmitCMoveSelect(MVT RetVT,
const Instruction *
I);
113 bool X86FastEmitSSESelect(MVT RetVT,
const Instruction *
I);
115 bool X86FastEmitPseudoSelect(MVT RetVT,
const Instruction *
I);
117 bool X86SelectSelect(
const Instruction *
I);
119 bool X86SelectTrunc(
const Instruction *
I);
121 bool X86SelectFPExtOrFPTrunc(
const Instruction *
I,
unsigned Opc,
122 const TargetRegisterClass *RC);
124 bool X86SelectFPExt(
const Instruction *
I);
125 bool X86SelectFPTrunc(
const Instruction *
I);
126 bool X86SelectSIToFP(
const Instruction *
I);
127 bool X86SelectUIToFP(
const Instruction *
I);
128 bool X86SelectIntToFP(
const Instruction *
I,
bool IsSigned);
129 bool X86SelectBitCast(
const Instruction *
I);
131 const X86InstrInfo *getInstrInfo()
const {
132 return Subtarget->getInstrInfo();
134 const X86TargetMachine *getTargetMachine()
const {
135 return static_cast<const X86TargetMachine *
>(&TM);
138 bool handleConstantAddresses(
const Value *V, X86AddressMode &AM);
140 Register X86MaterializeInt(
const ConstantInt *CI, MVT VT);
141 Register X86MaterializeFP(
const ConstantFP *CFP, MVT VT);
142 Register X86MaterializeGV(
const GlobalValue *GV, MVT VT);
143 Register fastMaterializeConstant(
const Constant *
C)
override;
145 Register fastMaterializeAlloca(
const AllocaInst *
C)
override;
147 Register fastMaterializeFloatZero(
const ConstantFP *CF)
override;
151 bool isScalarFPTypeInSSEReg(EVT VT)
const {
152 return (VT == MVT::f64 && Subtarget->hasSSE2()) ||
153 (VT == MVT::f32 && Subtarget->hasSSE1()) || VT == MVT::f16;
156 bool isTypeLegal(
Type *Ty, MVT &VT,
bool AllowI1 =
false);
158 bool IsMemcpySmall(uint64_t Len);
160 bool TryEmitSmallMemcpy(X86AddressMode DestAM,
161 X86AddressMode SrcAM, uint64_t Len);
163 bool foldX86XALUIntrinsic(
X86::CondCode &CC,
const Instruction *
I,
166 const MachineInstrBuilder &
addFullAddress(
const MachineInstrBuilder &MIB,
169 Register fastEmitInst_rrrr(
unsigned MachineInstOpcode,
170 const TargetRegisterClass *RC,
Register Op0,
176static std::pair<unsigned, bool>
179 bool NeedSwap =
false;
208 return std::make_pair(CC, NeedSwap);
222 return ::addFullAddress(MIB, AM);
227bool X86FastISel::foldX86XALUIntrinsic(
X86::CondCode &CC,
const Instruction *
I,
241 if (!isTypeLegal(RetTy, RetVT))
244 if (RetVT != MVT::i32 && RetVT != MVT::i64)
248 switch (
II->getIntrinsicID()) {
249 default:
return false;
250 case Intrinsic::sadd_with_overflow:
251 case Intrinsic::ssub_with_overflow:
252 case Intrinsic::smul_with_overflow:
253 case Intrinsic::umul_with_overflow: TmpCC =
X86::COND_O;
break;
254 case Intrinsic::uadd_with_overflow:
255 case Intrinsic::usub_with_overflow: TmpCC =
X86::COND_B;
break;
259 if (
II->getParent() !=
I->getParent())
265 for (
auto Itr = std::prev(Start); Itr != End; --Itr) {
273 if (EVI->getAggregateOperand() !=
II)
279 auto HasPhis = [](
const BasicBlock *Succ) {
return !Succ->phis().empty(); };
292bool X86FastISel::isTypeLegal(
Type *Ty, MVT &VT,
bool AllowI1) {
293 EVT evt = TLI.getValueType(
DL, Ty,
true);
294 if (evt == MVT::Other || !evt.
isSimple())
301 if (VT == MVT::f64 && !Subtarget->hasSSE2())
303 if (VT == MVT::f32 && !Subtarget->hasSSE1())
312 return (AllowI1 && VT == MVT::i1) || TLI.isTypeLegal(VT);
318bool X86FastISel::X86FastEmitLoad(MVT VT, X86AddressMode &AM,
319 MachineMemOperand *MMO,
Register &ResultReg,
320 unsigned Alignment) {
321 bool HasSSE1 = Subtarget->hasSSE1();
322 bool HasSSE2 = Subtarget->hasSSE2();
323 bool HasSSE41 = Subtarget->hasSSE41();
324 bool HasAVX = Subtarget->hasAVX();
325 bool HasAVX2 = Subtarget->hasAVX2();
326 bool HasAVX512 = Subtarget->hasAVX512();
327 bool HasVLX = Subtarget->hasVLX();
337 default:
return false;
352 Opc = HasAVX512 ? X86::VMOVSSZrm_alt
353 : HasAVX ? X86::VMOVSSrm_alt
354 : HasSSE1 ? X86::MOVSSrm_alt
358 Opc = HasAVX512 ? X86::VMOVSDZrm_alt
359 : HasAVX ? X86::VMOVSDrm_alt
360 : HasSSE2 ? X86::MOVSDrm_alt
367 if (IsNonTemporal && Alignment >= 16 && HasSSE41)
368 Opc = HasVLX ? X86::VMOVNTDQAZ128rm :
369 HasAVX ? X86::VMOVNTDQArm : X86::MOVNTDQArm;
370 else if (Alignment >= 16)
371 Opc = HasVLX ? X86::VMOVAPSZ128rm :
372 HasAVX ? X86::VMOVAPSrm : X86::MOVAPSrm;
374 Opc = HasVLX ? X86::VMOVUPSZ128rm :
375 HasAVX ? X86::VMOVUPSrm : X86::MOVUPSrm;
378 if (IsNonTemporal && Alignment >= 16 && HasSSE41)
379 Opc = HasVLX ? X86::VMOVNTDQAZ128rm :
380 HasAVX ? X86::VMOVNTDQArm : X86::MOVNTDQArm;
381 else if (Alignment >= 16)
382 Opc = HasVLX ? X86::VMOVAPDZ128rm :
383 HasAVX ? X86::VMOVAPDrm : X86::MOVAPDrm;
385 Opc = HasVLX ? X86::VMOVUPDZ128rm :
386 HasAVX ? X86::VMOVUPDrm : X86::MOVUPDrm;
392 if (IsNonTemporal && Alignment >= 16 && HasSSE41)
393 Opc = HasVLX ? X86::VMOVNTDQAZ128rm :
394 HasAVX ? X86::VMOVNTDQArm : X86::MOVNTDQArm;
395 else if (Alignment >= 16)
396 Opc = HasVLX ? X86::VMOVDQA64Z128rm :
397 HasAVX ? X86::VMOVDQArm : X86::MOVDQArm;
399 Opc = HasVLX ? X86::VMOVDQU64Z128rm :
400 HasAVX ? X86::VMOVDQUrm : X86::MOVDQUrm;
404 if (IsNonTemporal && Alignment >= 32 && HasAVX2)
405 Opc = HasVLX ? X86::VMOVNTDQAZ256rm : X86::VMOVNTDQAYrm;
406 else if (IsNonTemporal && Alignment >= 16)
408 else if (Alignment >= 32)
409 Opc = HasVLX ? X86::VMOVAPSZ256rm : X86::VMOVAPSYrm;
411 Opc = HasVLX ? X86::VMOVUPSZ256rm : X86::VMOVUPSYrm;
415 if (IsNonTemporal && Alignment >= 32 && HasAVX2)
416 Opc = HasVLX ? X86::VMOVNTDQAZ256rm : X86::VMOVNTDQAYrm;
417 else if (IsNonTemporal && Alignment >= 16)
419 else if (Alignment >= 32)
420 Opc = HasVLX ? X86::VMOVAPDZ256rm : X86::VMOVAPDYrm;
422 Opc = HasVLX ? X86::VMOVUPDZ256rm : X86::VMOVUPDYrm;
429 if (IsNonTemporal && Alignment >= 32 && HasAVX2)
430 Opc = HasVLX ? X86::VMOVNTDQAZ256rm : X86::VMOVNTDQAYrm;
431 else if (IsNonTemporal && Alignment >= 16)
433 else if (Alignment >= 32)
434 Opc = HasVLX ? X86::VMOVDQA64Z256rm : X86::VMOVDQAYrm;
436 Opc = HasVLX ? X86::VMOVDQU64Z256rm : X86::VMOVDQUYrm;
440 if (IsNonTemporal && Alignment >= 64)
441 Opc = X86::VMOVNTDQAZrm;
443 Opc = (Alignment >= 64) ? X86::VMOVAPSZrm :
X86::VMOVUPSZrm;
447 if (IsNonTemporal && Alignment >= 64)
448 Opc = X86::VMOVNTDQAZrm;
450 Opc = (Alignment >= 64) ? X86::VMOVAPDZrm :
X86::VMOVUPDZrm;
459 if (IsNonTemporal && Alignment >= 64)
460 Opc = X86::VMOVNTDQAZrm;
462 Opc = (Alignment >= 64) ? X86::VMOVDQA64Zrm :
X86::VMOVDQU64Zrm;
466 const TargetRegisterClass *RC = TLI.getRegClassFor(VT);
468 ResultReg = createResultReg(RC);
469 MachineInstrBuilder MIB =
470 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(
Opc), ResultReg);
481bool X86FastISel::X86FastEmitStore(EVT VT,
Register ValReg, X86AddressMode &AM,
482 MachineMemOperand *MMO,
bool Aligned) {
483 bool HasSSE1 = Subtarget->hasSSE1();
484 bool HasSSE2 = Subtarget->hasSSE2();
485 bool HasSSE4A = Subtarget->hasSSE4A();
486 bool HasAVX = Subtarget->hasAVX();
487 bool HasAVX512 = Subtarget->hasAVX512();
488 bool HasVLX = Subtarget->hasVLX();
495 default:
return false;
498 Register AndResult = createResultReg(&X86::GR8RegClass);
499 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
500 TII.get(X86::AND8ri), AndResult)
505 case MVT::i8:
Opc = X86::MOV8mr;
break;
506 case MVT::i16:
Opc = X86::MOV16mr;
break;
508 Opc = (IsNonTemporal && HasSSE2) ? X86::MOVNTImr :
X86::MOV32mr;
512 Opc = (IsNonTemporal && HasSSE2) ? X86::MOVNTI_64mr :
X86::MOV64mr;
516 if (IsNonTemporal && HasSSE4A)
519 Opc = HasAVX512 ? X86::VMOVSSZmr :
520 HasAVX ? X86::VMOVSSmr : X86::MOVSSmr;
526 if (IsNonTemporal && HasSSE4A)
529 Opc = HasAVX512 ? X86::VMOVSDZmr :
530 HasAVX ? X86::VMOVSDmr : X86::MOVSDmr;
535 Opc = (IsNonTemporal && HasSSE1) ? X86::MMX_MOVNTQmr :
X86::MMX_MOVQ64mr;
540 Opc = HasVLX ? X86::VMOVNTPSZ128mr :
541 HasAVX ? X86::VMOVNTPSmr : X86::MOVNTPSmr;
543 Opc = HasVLX ? X86::VMOVAPSZ128mr :
544 HasAVX ? X86::VMOVAPSmr : X86::MOVAPSmr;
546 Opc = HasVLX ? X86::VMOVUPSZ128mr :
547 HasAVX ? X86::VMOVUPSmr : X86::MOVUPSmr;
552 Opc = HasVLX ? X86::VMOVNTPDZ128mr :
553 HasAVX ? X86::VMOVNTPDmr : X86::MOVNTPDmr;
555 Opc = HasVLX ? X86::VMOVAPDZ128mr :
556 HasAVX ? X86::VMOVAPDmr : X86::MOVAPDmr;
558 Opc = HasVLX ? X86::VMOVUPDZ128mr :
559 HasAVX ? X86::VMOVUPDmr : X86::MOVUPDmr;
567 Opc = HasVLX ? X86::VMOVNTDQZ128mr :
568 HasAVX ? X86::VMOVNTDQmr : X86::MOVNTDQmr;
570 Opc = HasVLX ? X86::VMOVDQA64Z128mr :
571 HasAVX ? X86::VMOVDQAmr : X86::MOVDQAmr;
573 Opc = HasVLX ? X86::VMOVDQU64Z128mr :
574 HasAVX ? X86::VMOVDQUmr : X86::MOVDQUmr;
580 Opc = HasVLX ? X86::VMOVNTPSZ256mr : X86::VMOVNTPSYmr;
582 Opc = HasVLX ? X86::VMOVAPSZ256mr : X86::VMOVAPSYmr;
584 Opc = HasVLX ? X86::VMOVUPSZ256mr : X86::VMOVUPSYmr;
590 Opc = HasVLX ? X86::VMOVNTPDZ256mr : X86::VMOVNTPDYmr;
592 Opc = HasVLX ? X86::VMOVAPDZ256mr : X86::VMOVAPDYmr;
594 Opc = HasVLX ? X86::VMOVUPDZ256mr : X86::VMOVUPDYmr;
603 Opc = HasVLX ? X86::VMOVNTDQZ256mr : X86::VMOVNTDQYmr;
605 Opc = HasVLX ? X86::VMOVDQA64Z256mr : X86::VMOVDQAYmr;
607 Opc = HasVLX ? X86::VMOVDQU64Z256mr : X86::VMOVDQUYmr;
612 Opc = IsNonTemporal ? X86::VMOVNTPSZmr : X86::VMOVAPSZmr;
614 Opc = X86::VMOVUPSZmr;
619 Opc = IsNonTemporal ? X86::VMOVNTPDZmr : X86::VMOVAPDZmr;
621 Opc = X86::VMOVUPDZmr;
631 Opc = IsNonTemporal ? X86::VMOVNTDQZmr : X86::VMOVDQA64Zmr;
633 Opc = X86::VMOVDQU64Zmr;
645 MachineInstrBuilder MIB =
646 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
Desc);
654bool X86FastISel::X86FastEmitStore(EVT VT,
const Value *Val,
656 MachineMemOperand *MMO,
bool Aligned) {
670 case MVT::i8:
Opc = X86::MOV8mi;
break;
671 case MVT::i16:
Opc = X86::MOV16mi;
break;
672 case MVT::i32:
Opc = X86::MOV32mi;
break;
676 Opc = X86::MOV64mi32;
681 MachineInstrBuilder MIB =
682 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(
Opc));
684 : CI->getZExtValue());
691 Register ValReg = getRegForValue(Val);
695 return X86FastEmitStore(VT, ValReg, AM, MMO,
Aligned);
711bool X86FastISel::handleConstantAddresses(
const Value *V, X86AddressMode &AM) {
720 if (TM.isLargeGlobalValue(GV))
724 if (GV->isThreadLocal())
728 if (GV->isAbsoluteSymbolRef())
734 if (!Subtarget->isPICStyleRIPRel() ||
740 unsigned char GVFlags = Subtarget->classifyGlobalReference(GV);
745 AM.
Base.
Reg = getInstrInfo()->getGlobalBaseReg(FuncInfo.MF);
751 if (Subtarget->isPICStyleRIPRel()) {
763 DenseMap<const Value *, Register>::iterator
I = LocalValueMap.find(V);
765 if (
I != LocalValueMap.end() &&
I->second) {
770 const TargetRegisterClass *RC =
nullptr;
771 X86AddressMode StubAM;
777 SavePoint SaveInsertPt = enterLocalValueArea();
779 if (TLI.getPointerTy(
DL) == MVT::i64) {
781 RC = &X86::GR64RegClass;
784 RC = &X86::GR32RegClass;
791 LoadReg = createResultReg(RC);
792 MachineInstrBuilder LoadMI =
793 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(
Opc), LoadReg);
797 leaveLocalValueArea(SaveInsertPt);
800 LocalValueMap[
V] = LoadReg;
812 if (!AM.
GV || !Subtarget->isPICStyleRIPRel()) {
814 AM.
Base.
Reg = getRegForValue(V);
829bool X86FastISel::X86SelectAddress(
const Value *V, X86AddressMode &AM) {
832 const User *
U =
nullptr;
833 unsigned Opcode = Instruction::UserOp1;
838 if (FuncInfo.StaticAllocaMap.count(
static_cast<const AllocaInst *
>(V)) ||
839 FuncInfo.getMBB(
I->getParent()) == FuncInfo.MBB) {
840 Opcode =
I->getOpcode();
844 Opcode =
C->getOpcode();
849 if (Ty->getAddressSpace() > 255)
856 case Instruction::BitCast:
860 case Instruction::IntToPtr:
862 if (TLI.getValueType(
DL,
U->getOperand(0)->getType()) ==
863 TLI.getPointerTy(
DL))
867 case Instruction::PtrToInt:
869 if (TLI.getValueType(
DL,
U->getType()) == TLI.getPointerTy(
DL))
873 case Instruction::Alloca: {
876 DenseMap<const AllocaInst *, int>::iterator
SI =
877 FuncInfo.StaticAllocaMap.find(
A);
878 if (SI != FuncInfo.StaticAllocaMap.end()) {
886 case Instruction::Add: {
889 uint64_t Disp = (int32_t)AM.
Disp + (uint64_t)CI->getSExtValue();
892 AM.
Disp = (uint32_t)Disp;
899 case Instruction::GetElementPtr: {
900 X86AddressMode SavedAM = AM;
903 uint64_t Disp = (int32_t)AM.
Disp;
905 unsigned Scale = AM.
Scale;
906 MVT PtrVT = TLI.getValueType(
DL,
U->getType()).getSimpleVT();
912 i != e; ++i, ++GTI) {
915 const StructLayout *SL =
DL.getStructLayout(STy);
926 Disp += CI->getSExtValue() * S;
929 if (canFoldAddIntoGEP(U,
Op)) {
938 if (!IndexReg && (!AM.
GV || !Subtarget->isPICStyleRIPRel()) &&
939 (S == 1 || S == 2 || S == 4 || S == 8)) {
942 IndexReg = getRegForGEPIndex(PtrVT,
Op);
948 goto unsupported_gep;
958 AM.
Disp = (uint32_t)Disp;
961 if (
const GetElementPtrInst *
GEP =
976 if (handleConstantAddresses(
I, AM))
986 return handleConstantAddresses(V, AM);
991bool X86FastISel::X86SelectCallAddress(
const Value *V, X86AddressMode &AM) {
992 const User *
U =
nullptr;
993 unsigned Opcode = Instruction::UserOp1;
1020 Opcode =
I->getOpcode();
1022 InMBB =
I->getParent() == FuncInfo.MBB->getBasicBlock();
1024 Opcode =
C->getOpcode();
1030 case Instruction::BitCast:
1033 return X86SelectCallAddress(
U->getOperand(0), AM);
1036 case Instruction::IntToPtr:
1039 TLI.getValueType(
DL,
U->getOperand(0)->getType()) ==
1040 TLI.getPointerTy(
DL))
1041 return X86SelectCallAddress(
U->getOperand(0), AM);
1044 case Instruction::PtrToInt:
1046 if (InMBB && TLI.getValueType(
DL,
U->getType()) == TLI.getPointerTy(
DL))
1047 return X86SelectCallAddress(
U->getOperand(0), AM);
1059 if (Subtarget->isPICStyleRIPRel() &&
1065 if (GVar->isThreadLocal())
1074 if (Subtarget->isPICStyleRIPRel()) {
1080 AM.
GVOpFlags = Subtarget->classifyLocalReference(
nullptr);
1087 if (!AM.
GV || !Subtarget->isPICStyleRIPRel()) {
1088 auto GetCallRegForValue = [
this](
const Value *
V) {
1092 if (
Reg && Subtarget->isTarget64BitILP32()) {
1093 Register CopyReg = createResultReg(&X86::GR32RegClass);
1094 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOV32rr),
1098 Register ExtReg = createResultReg(&X86::GR64RegClass);
1099 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1100 TII.get(TargetOpcode::SUBREG_TO_REG), ExtReg)
1110 AM.
Base.
Reg = GetCallRegForValue(V);
1115 AM.
IndexReg = GetCallRegForValue(V);
1125bool X86FastISel::X86SelectStore(
const Instruction *
I) {
1132 const Value *PtrV =
I->getOperand(1);
1133 if (TLI.supportSwiftError()) {
1137 if (Arg->hasSwiftErrorAttr())
1142 if (Alloca->isSwiftError())
1151 if (!isTypeLegal(Val->
getType(), VT,
true))
1156 bool Aligned = Alignment >= ABIAlignment;
1162 return X86FastEmitStore(VT, Val, AM, createMachineMemOperandFor(
I),
Aligned);
1166bool X86FastISel::X86SelectRet(
const Instruction *
I) {
1168 const Function &
F = *
I->getParent()->getParent();
1169 const X86MachineFunctionInfo *X86MFInfo =
1170 FuncInfo.MF->getInfo<X86MachineFunctionInfo>();
1172 if (!FuncInfo.CanLowerReturn)
1175 if (TLI.supportSwiftError() &&
1176 F.getAttributes().hasAttrSomewhere(Attribute::SwiftError))
1179 if (TLI.supportSplitCSR(FuncInfo.MF))
1182 CallingConv::ID CC =
F.getCallingConv();
1183 if (CC != CallingConv::C &&
1184 CC != CallingConv::Fast &&
1185 CC != CallingConv::Tail &&
1186 CC != CallingConv::SwiftTail &&
1187 CC != CallingConv::X86_FastCall &&
1188 CC != CallingConv::X86_StdCall &&
1189 CC != CallingConv::X86_ThisCall &&
1190 CC != CallingConv::X86_64_SysV &&
1191 CC != CallingConv::Win64)
1200 if ((CC == CallingConv::Fast && TM.Options.GuaranteedTailCallOpt) ||
1201 CC == CallingConv::Tail || CC == CallingConv::SwiftTail)
1217 CCState CCInfo(CC,
F.isVarArg(), *FuncInfo.MF, ValLocs,
I->getContext());
1226 if (ValLocs.
size() != 1)
1229 CCValAssign &VA = ValLocs[0];
1244 EVT SrcVT = TLI.getValueType(
DL, RV->
getType());
1247 if (SrcVT != DstVT) {
1248 if (SrcVT != MVT::i1 && SrcVT != MVT::i8 && SrcVT != MVT::i16)
1251 if (!Outs[0].
Flags.isZExt() && !Outs[0].Flags.isSExt())
1254 if (SrcVT == MVT::i1) {
1255 if (Outs[0].
Flags.isSExt())
1257 SrcReg = fastEmitZExtFromI1(MVT::i8, SrcReg);
1260 if (SrcVT != DstVT) {
1270 const TargetRegisterClass *SrcRC =
MRI.getRegClass(SrcReg);
1274 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1275 TII.get(TargetOpcode::COPY), DstReg).
addReg(SrcReg);
1288 if (
F.hasStructRetAttr() && CC != CallingConv::Swift &&
1289 CC != CallingConv::SwiftTail) {
1292 "SRetReturnReg should have been set in LowerFormalArguments()!");
1293 Register RetReg = Subtarget->isTarget64BitLP64() ? X86::RAX : X86::EAX;
1294 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1300 MachineInstrBuilder MIB;
1302 MIB =
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1303 TII.get(Subtarget->is64Bit() ? X86::RETI64 : X86::RETI32))
1306 MIB =
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1307 TII.get(Subtarget->is64Bit() ? X86::RET64 : X86::RET32));
1316bool X86FastISel::X86SelectLoad(
const Instruction *
I) {
1323 const Value *SV =
I->getOperand(0);
1324 if (TLI.supportSwiftError()) {
1328 if (Arg->hasSwiftErrorAttr())
1333 if (Alloca->isSwiftError())
1339 if (!isTypeLegal(LI->
getType(), VT,
true))
1349 if (!X86FastEmitLoad(VT, AM, createMachineMemOperandFor(LI), ResultReg,
1353 updateValueMap(
I, ResultReg);
1358 bool HasAVX512 = Subtarget->
hasAVX512();
1359 bool HasAVX = Subtarget->
hasAVX();
1360 bool HasSSE1 = Subtarget->
hasSSE1();
1361 bool HasSSE2 = Subtarget->
hasSSE2();
1365 case MVT::i8:
return X86::CMP8rr;
1366 case MVT::i16:
return X86::CMP16rr;
1367 case MVT::i32:
return X86::CMP32rr;
1368 case MVT::i64:
return X86::CMP64rr;
1370 return HasAVX512 ? X86::VUCOMISSZrr
1371 : HasAVX ? X86::VUCOMISSrr
1372 : HasSSE1 ? X86::UCOMISSrr
1375 return HasAVX512 ? X86::VUCOMISDZrr
1376 : HasAVX ? X86::VUCOMISDrr
1377 : HasSSE2 ? X86::UCOMISDrr
1392 return X86::CMP16ri;
1394 return X86::CMP32ri;
1402bool X86FastISel::X86FastEmitCompare(
const Value *Op0,
const Value *Op1, EVT VT,
1404 Register Op0Reg = getRegForValue(Op0);
1417 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, CurMIMD,
TII.get(CompareImmOpc))
1419 .
addImm(Op1C->getSExtValue());
1425 if (CompareOpc == 0)
return false;
1427 Register Op1Reg = getRegForValue(Op1);
1430 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, CurMIMD,
TII.get(CompareOpc))
1438 ((!Subtarget->hasZU() || Subtarget->preferLegacySetCC()) ? X86::SETCCr \
1445 if (!isTypeLegal(
I->getOperand(0)->getType(), VT))
1458 ResultReg = createResultReg(&X86::GR32RegClass);
1459 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOV32r0),
1461 ResultReg = fastEmitInst_extractsubreg(MVT::i8, ResultReg, X86::sub_8bit);
1467 ResultReg = createResultReg(&X86::GR8RegClass);
1468 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOV8ri),
1475 updateValueMap(
I, ResultReg);
1487 if (RHSC && RHSC->isNullValue())
1492 static const uint16_t SETFOpcTable[2][3] = {
1503 ResultReg = createResultReg(&X86::GR8RegClass);
1505 if (!X86FastEmitCompare(
LHS,
RHS, VT,
I->getDebugLoc()))
1508 Register FlagReg1 = createResultReg(&X86::GR8RegClass);
1509 Register FlagReg2 = createResultReg(&X86::GR8RegClass);
1516 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(SETFOpc[2]),
1518 updateValueMap(
I, ResultReg);
1531 if (!X86FastEmitCompare(
LHS,
RHS, VT,
I->getDebugLoc()))
1536 updateValueMap(
I, ResultReg);
1540bool X86FastISel::X86SelectZExt(
const Instruction *
I) {
1541 EVT DstVT = TLI.getValueType(
DL,
I->getType());
1542 if (!TLI.isTypeLegal(DstVT))
1545 Register ResultReg = getRegForValue(
I->getOperand(0));
1550 MVT SrcVT = TLI.getSimpleValueType(
DL,
I->getOperand(0)->getType());
1551 if (SrcVT == MVT::i1) {
1553 ResultReg = fastEmitZExtFromI1(MVT::i8, ResultReg);
1560 if (DstVT == MVT::i64) {
1565 case MVT::i8: MovInst = X86::MOVZX32rr8;
break;
1566 case MVT::i16: MovInst = X86::MOVZX32rr16;
break;
1567 case MVT::i32: MovInst = X86::MOV32rr;
break;
1571 Register Result32 = createResultReg(&X86::GR32RegClass);
1572 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(MovInst), Result32)
1575 ResultReg = createResultReg(&X86::GR64RegClass);
1576 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1577 TII.get(TargetOpcode::SUBREG_TO_REG), ResultReg)
1580 }
else if (DstVT == MVT::i16) {
1583 Register Result32 = createResultReg(&X86::GR32RegClass);
1584 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOVZX32rr8),
1585 Result32).
addReg(ResultReg);
1587 ResultReg = fastEmitInst_extractsubreg(MVT::i16, Result32, X86::sub_16bit);
1588 }
else if (DstVT != MVT::i8) {
1595 updateValueMap(
I, ResultReg);
1599bool X86FastISel::X86SelectSExt(
const Instruction *
I) {
1600 EVT DstVT = TLI.getValueType(
DL,
I->getType());
1601 if (!TLI.isTypeLegal(DstVT))
1604 Register ResultReg = getRegForValue(
I->getOperand(0));
1609 MVT SrcVT = TLI.getSimpleValueType(
DL,
I->getOperand(0)->getType());
1610 if (SrcVT == MVT::i1) {
1612 Register ZExtReg = fastEmitZExtFromI1(MVT::i8, ResultReg);
1617 ResultReg = createResultReg(&X86::GR8RegClass);
1618 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::NEG8r),
1619 ResultReg).
addReg(ZExtReg);
1624 if (DstVT == MVT::i16) {
1627 Register Result32 = createResultReg(&X86::GR32RegClass);
1628 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOVSX32rr8),
1629 Result32).
addReg(ResultReg);
1631 ResultReg = fastEmitInst_extractsubreg(MVT::i16, Result32, X86::sub_16bit);
1632 }
else if (DstVT != MVT::i8) {
1639 updateValueMap(
I, ResultReg);
1643bool X86FastISel::X86SelectBranch(
const Instruction *
I) {
1647 MachineBasicBlock *TrueMBB = FuncInfo.getMBB(BI->
getSuccessor(0));
1648 MachineBasicBlock *FalseMBB = FuncInfo.getMBB(BI->
getSuccessor(1));
1655 if (CI->
hasOneUse() && CI->getParent() ==
I->getParent()) {
1660 switch (Predicate) {
1675 if (CmpRHSC && CmpRHSC->isNullValue())
1680 if (FuncInfo.MBB->isLayoutSuccessor(TrueMBB)) {
1690 bool NeedExtraBranch =
false;
1691 switch (Predicate) {
1697 NeedExtraBranch =
true;
1710 if (!X86FastEmitCompare(CmpLHS, CmpRHS, VT, CI->getDebugLoc()))
1713 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::JCC_1))
1718 if (NeedExtraBranch) {
1719 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::JCC_1))
1723 finishCondBranch(BI->
getParent(), TrueMBB, FalseMBB);
1730 if (TI->hasOneUse() && TI->getParent() ==
I->getParent() &&
1731 isTypeLegal(TI->getOperand(0)->getType(), SourceVT)) {
1732 unsigned TestOpc = 0;
1735 case MVT::i8: TestOpc = X86::TEST8ri;
break;
1736 case MVT::i16: TestOpc = X86::TEST16ri;
break;
1737 case MVT::i32: TestOpc = X86::TEST32ri;
break;
1738 case MVT::i64: TestOpc = X86::TEST64ri32;
break;
1741 Register OpReg = getRegForValue(TI->getOperand(0));
1745 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(TestOpc))
1749 if (FuncInfo.MBB->isLayoutSuccessor(TrueMBB)) {
1754 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::JCC_1))
1757 finishCondBranch(BI->
getParent(), TrueMBB, FalseMBB);
1761 }
else if (foldX86XALUIntrinsic(CC, BI, BI->
getCondition())) {
1768 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::JCC_1))
1770 finishCondBranch(BI->
getParent(), TrueMBB, FalseMBB);
1782 if (
MRI.getRegClass(OpReg) == &X86::VK1RegClass) {
1784 OpReg = createResultReg(&X86::GR32RegClass);
1785 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1786 TII.get(TargetOpcode::COPY), OpReg)
1788 OpReg = fastEmitInst_extractsubreg(MVT::i8, OpReg, X86::sub_8bit);
1790 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::TEST8ri))
1793 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::JCC_1))
1795 finishCondBranch(BI->
getParent(), TrueMBB, FalseMBB);
1799bool X86FastISel::X86SelectShift(
const Instruction *
I) {
1802 const TargetRegisterClass *RC =
nullptr;
1803 if (
I->getType()->isIntegerTy(8)) {
1805 RC = &X86::GR8RegClass;
1806 switch (
I->getOpcode()) {
1807 case Instruction::LShr: OpReg = X86::SHR8rCL;
break;
1808 case Instruction::AShr: OpReg = X86::SAR8rCL;
break;
1809 case Instruction::Shl: OpReg = X86::SHL8rCL;
break;
1810 default:
return false;
1812 }
else if (
I->getType()->isIntegerTy(16)) {
1814 RC = &X86::GR16RegClass;
1815 switch (
I->getOpcode()) {
1817 case Instruction::LShr: OpReg = X86::SHR16rCL;
break;
1818 case Instruction::AShr: OpReg = X86::SAR16rCL;
break;
1819 case Instruction::Shl: OpReg = X86::SHL16rCL;
break;
1821 }
else if (
I->getType()->isIntegerTy(32)) {
1823 RC = &X86::GR32RegClass;
1824 switch (
I->getOpcode()) {
1826 case Instruction::LShr: OpReg = X86::SHR32rCL;
break;
1827 case Instruction::AShr: OpReg = X86::SAR32rCL;
break;
1828 case Instruction::Shl: OpReg = X86::SHL32rCL;
break;
1830 }
else if (
I->getType()->isIntegerTy(64)) {
1832 RC = &X86::GR64RegClass;
1833 switch (
I->getOpcode()) {
1835 case Instruction::LShr: OpReg = X86::SHR64rCL;
break;
1836 case Instruction::AShr: OpReg = X86::SAR64rCL;
break;
1837 case Instruction::Shl: OpReg = X86::SHL64rCL;
break;
1844 if (!isTypeLegal(
I->getType(), VT))
1847 Register Op0Reg = getRegForValue(
I->getOperand(0));
1851 Register Op1Reg = getRegForValue(
I->getOperand(1));
1854 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(TargetOpcode::COPY),
1859 if (CReg != X86::CL)
1860 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1861 TII.get(TargetOpcode::KILL), X86::CL)
1862 .
addReg(CReg, RegState::Kill);
1864 Register ResultReg = createResultReg(RC);
1865 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(OpReg), ResultReg)
1867 updateValueMap(
I, ResultReg);
1871bool X86FastISel::X86SelectDivRem(
const Instruction *
I) {
1872 const static unsigned NumTypes = 4;
1873 const static unsigned NumOps = 4;
1874 const static bool S =
true;
1875 const static bool U =
false;
1876 const static unsigned Copy = TargetOpcode::COPY;
1886 const static struct DivRemEntry {
1888 const TargetRegisterClass *RC;
1892 struct DivRemResult {
1894 unsigned OpSignExtend;
1898 unsigned DivRemResultReg;
1901 } OpTable[NumTypes] = {
1902 { &X86::GR8RegClass, X86::AX, 0, {
1903 { X86::IDIV8r, 0, X86::MOVSX16rr8, X86::AL, S },
1904 { X86::IDIV8r, 0, X86::MOVSX16rr8, X86::AH, S },
1905 { X86::DIV8r, 0, X86::MOVZX16rr8, X86::AL,
U },
1906 { X86::DIV8r, 0, X86::MOVZX16rr8, X86::AH,
U },
1909 { &X86::GR16RegClass, X86::AX, X86::DX, {
1910 { X86::IDIV16r, X86::CWD,
Copy, X86::AX, S },
1911 { X86::IDIV16r, X86::CWD,
Copy, X86::DX, S },
1912 { X86::DIV16r, X86::MOV32r0,
Copy, X86::AX,
U },
1913 { X86::DIV16r, X86::MOV32r0,
Copy, X86::DX,
U },
1916 { &X86::GR32RegClass, X86::EAX, X86::EDX, {
1917 { X86::IDIV32r, X86::CDQ,
Copy, X86::EAX, S },
1918 { X86::IDIV32r, X86::CDQ,
Copy, X86::EDX, S },
1919 { X86::DIV32r, X86::MOV32r0,
Copy, X86::EAX,
U },
1920 { X86::DIV32r, X86::MOV32r0,
Copy, X86::EDX,
U },
1923 { &X86::GR64RegClass, X86::RAX, X86::RDX, {
1924 { X86::IDIV64r, X86::CQO,
Copy, X86::RAX, S },
1925 { X86::IDIV64r, X86::CQO,
Copy, X86::RDX, S },
1926 { X86::DIV64r, X86::MOV32r0,
Copy, X86::RAX,
U },
1927 { X86::DIV64r, X86::MOV32r0,
Copy, X86::RDX,
U },
1933 if (!isTypeLegal(
I->getType(), VT))
1938 default:
return false;
1939 case MVT::i8: TypeIndex = 0;
break;
1940 case MVT::i16: TypeIndex = 1;
break;
1941 case MVT::i32: TypeIndex = 2;
break;
1942 case MVT::i64: TypeIndex = 3;
1943 if (!Subtarget->is64Bit())
1948 switch (
I->getOpcode()) {
1950 case Instruction::SDiv:
OpIndex = 0;
break;
1951 case Instruction::SRem:
OpIndex = 1;
break;
1952 case Instruction::UDiv:
OpIndex = 2;
break;
1953 case Instruction::URem:
OpIndex = 3;
break;
1956 const DivRemEntry &
TypeEntry = OpTable[TypeIndex];
1958 Register Op0Reg = getRegForValue(
I->getOperand(0));
1961 Register Op1Reg = getRegForValue(
I->getOperand(1));
1966 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1969 if (OpEntry.OpSignExtend) {
1970 if (OpEntry.IsOpSigned)
1971 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1972 TII.get(OpEntry.OpSignExtend));
1974 Register Zero32 = createResultReg(&X86::GR32RegClass);
1975 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1976 TII.get(X86::MOV32r0), Zero32);
1981 if (VT == MVT::i16) {
1982 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(Copy),
1984 .
addReg(Zero32, {}, X86::sub_16bit);
1985 }
else if (VT == MVT::i32) {
1986 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1989 }
else if (VT == MVT::i64) {
1990 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1991 TII.get(TargetOpcode::SUBREG_TO_REG),
TypeEntry.HighInReg)
1998 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
1999 TII.get(OpEntry.OpDivRem)).
addReg(Op1Reg);
2009 if ((
I->getOpcode() == Instruction::SRem ||
2010 I->getOpcode() == Instruction::URem) &&
2011 OpEntry.DivRemResultReg == X86::AH && Subtarget->is64Bit()) {
2012 Register SourceSuperReg = createResultReg(&X86::GR16RegClass);
2013 Register ResultSuperReg = createResultReg(&X86::GR16RegClass);
2014 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2015 TII.get(Copy), SourceSuperReg).
addReg(X86::AX);
2018 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::SHR16ri),
2022 ResultReg = fastEmitInst_extractsubreg(MVT::i8, ResultSuperReg,
2027 ResultReg = createResultReg(
TypeEntry.RC);
2028 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(Copy), ResultReg)
2029 .
addReg(OpEntry.DivRemResultReg);
2031 updateValueMap(
I, ResultReg);
2038bool X86FastISel::X86FastEmitCMoveSelect(MVT RetVT,
const Instruction *
I) {
2040 if (!Subtarget->canUseCMOV())
2044 if (RetVT < MVT::i16 || RetVT > MVT::i64)
2048 const TargetRegisterClass *RC = TLI.getRegClassFor(RetVT);
2049 bool NeedTest =
true;
2056 if (CI && (CI->getParent() ==
I->getParent())) {
2060 static const uint16_t SETFOpcTable[2][3] = {
2064 const uint16_t *SETFOpc =
nullptr;
2065 switch (Predicate) {
2068 SETFOpc = &SETFOpcTable[0][0];
2072 SETFOpc = &SETFOpcTable[1][0];
2086 EVT CmpVT = TLI.getValueType(
DL, CmpLHS->
getType());
2088 if (!X86FastEmitCompare(CmpLHS, CmpRHS, CmpVT, CI->getDebugLoc()))
2092 Register FlagReg1 = createResultReg(&X86::GR8RegClass);
2093 Register FlagReg2 = createResultReg(&X86::GR8RegClass);
2100 auto const &
II =
TII.get(SETFOpc[2]);
2101 if (
II.getNumDefs()) {
2102 Register TmpReg = createResultReg(&X86::GR8RegClass);
2103 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
II, TmpReg)
2106 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
II)
2111 }
else if (foldX86XALUIntrinsic(CC,
I,
Cond)) {
2132 if (
MRI.getRegClass(CondReg) == &X86::VK1RegClass) {
2134 CondReg = createResultReg(&X86::GR32RegClass);
2135 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2136 TII.get(TargetOpcode::COPY), CondReg)
2138 CondReg = fastEmitInst_extractsubreg(MVT::i8, CondReg, X86::sub_8bit);
2140 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::TEST8ri))
2150 if (!LHSReg || !RHSReg)
2153 const TargetRegisterInfo &
TRI = *Subtarget->getRegisterInfo();
2155 Subtarget->hasNDD());
2156 Register ResultReg = fastEmitInst_rri(
Opc, RC, RHSReg, LHSReg, CC);
2157 updateValueMap(
I, ResultReg);
2166bool X86FastISel::X86FastEmitSSESelect(MVT RetVT,
const Instruction *
I) {
2171 if (!CI || (CI->getParent() !=
I->getParent()))
2175 !((Subtarget->hasSSE1() && RetVT == MVT::f32) ||
2176 (Subtarget->hasSSE2() && RetVT == MVT::f64)))
2188 if (CmpRHSC && CmpRHSC->isNullValue())
2195 if (CC > 7 && !Subtarget->hasAVX())
2206 Register CmpLHSReg = getRegForValue(CmpLHS);
2207 Register CmpRHSReg = getRegForValue(CmpRHS);
2208 if (!LHSReg || !RHSReg || !CmpLHSReg || !CmpRHSReg)
2211 const TargetRegisterClass *RC = TLI.getRegClassFor(RetVT);
2214 if (Subtarget->hasAVX512()) {
2216 const TargetRegisterClass *VR128X = &X86::VR128XRegClass;
2217 const TargetRegisterClass *VK1 = &X86::VK1RegClass;
2219 unsigned CmpOpcode =
2220 (RetVT == MVT::f32) ? X86::VCMPSSZrri :
X86::VCMPSDZrri;
2221 Register CmpReg = fastEmitInst_rri(CmpOpcode, VK1, CmpLHSReg, CmpRHSReg,
2226 Register ImplicitDefReg = createResultReg(VR128X);
2227 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2228 TII.get(TargetOpcode::IMPLICIT_DEF), ImplicitDefReg);
2232 unsigned MovOpcode =
2233 (RetVT == MVT::f32) ? X86::VMOVSSZrrk :
X86::VMOVSDZrrk;
2234 Register MovReg = fastEmitInst_rrrr(MovOpcode, VR128X, RHSReg, CmpReg,
2235 ImplicitDefReg, LHSReg);
2237 ResultReg = createResultReg(RC);
2238 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2239 TII.get(TargetOpcode::COPY), ResultReg).
addReg(MovReg);
2241 }
else if (Subtarget->hasAVX()) {
2242 const TargetRegisterClass *VR128 = &X86::VR128RegClass;
2249 unsigned CmpOpcode =
2250 (RetVT == MVT::f32) ? X86::VCMPSSrri :
X86::VCMPSDrri;
2251 unsigned BlendOpcode =
2252 (RetVT == MVT::f32) ? X86::VBLENDVPSrrr :
X86::VBLENDVPDrrr;
2254 Register CmpReg = fastEmitInst_rri(CmpOpcode, RC, CmpLHSReg, CmpRHSReg,
2256 Register VBlendReg = fastEmitInst_rrr(BlendOpcode, VR128, RHSReg, LHSReg,
2258 ResultReg = createResultReg(RC);
2259 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2260 TII.get(TargetOpcode::COPY), ResultReg).
addReg(VBlendReg);
2263 static const uint16_t OpcTable[2][4] = {
2264 { X86::CMPSSrri, X86::ANDPSrr, X86::ANDNPSrr, X86::ORPSrr },
2265 { X86::CMPSDrri, X86::ANDPDrr, X86::ANDNPDrr, X86::ORPDrr }
2268 const uint16_t *
Opc =
nullptr;
2270 default:
return false;
2271 case MVT::f32:
Opc = &OpcTable[0][0];
break;
2272 case MVT::f64:
Opc = &OpcTable[1][0];
break;
2275 const TargetRegisterClass *VR128 = &X86::VR128RegClass;
2276 Register CmpReg = fastEmitInst_rri(
Opc[0], RC, CmpLHSReg, CmpRHSReg, CC);
2277 Register AndReg = fastEmitInst_rr(
Opc[1], VR128, CmpReg, LHSReg);
2278 Register AndNReg = fastEmitInst_rr(
Opc[2], VR128, CmpReg, RHSReg);
2279 Register OrReg = fastEmitInst_rr(
Opc[3], VR128, AndNReg, AndReg);
2280 ResultReg = createResultReg(RC);
2281 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2282 TII.get(TargetOpcode::COPY), ResultReg).
addReg(OrReg);
2284 updateValueMap(
I, ResultReg);
2288bool X86FastISel::X86FastEmitPseudoSelect(MVT RetVT,
const Instruction *
I) {
2293 default:
return false;
2294 case MVT::i8:
Opc = X86::CMOV_GR8;
break;
2295 case MVT::i16:
Opc = X86::CMOV_GR16;
break;
2296 case MVT::i32:
Opc = X86::CMOV_GR32;
break;
2298 Opc = Subtarget->hasAVX512() ? X86::CMOV_FR16X : X86::CMOV_FR16;
break;
2300 Opc = Subtarget->hasAVX512() ? X86::CMOV_FR32X : X86::CMOV_FR32;
break;
2302 Opc = Subtarget->hasAVX512() ? X86::CMOV_FR64X : X86::CMOV_FR64;
break;
2312 if (CI && (CI->getParent() ==
I->getParent())) {
2324 EVT CmpVT = TLI.getValueType(
DL, CmpLHS->
getType());
2325 if (!X86FastEmitCompare(CmpLHS, CmpRHS, CmpVT, CI->getDebugLoc()))
2333 if (
MRI.getRegClass(CondReg) == &X86::VK1RegClass) {
2335 CondReg = createResultReg(&X86::GR32RegClass);
2336 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2337 TII.get(TargetOpcode::COPY), CondReg)
2339 CondReg = fastEmitInst_extractsubreg(MVT::i8, CondReg, X86::sub_8bit);
2341 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::TEST8ri))
2351 if (!LHSReg || !RHSReg)
2354 const TargetRegisterClass *RC = TLI.getRegClassFor(RetVT);
2357 fastEmitInst_rri(
Opc, RC, RHSReg, LHSReg, CC);
2358 updateValueMap(
I, ResultReg);
2362bool X86FastISel::X86SelectSelect(
const Instruction *
I) {
2364 if (!isTypeLegal(
I->getType(), RetVT))
2370 const Value *Opnd =
nullptr;
2371 switch (Predicate) {
2378 Register OpReg = getRegForValue(Opnd);
2381 const TargetRegisterClass *RC = TLI.getRegClassFor(RetVT);
2382 Register ResultReg = createResultReg(RC);
2383 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2384 TII.get(TargetOpcode::COPY), ResultReg)
2386 updateValueMap(
I, ResultReg);
2392 if (X86FastEmitCMoveSelect(RetVT,
I))
2396 if (X86FastEmitSSESelect(RetVT,
I))
2401 if (X86FastEmitPseudoSelect(RetVT,
I))
2408bool X86FastISel::X86SelectIntToFP(
const Instruction *
I,
bool IsSigned) {
2413 bool HasAVX512 = Subtarget->hasAVX512();
2414 if (!Subtarget->hasAVX() || (!IsSigned && !HasAVX512))
2418 EVT SrcVT = TLI.getValueType(
DL,
I->getOperand(0)->getType());
2419 if (SrcVT != MVT::i32 && SrcVT != MVT::i64)
2423 Register OpReg = getRegForValue(
I->getOperand(0));
2429 static const uint16_t SCvtOpc[2][2][2] = {
2430 { { X86::VCVTSI2SSrr, X86::VCVTSI642SSrr },
2431 { X86::VCVTSI2SDrr, X86::VCVTSI642SDrr } },
2432 { { X86::VCVTSI2SSZrr, X86::VCVTSI642SSZrr },
2433 { X86::VCVTSI2SDZrr, X86::VCVTSI642SDZrr } },
2435 static const uint16_t UCvtOpc[2][2] = {
2436 { X86::VCVTUSI2SSZrr, X86::VCVTUSI642SSZrr },
2437 { X86::VCVTUSI2SDZrr, X86::VCVTUSI642SDZrr },
2439 bool Is64Bit = SrcVT == MVT::i64;
2441 if (
I->getType()->isDoubleTy()) {
2443 Opcode = IsSigned ? SCvtOpc[HasAVX512][1][Is64Bit] : UCvtOpc[1][Is64Bit];
2444 }
else if (
I->getType()->isFloatTy()) {
2446 Opcode = IsSigned ? SCvtOpc[HasAVX512][0][Is64Bit] : UCvtOpc[0][Is64Bit];
2451 const TargetRegisterClass *RC = TLI.getRegClassFor(DstVT);
2452 Register ImplicitDefReg = createResultReg(RC);
2453 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2454 TII.get(TargetOpcode::IMPLICIT_DEF), ImplicitDefReg);
2455 Register ResultReg = fastEmitInst_rr(Opcode, RC, ImplicitDefReg, OpReg);
2456 updateValueMap(
I, ResultReg);
2460bool X86FastISel::X86SelectSIToFP(
const Instruction *
I) {
2461 return X86SelectIntToFP(
I,
true);
2464bool X86FastISel::X86SelectUIToFP(
const Instruction *
I) {
2465 return X86SelectIntToFP(
I,
false);
2469bool X86FastISel::X86SelectFPExtOrFPTrunc(
const Instruction *
I,
2471 const TargetRegisterClass *RC) {
2472 assert((
I->getOpcode() == Instruction::FPExt ||
2473 I->getOpcode() == Instruction::FPTrunc) &&
2474 "Instruction must be an FPExt or FPTrunc!");
2475 bool HasAVX = Subtarget->hasAVX();
2477 Register OpReg = getRegForValue(
I->getOperand(0));
2483 ImplicitDefReg = createResultReg(RC);
2484 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2485 TII.get(TargetOpcode::IMPLICIT_DEF), ImplicitDefReg);
2489 Register ResultReg = createResultReg(RC);
2490 MachineInstrBuilder MIB;
2491 MIB =
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(TargetOpc),
2495 MIB.
addReg(ImplicitDefReg);
2498 updateValueMap(
I, ResultReg);
2502bool X86FastISel::X86SelectFPExt(
const Instruction *
I) {
2503 if (Subtarget->hasSSE2() &&
I->getType()->isDoubleTy() &&
2504 I->getOperand(0)->getType()->isFloatTy()) {
2505 bool HasAVX512 = Subtarget->hasAVX512();
2508 HasAVX512 ? X86::VCVTSS2SDZrr
2509 : Subtarget->hasAVX() ? X86::VCVTSS2SDrr : X86::CVTSS2SDrr;
2510 return X86SelectFPExtOrFPTrunc(
I,
Opc, TLI.getRegClassFor(MVT::f64));
2516bool X86FastISel::X86SelectFPTrunc(
const Instruction *
I) {
2517 if (Subtarget->hasSSE2() &&
I->getType()->isFloatTy() &&
2518 I->getOperand(0)->getType()->isDoubleTy()) {
2519 bool HasAVX512 = Subtarget->hasAVX512();
2522 HasAVX512 ? X86::VCVTSD2SSZrr
2523 : Subtarget->hasAVX() ? X86::VCVTSD2SSrr : X86::CVTSD2SSrr;
2524 return X86SelectFPExtOrFPTrunc(
I,
Opc, TLI.getRegClassFor(MVT::f32));
2530bool X86FastISel::X86SelectTrunc(
const Instruction *
I) {
2531 EVT SrcVT = TLI.getValueType(
DL,
I->getOperand(0)->getType());
2532 EVT DstVT = TLI.getValueType(
DL,
I->getType());
2535 if (DstVT != MVT::i8 && DstVT != MVT::i1)
2537 if (!TLI.isTypeLegal(SrcVT))
2540 Register InputReg = getRegForValue(
I->getOperand(0));
2545 if (SrcVT == MVT::i8) {
2547 updateValueMap(
I, InputReg);
2552 Register ResultReg = fastEmitInst_extractsubreg(MVT::i8, InputReg,
2557 updateValueMap(
I, ResultReg);
2561bool X86FastISel::X86SelectBitCast(
const Instruction *
I) {
2564 if (!Subtarget->hasSSE2() ||
2565 !isTypeLegal(
I->getOperand(0)->getType(), SrcVT) ||
2566 !isTypeLegal(
I->getType(), DstVT))
2581 const TargetRegisterClass *DstClass = TLI.getRegClassFor(DstVT);
2582 Register ResultReg = createResultReg(DstClass);
2583 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(TargetOpcode::COPY),
2587 updateValueMap(
I, ResultReg);
2591bool X86FastISel::IsMemcpySmall(uint64_t Len) {
2592 return Len <= (Subtarget->is64Bit() ? 32 : 16);
2595bool X86FastISel::TryEmitSmallMemcpy(X86AddressMode DestAM,
2596 X86AddressMode SrcAM, uint64_t Len) {
2599 if (!IsMemcpySmall(Len))
2602 bool i64Legal = Subtarget->is64Bit();
2607 if (Len >= 8 && i64Legal)
2617 bool RV = X86FastEmitLoad(VT, SrcAM,
nullptr,
Reg);
2618 RV &= X86FastEmitStore(VT,
Reg, DestAM);
2619 assert(RV &&
"Failed to emit load or store??");
2631bool X86FastISel::fastLowerIntrinsicCall(
const IntrinsicInst *
II) {
2633 switch (
II->getIntrinsicID()) {
2636 case Intrinsic::frameaddress: {
2637 MachineFunction *MF = FuncInfo.MF;
2641 Type *RetTy =
II->getCalledFunction()->getReturnType();
2644 if (!isTypeLegal(RetTy, VT))
2648 const TargetRegisterClass *RC =
nullptr;
2652 case MVT::i32:
Opc = X86::MOV32rm; RC = &X86::GR32RegClass;
break;
2653 case MVT::i64:
Opc = X86::MOV64rm; RC = &X86::GR64RegClass;
break;
2661 const X86RegisterInfo *RegInfo = Subtarget->getRegisterInfo();
2663 assert(((FrameReg == X86::RBP && VT == MVT::i64) ||
2664 (FrameReg == X86::EBP && VT == MVT::i32)) &&
2665 "Invalid Frame Register!");
2670 Register SrcReg = createResultReg(RC);
2671 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2672 TII.get(TargetOpcode::COPY), SrcReg).
addReg(FrameReg);
2681 Register DestReg = createResultReg(RC);
2683 TII.get(
Opc), DestReg), SrcReg);
2687 updateValueMap(
II, SrcReg);
2690 case Intrinsic::memcpy: {
2700 if (IsMemcpySmall(Len)) {
2701 X86AddressMode DestAM, SrcAM;
2705 TryEmitSmallMemcpy(DestAM, SrcAM, Len);
2710 unsigned SizeWidth = Subtarget->is64Bit() ? 64 : 32;
2717 return lowerCallTo(
II,
"memcpy",
II->arg_size() - 1);
2719 case Intrinsic::memset: {
2725 unsigned SizeWidth = Subtarget->is64Bit() ? 64 : 32;
2732 return lowerCallTo(
II,
"memset",
II->arg_size() - 1);
2734 case Intrinsic::stackprotector: {
2736 EVT PtrTy = TLI.getPointerTy(
DL);
2738 const Value *Op1 =
II->getArgOperand(0);
2746 if (!X86FastEmitStore(PtrTy, Op1, AM))
return false;
2749 case Intrinsic::dbg_declare: {
2755 const MCInstrDesc &
II =
TII.get(TargetOpcode::DBG_VALUE);
2757 "Expected inlined-at fields to agree");
2764 case Intrinsic::trap: {
2765 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::TRAP));
2768 case Intrinsic::sqrt: {
2769 if (!Subtarget->hasSSE1())
2772 Type *RetTy =
II->getCalledFunction()->getReturnType();
2775 if (!isTypeLegal(RetTy, VT))
2781 static const uint16_t SqrtOpc[3][2] = {
2782 { X86::SQRTSSr, X86::SQRTSDr },
2783 { X86::VSQRTSSr, X86::VSQRTSDr },
2784 { X86::VSQRTSSZr, X86::VSQRTSDZr },
2786 unsigned AVXLevel = Subtarget->hasAVX512() ? 2 :
2787 Subtarget->hasAVX() ? 1 :
2791 default:
return false;
2792 case MVT::f32:
Opc = SqrtOpc[AVXLevel][0];
break;
2793 case MVT::f64:
Opc = SqrtOpc[AVXLevel][1];
break;
2796 const Value *SrcVal =
II->getArgOperand(0);
2797 Register SrcReg = getRegForValue(SrcVal);
2802 const TargetRegisterClass *RC = TLI.getRegClassFor(VT);
2805 ImplicitDefReg = createResultReg(RC);
2806 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2807 TII.get(TargetOpcode::IMPLICIT_DEF), ImplicitDefReg);
2810 Register ResultReg = createResultReg(RC);
2811 MachineInstrBuilder MIB;
2812 MIB =
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(
Opc),
2816 MIB.
addReg(ImplicitDefReg);
2820 updateValueMap(
II, ResultReg);
2823 case Intrinsic::sadd_with_overflow:
2824 case Intrinsic::uadd_with_overflow:
2825 case Intrinsic::ssub_with_overflow:
2826 case Intrinsic::usub_with_overflow:
2827 case Intrinsic::smul_with_overflow:
2828 case Intrinsic::umul_with_overflow: {
2833 Type *RetTy = Ty->getTypeAtIndex(0U);
2836 "Overflow value expected to be an i1");
2839 if (!isTypeLegal(RetTy, VT))
2842 if (VT < MVT::i8 || VT > MVT::i64)
2853 switch (
II->getIntrinsicID()) {
2855 case Intrinsic::sadd_with_overflow:
2857 case Intrinsic::uadd_with_overflow:
2859 case Intrinsic::ssub_with_overflow:
2861 case Intrinsic::usub_with_overflow:
2863 case Intrinsic::smul_with_overflow:
2865 case Intrinsic::umul_with_overflow:
2876 static const uint16_t
Opc[2][4] = {
2877 { X86::INC8r, X86::INC16r, X86::INC32r, X86::INC64r },
2878 { X86::DEC8r, X86::DEC16r, X86::DEC32r, X86::DEC64r }
2884 ResultReg = createResultReg(TLI.getRegClassFor(VT));
2886 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2890 ResultReg = fastEmit_ri(VT, VT, BaseOpc, LHSReg, CI->
getZExtValue());
2895 RHSReg = getRegForValue(
RHS);
2898 ResultReg = fastEmit_rr(VT, VT, BaseOpc, LHSReg, RHSReg);
2904 static const uint16_t MULOpc[] =
2905 { X86::MUL8r, X86::MUL16r, X86::MUL32r, X86::MUL64r };
2906 static const MCPhysReg Reg[] = { X86::AL, X86::AX, X86::EAX, X86::RAX };
2909 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2912 ResultReg = fastEmitInst_r(MULOpc[VT.
SimpleTy-MVT::i8],
2913 TLI.getRegClassFor(VT), RHSReg);
2915 static const uint16_t MULOpc[] =
2916 { X86::IMUL8r, X86::IMUL16rr, X86::IMUL32rr, X86::IMUL64rr };
2917 if (VT == MVT::i8) {
2920 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
2921 TII.get(TargetOpcode::COPY), X86::AL)
2923 ResultReg = fastEmitInst_r(MULOpc[0], TLI.getRegClassFor(VT), RHSReg);
2925 ResultReg = fastEmitInst_rr(MULOpc[VT.
SimpleTy-MVT::i8],
2926 TLI.getRegClassFor(VT), LHSReg, RHSReg);
2933 Register ResultReg2 = createResultReg(&X86::GR8RegClass);
2934 assert((ResultReg+1) == ResultReg2 &&
"Nonconsecutive result registers.");
2939 updateValueMap(
II, ResultReg, 2);
2942 case Intrinsic::x86_sse_cvttss2si:
2943 case Intrinsic::x86_sse_cvttss2si64:
2944 case Intrinsic::x86_sse2_cvttsd2si:
2945 case Intrinsic::x86_sse2_cvttsd2si64: {
2947 switch (
II->getIntrinsicID()) {
2949 case Intrinsic::x86_sse_cvttss2si:
2950 case Intrinsic::x86_sse_cvttss2si64:
2951 if (!Subtarget->hasSSE1())
2953 IsInputDouble =
false;
2955 case Intrinsic::x86_sse2_cvttsd2si:
2956 case Intrinsic::x86_sse2_cvttsd2si64:
2957 if (!Subtarget->hasSSE2())
2959 IsInputDouble =
true;
2963 Type *RetTy =
II->getCalledFunction()->getReturnType();
2965 if (!isTypeLegal(RetTy, VT))
2968 static const uint16_t CvtOpc[3][2][2] = {
2969 { { X86::CVTTSS2SIrr, X86::CVTTSS2SI64rr },
2970 { X86::CVTTSD2SIrr, X86::CVTTSD2SI64rr } },
2971 { { X86::VCVTTSS2SIrr, X86::VCVTTSS2SI64rr },
2972 { X86::VCVTTSD2SIrr, X86::VCVTTSD2SI64rr } },
2973 { { X86::VCVTTSS2SIZrr, X86::VCVTTSS2SI64Zrr },
2974 { X86::VCVTTSD2SIZrr, X86::VCVTTSD2SI64Zrr } },
2976 unsigned AVXLevel = Subtarget->hasAVX512() ? 2 :
2977 Subtarget->hasAVX() ? 1 :
2982 case MVT::i32:
Opc = CvtOpc[AVXLevel][IsInputDouble][0];
break;
2983 case MVT::i64:
Opc = CvtOpc[AVXLevel][IsInputDouble][1];
break;
2995 Op =
IE->getOperand(1);
2998 Op =
IE->getOperand(0);
3005 Register ResultReg = createResultReg(TLI.getRegClassFor(VT));
3006 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(
Opc), ResultReg)
3009 updateValueMap(
II, ResultReg);
3012 case Intrinsic::x86_sse42_crc32_32_8:
3013 case Intrinsic::x86_sse42_crc32_32_16:
3014 case Intrinsic::x86_sse42_crc32_32_32:
3015 case Intrinsic::x86_sse42_crc32_64_64: {
3016 if (!Subtarget->hasCRC32())
3019 Type *RetTy =
II->getCalledFunction()->getReturnType();
3022 if (!isTypeLegal(RetTy, VT))
3026 const TargetRegisterClass *RC =
nullptr;
3028 switch (
II->getIntrinsicID()) {
3031#define GET_EGPR_IF_ENABLED(OPC) Subtarget->hasEGPR() ? OPC##_EVEX : OPC
3032 case Intrinsic::x86_sse42_crc32_32_8:
3034 RC = &X86::GR32RegClass;
3036 case Intrinsic::x86_sse42_crc32_32_16:
3038 RC = &X86::GR32RegClass;
3040 case Intrinsic::x86_sse42_crc32_32_32:
3042 RC = &X86::GR32RegClass;
3044 case Intrinsic::x86_sse42_crc32_64_64:
3046 RC = &X86::GR64RegClass;
3048#undef GET_EGPR_IF_ENABLED
3056 if (!LHSReg || !RHSReg)
3059 Register ResultReg = fastEmitInst_rr(
Opc, RC, LHSReg, RHSReg);
3063 updateValueMap(
II, ResultReg);
3069bool X86FastISel::fastLowerArguments() {
3070 if (!FuncInfo.CanLowerReturn)
3077 CallingConv::ID CC =
F->getCallingConv();
3078 if (CC != CallingConv::C)
3081 if (Subtarget->isCallingConvWin64(CC))
3084 if (!Subtarget->is64Bit())
3087 if (Subtarget->useSoftFloat())
3091 unsigned GPRCnt = 0;
3092 unsigned FPRCnt = 0;
3093 for (
auto const &Arg :
F->args()) {
3094 if (Arg.hasAttribute(Attribute::ByVal) ||
3095 Arg.hasAttribute(Attribute::InReg) ||
3096 Arg.hasAttribute(Attribute::StructRet) ||
3097 Arg.hasAttribute(Attribute::SwiftSelf) ||
3098 Arg.hasAttribute(Attribute::SwiftAsync) ||
3099 Arg.hasAttribute(Attribute::SwiftError) ||
3100 Arg.hasAttribute(Attribute::Nest))
3103 Type *ArgTy = Arg.getType();
3107 EVT ArgVT = TLI.getValueType(
DL, ArgTy);
3108 if (!ArgVT.
isSimple())
return false;
3110 default:
return false;
3117 if (!Subtarget->hasSSE1())
3130 static const MCPhysReg GPR32ArgRegs[] = {
3131 X86::EDI, X86::ESI, X86::EDX, X86::ECX, X86::R8D, X86::R9D
3133 static const MCPhysReg GPR64ArgRegs[] = {
3134 X86::RDI, X86::RSI, X86::RDX, X86::RCX, X86::R8 , X86::R9
3137 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3,
3138 X86::XMM4, X86::XMM5, X86::XMM6, X86::XMM7
3141 unsigned GPRIdx = 0;
3142 unsigned FPRIdx = 0;
3143 for (
auto const &Arg :
F->args()) {
3144 MVT VT = TLI.getSimpleValueType(
DL, Arg.getType());
3145 const TargetRegisterClass *RC = TLI.getRegClassFor(VT);
3149 case MVT::i32: SrcReg = GPR32ArgRegs[GPRIdx++];
break;
3150 case MVT::i64: SrcReg = GPR64ArgRegs[GPRIdx++];
break;
3151 case MVT::f32: [[fallthrough]];
3152 case MVT::f64: SrcReg = XMMArgRegs[FPRIdx++];
break;
3154 Register DstReg = FuncInfo.MF->addLiveIn(SrcReg, RC);
3158 Register ResultReg = createResultReg(RC);
3159 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
3160 TII.get(TargetOpcode::COPY), ResultReg)
3162 updateValueMap(&Arg, ResultReg);
3170 if (Subtarget->is64Bit())
3187bool X86FastISel::fastLowerCall(CallLoweringInfo &CLI) {
3188 auto &OutVals = CLI.OutVals;
3189 auto &OutFlags = CLI.OutFlags;
3190 auto &OutRegs = CLI.OutRegs;
3191 auto &Ins = CLI.Ins;
3192 auto &InRegs = CLI.InRegs;
3193 CallingConv::ID CC = CLI.CallConv;
3194 bool &IsTailCall = CLI.IsTailCall;
3195 bool IsVarArg = CLI.IsVarArg;
3198 const auto *CB = CLI.CB;
3200 bool Is64Bit = Subtarget->is64Bit();
3201 bool IsWin64 = Subtarget->isCallingConvWin64(CC);
3205 if (CB && CB->doesNoCfCheck())
3209 if ((CB &&
isa<CallInst>(CB) && CB->hasFnAttr(
"no_caller_saved_registers")))
3213 if ((CB && CB->hasFnAttr(
"no_callee_saved_registers")))
3221 if (Subtarget->useIndirectThunkCalls())
3226 default:
return false;
3227 case CallingConv::C:
3228 case CallingConv::Fast:
3229 case CallingConv::Tail:
3230 case CallingConv::Swift:
3231 case CallingConv::SwiftTail:
3232 case CallingConv::X86_FastCall:
3233 case CallingConv::X86_StdCall:
3234 case CallingConv::X86_ThisCall:
3235 case CallingConv::Win64:
3236 case CallingConv::X86_64_SysV:
3237 case CallingConv::CFGuard_Check:
3247 if ((CC == CallingConv::Fast && TM.Options.GuaranteedTailCallOpt) ||
3248 CC == CallingConv::Tail || CC == CallingConv::SwiftTail)
3253 if (IsVarArg && IsWin64)
3257 if (CLI.CB && CLI.CB->hasInAllocaArgument())
3260 for (
auto Flag : CLI.OutFlags)
3261 if (
Flag.isSwiftError() ||
Flag.isPreallocated())
3271 SmallVector<Register, 16> ArgRegs;
3276 for (
int i = 0, e = OutVals.size(); i != e; ++i) {
3277 Value *&Val = OutVals[i];
3278 ISD::ArgFlagsTy
Flags = OutFlags[i];
3293 if (TI && TI->getType()->isIntegerTy(1) && CLI.CB &&
3294 (TI->getParent() == CLI.CB->getParent()) && TI->hasOneUse()) {
3295 Value *PrevVal = TI->getOperand(0);
3296 ResultReg = getRegForValue(PrevVal);
3301 if (!isTypeLegal(PrevVal->
getType(), VT))
3304 ResultReg = fastEmit_ri(VT, VT,
ISD::AND, ResultReg, 1);
3306 if (!isTypeLegal(Val->
getType(), VT) ||
3309 ResultReg = getRegForValue(Val);
3322 CCState CCInfo(CC, IsVarArg, *FuncInfo.MF, ArgLocs, CLI.RetTy->getContext());
3326 CCInfo.AllocateStack(32,
Align(8));
3328 CCInfo.AnalyzeCallOperands(OutVTs, OutFlags, ArgTys,
CC_X86);
3331 unsigned NumBytes = CCInfo.getAlignedCallFrameSize();
3334 unsigned AdjStackDown =
TII.getCallFrameSetupOpcode();
3335 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(AdjStackDown))
3339 const X86RegisterInfo *RegInfo = Subtarget->getRegisterInfo();
3340 for (
const CCValAssign &VA : ArgLocs) {
3344 if (ArgVT == MVT::x86mmx)
3354 "Unexpected extend");
3356 if (ArgVT == MVT::i1)
3361 assert(Emitted &&
"Failed to emit a sext!"); (void)Emitted;
3367 "Unexpected extend");
3370 if (ArgVT == MVT::i1) {
3372 ArgReg = fastEmitZExtFromI1(MVT::i8, ArgReg);
3381 assert(Emitted &&
"Failed to emit a zext!"); (void)Emitted;
3387 "Unexpected extend");
3397 assert(Emitted &&
"Failed to emit a aext!"); (void)Emitted;
3403 assert(ArgReg &&
"Failed to emit a bitcast!");
3424 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
3437 AM.
Disp = LocMemOffset;
3440 MachineMemOperand *MMO = FuncInfo.MF->getMachineMemOperand(
3443 if (
Flags.isByVal()) {
3444 X86AddressMode SrcAM;
3446 if (!TryEmitSmallMemcpy(AM, SrcAM,
Flags.getByValSize()))
3452 if (!X86FastEmitStore(ArgVT, ArgVal, AM, MMO))
3455 if (!X86FastEmitStore(ArgVT, ArgReg, AM, MMO))
3463 if (Subtarget->isPICStyleGOT()) {
3464 Register Base = getInstrInfo()->getGlobalBaseReg(FuncInfo.MF);
3465 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
3469 if (Is64Bit && IsVarArg && !IsWin64) {
3480 X86::XMM0, X86::XMM1, X86::XMM2, X86::XMM3,
3481 X86::XMM4, X86::XMM5, X86::XMM6, X86::XMM7
3483 unsigned NumXMMRegs = CCInfo.getFirstUnallocated(XMMArgRegs);
3484 assert((Subtarget->hasSSE1() || !NumXMMRegs)
3485 &&
"SSE registers cannot be used when SSE is disabled");
3486 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOV8ri),
3487 X86::AL).
addImm(NumXMMRegs);
3492 X86AddressMode CalleeAM;
3493 if (!X86SelectCallAddress(Callee, CalleeAM))
3497 const GlobalValue *GV =
nullptr;
3498 if (CalleeAM.
GV !=
nullptr) {
3500 }
else if (CalleeAM.
Base.
Reg) {
3506 MachineInstrBuilder MIB;
3509 unsigned CallOpc = Is64Bit ? X86::CALL64r : X86::CALL32r;
3510 MIB =
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(CallOpc))
3514 assert(GV &&
"Not a direct call");
3516 unsigned char OpFlags = Subtarget->classifyGlobalFunctionReference(GV);
3528 unsigned CallOpc = NeedLoad
3529 ? (Is64Bit ? X86::CALL64m : X86::CALL32m)
3530 : (Is64Bit ?
X86::CALL64pcrel32 :
X86::CALLpcrel32);
3532 MIB =
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(CallOpc));
3536 MIB.
addSym(Symbol, OpFlags);
3548 if (Subtarget->isPICStyleGOT())
3549 MIB.
addReg(X86::EBX, RegState::Implicit);
3551 if (Is64Bit && IsVarArg && !IsWin64)
3552 MIB.
addReg(X86::AL, RegState::Implicit);
3555 for (
auto Reg : OutRegs)
3559 unsigned NumBytesForCalleeToPop =
3561 TM.Options.GuaranteedTailCallOpt)
3564 unsigned AdjStackUp =
TII.getCallFrameDestroyOpcode();
3565 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(AdjStackUp))
3570 CCState CCRetInfo(CC, IsVarArg, *FuncInfo.MF, RVLocs,
3571 CLI.RetTy->getContext());
3572 CCRetInfo.AnalyzeCallResult(Ins,
RetCC_X86);
3575 Register ResultReg = FuncInfo.CreateRegs(CLI.RetTy);
3576 for (
unsigned i = 0; i != RVLocs.
size(); ++i) {
3577 CCValAssign &VA = RVLocs[i];
3583 if ((CopyVT == MVT::f32 || CopyVT == MVT::f64) &&
3584 ((Is64Bit || Ins[i].
Flags.isInReg()) && !Subtarget->hasSSE1())) {
3590 if ((SrcReg == X86::FP0 || SrcReg == X86::FP1) &&
3591 isScalarFPTypeInSSEReg(VA.
getValVT())) {
3593 CopyReg = createResultReg(&X86::RFP80RegClass);
3597 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
3598 TII.get(TargetOpcode::COPY), CopyReg).
addReg(SrcReg);
3606 unsigned Opc = ResVT == MVT::f32 ? X86::ST_Fp80m32 : X86::ST_Fp80m64;
3612 Opc = ResVT == MVT::f32 ? X86::MOVSSrm_alt : X86::MOVSDrm_alt;
3614 TII.get(
Opc), ResultReg + i), FI);
3618 CLI.ResultReg = ResultReg;
3619 CLI.NumResultRegs = RVLocs.
size();
3623 if (TM.Options.EmitCallGraphSection && CB && CB->isIndirectCall()) {
3624 MachineFunction::CallSiteInfo CSInfo(*CB);
3632X86FastISel::fastSelectInstruction(
const Instruction *
I) {
3633 switch (
I->getOpcode()) {
3635 case Instruction::Load:
3636 return X86SelectLoad(
I);
3637 case Instruction::Store:
3638 return X86SelectStore(
I);
3639 case Instruction::Ret:
3640 return X86SelectRet(
I);
3641 case Instruction::ICmp:
3642 case Instruction::FCmp:
3643 return X86SelectCmp(
I);
3644 case Instruction::ZExt:
3645 return X86SelectZExt(
I);
3646 case Instruction::SExt:
3647 return X86SelectSExt(
I);
3648 case Instruction::Br:
3649 return X86SelectBranch(
I);
3650 case Instruction::LShr:
3651 case Instruction::AShr:
3652 case Instruction::Shl:
3653 return X86SelectShift(
I);
3654 case Instruction::SDiv:
3655 case Instruction::UDiv:
3656 case Instruction::SRem:
3657 case Instruction::URem:
3658 return X86SelectDivRem(
I);
3659 case Instruction::Select:
3660 return X86SelectSelect(
I);
3661 case Instruction::Trunc:
3662 return X86SelectTrunc(
I);
3663 case Instruction::FPExt:
3664 return X86SelectFPExt(
I);
3665 case Instruction::FPTrunc:
3666 return X86SelectFPTrunc(
I);
3667 case Instruction::SIToFP:
3668 return X86SelectSIToFP(
I);
3669 case Instruction::UIToFP:
3670 return X86SelectUIToFP(
I);
3671 case Instruction::IntToPtr:
3672 case Instruction::PtrToInt: {
3673 EVT SrcVT = TLI.getValueType(
DL,
I->getOperand(0)->getType());
3674 EVT DstVT = TLI.getValueType(
DL,
I->getType());
3676 return X86SelectZExt(
I);
3678 return X86SelectTrunc(
I);
3682 updateValueMap(
I,
Reg);
3685 case Instruction::BitCast:
3686 return X86SelectBitCast(
I);
3692Register X86FastISel::X86MaterializeInt(
const ConstantInt *CI, MVT VT) {
3698 Register SrcReg = fastEmitInst_(X86::MOV32r0, &X86::GR32RegClass);
3703 return fastEmitInst_extractsubreg(MVT::i8, SrcReg, X86::sub_8bit);
3705 return fastEmitInst_extractsubreg(MVT::i16, SrcReg, X86::sub_16bit);
3709 Register ResultReg = createResultReg(&X86::GR64RegClass);
3710 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
3711 TII.get(TargetOpcode::SUBREG_TO_REG), ResultReg)
3725 case MVT::i8:
Opc = X86::MOV8ri;
break;
3726 case MVT::i16:
Opc = X86::MOV16ri;
break;
3727 case MVT::i32:
Opc = X86::MOV32ri;
break;
3730 Opc = X86::MOV32ri64;
3732 Opc = X86::MOV64ri32;
3738 return fastEmitInst_i(
Opc, TLI.getRegClassFor(VT), Imm);
3741Register X86FastISel::X86MaterializeFP(
const ConstantFP *CFP, MVT VT) {
3743 return fastMaterializeFloatZero(CFP);
3753 bool HasSSE1 = Subtarget->hasSSE1();
3754 bool HasSSE2 = Subtarget->hasSSE2();
3755 bool HasAVX = Subtarget->hasAVX();
3756 bool HasAVX512 = Subtarget->hasAVX512();
3761 Opc = HasAVX512 ? X86::VMOVSSZrm_alt
3762 : HasAVX ? X86::VMOVSSrm_alt
3763 : HasSSE1 ? X86::MOVSSrm_alt
3767 Opc = HasAVX512 ? X86::VMOVSDZrm_alt
3768 : HasAVX ? X86::VMOVSDrm_alt
3769 : HasSSE2 ? X86::MOVSDrm_alt
3782 unsigned char OpFlag = Subtarget->classifyLocalReference(
nullptr);
3784 PICBase = getInstrInfo()->getGlobalBaseReg(FuncInfo.MF);
3786 PICBase = getInstrInfo()->getGlobalBaseReg(FuncInfo.MF);
3791 unsigned CPI = MCP.getConstantPoolIndex(CFP, Alignment);
3796 Register AddrReg = createResultReg(&X86::GR64RegClass);
3797 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOV64ri),
3800 MachineInstrBuilder MIB =
BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
3801 TII.get(
Opc), ResultReg);
3802 addRegReg(MIB, AddrReg,
false, X86::NoSubRegister, PICBase,
false,
3803 X86::NoSubRegister);
3804 MachineMemOperand *MMO = FuncInfo.MF->getMachineMemOperand(
3812 TII.get(
Opc), ResultReg),
3813 CPI, PICBase, OpFlag);
3817Register X86FastISel::X86MaterializeGV(
const GlobalValue *GV, MVT VT) {
3822 if (TM.isLargeGlobalValue(GV))
3834 Register ResultReg = createResultReg(TLI.getRegClassFor(VT));
3836 TLI.getPointerTy(
DL) == MVT::i64) {
3839 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(X86::MOV64ri),
3844 TLI.getPointerTy(
DL) == MVT::i32
3845 ? (Subtarget->isTarget64BitILP32() ? X86::LEA64_32r : X86::LEA32r)
3848 TII.get(
Opc), ResultReg), AM);
3855Register X86FastISel::fastMaterializeConstant(
const Constant *
C) {
3856 EVT CEVT = TLI.getValueType(
DL,
C->getType(),
true);
3864 return X86MaterializeInt(CI, VT);
3866 return X86MaterializeFP(CFP, VT);
3868 return X86MaterializeGV(GV, VT);
3875 if (!Subtarget->hasSSE1())
3876 Opc = X86::LD_Fp032;
3879 if (!Subtarget->hasSSE2())
3880 Opc = X86::LD_Fp064;
3883 Opc = X86::LD_Fp080;
3888 Register ResultReg = createResultReg(TLI.getRegClassFor(VT));
3889 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(
Opc),
3898Register X86FastISel::fastMaterializeAlloca(
const AllocaInst *
C) {
3906 if (!FuncInfo.StaticAllocaMap.count(
C))
3908 assert(
C->isStaticAlloca() &&
"dynamic alloca in the static alloca map?");
3914 TLI.getPointerTy(
DL) == MVT::i32
3915 ? (Subtarget->isTarget64BitILP32() ? X86::LEA64_32r : X86::LEA32r)
3917 const TargetRegisterClass *RC = TLI.getRegClassFor(TLI.getPointerTy(
DL));
3918 Register ResultReg = createResultReg(RC);
3920 TII.get(
Opc), ResultReg), AM);
3924Register X86FastISel::fastMaterializeFloatZero(
const ConstantFP *CF) {
3926 if (!isTypeLegal(CF->
getType(), VT))
3930 bool HasSSE1 = Subtarget->hasSSE1();
3931 bool HasSSE2 = Subtarget->hasSSE2();
3932 bool HasAVX512 = Subtarget->hasAVX512();
3937 Opc = HasAVX512 ? X86::AVX512_FsFLD0SH : X86::FsFLD0SH;
3940 Opc = HasAVX512 ? X86::AVX512_FsFLD0SS
3941 : HasSSE1 ? X86::FsFLD0SS
3945 Opc = HasAVX512 ? X86::AVX512_FsFLD0SD
3946 : HasSSE2 ? X86::FsFLD0SD
3954 Register ResultReg = createResultReg(TLI.getRegClassFor(VT));
3955 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(
Opc), ResultReg);
3959bool X86FastISel::tryToFoldLoadIntoMI(MachineInstr *
MI,
unsigned OpNo,
3960 const LoadInst *LI) {
3966 const X86InstrInfo &XII = (
const X86InstrInfo &)
TII;
3974 *FuncInfo.MF, *
MI, OpNo, AddrOps, FuncInfo.InsertPt,
Size, LI->
getAlign(),
3984 unsigned OperandNo = 0;
3986 E =
Result->operands_end();
I !=
E; ++
I, ++OperandNo) {
3987 MachineOperand &MO = *
I;
3993 if (IndexReg == MO.
getReg())
3999 FuncInfo.MF->moveAdditionalCallInfo(
MI, Result);
4000 Result->addMemOperand(*FuncInfo.MF, createMachineMemOperandFor(LI));
4001 Result->cloneInstrSymbols(*FuncInfo.MF, *
MI);
4003 removeDeadCode(
I, std::next(
I));
4007Register X86FastISel::fastEmitInst_rrrr(
unsigned MachineInstOpcode,
4008 const TargetRegisterClass *RC,
4011 const MCInstrDesc &
II =
TII.get(MachineInstOpcode);
4013 Register ResultReg = createResultReg(RC);
4019 if (
II.getNumDefs() >= 1)
4020 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
II, ResultReg)
4026 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
II)
4031 BuildMI(*FuncInfo.MBB, FuncInfo.InsertPt, MIMD,
TII.get(TargetOpcode::COPY),
4042 return new X86FastISel(funcInfo, libInfo, libcallLowering);
unsigned const MachineRegisterInfo * MRI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< ErlangGC > A("erlang", "erlang-compatible garbage collector")
static GCRegistry::Add< CoreCLRGC > E("coreclr", "CoreCLR-compatible GC")
This file defines the FastISel class.
const HexagonInstrInfo * TII
Module.h This file contains the declarations for the Module class.
const size_t AbstractManglingParser< Derived, Alloc >::NumOps
This file declares the MachineConstantPool class which is an abstract constant pool to keep track of ...
Register const TargetRegisterInfo * TRI
Promote Memory to Register
uint64_t IntrinsicInst * II
const SmallVectorImpl< MachineOperand > & Cond
#define GET_EGPR_IF_ENABLED(OPC)
static unsigned X86ChooseCmpImmediateOpcode(EVT VT, const ConstantInt *RHSC)
If we have a comparison with RHS as the RHS of the comparison, return an opcode that works for the co...
static std::pair< unsigned, bool > getX86SSEConditionCode(CmpInst::Predicate Predicate)
static unsigned computeBytesPoppedByCalleeForSRet(const X86Subtarget *Subtarget, CallingConv::ID CC, const CallBase *CB)
static unsigned X86ChooseCmpOpcode(EVT VT, const X86Subtarget *Subtarget)
static bool X86SelectAddress(MachineInstr &I, const X86TargetMachine &TM, const MachineRegisterInfo &MRI, const X86Subtarget &STI, X86AddressMode &AM)
LLVM_ABI APInt zext(unsigned width) const
Zero extend to a new width.
LLVM_ABI APInt sext(unsigned width) const
Sign extend to a new width.
InstListType::const_iterator const_iterator
BasicBlock * getSuccessor(unsigned i) const
Value * getCondition() const
Register getLocReg() const
LocInfo getLocInfo() const
int64_t getLocMemOffset() const
unsigned getValNo() const
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
LLVM_ABI bool paramHasAttr(unsigned ArgNo, Attribute::AttrKind Kind) const
Determine whether the argument or parameter has the given attribute.
This class is the base class for the comparison instructions.
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
@ FCMP_OEQ
0 0 0 1 True if ordered and equal
@ FCMP_TRUE
1 1 1 1 Always true (always folded)
@ FCMP_OLT
0 1 0 0 True if ordered and less than
@ FCMP_ULE
1 1 0 1 True if unordered, less than, or equal
@ FCMP_OGT
0 0 1 0 True if ordered and greater than
@ FCMP_OGE
0 0 1 1 True if ordered and greater than or equal
@ FCMP_ULT
1 1 0 0 True if unordered or less than
@ FCMP_ONE
0 1 1 0 True if ordered and operands are unequal
@ FCMP_UEQ
1 0 0 1 True if unordered or equal
@ FCMP_UGT
1 0 1 0 True if unordered or greater than
@ FCMP_OLE
0 1 0 1 True if ordered and less than or equal
@ FCMP_ORD
0 1 1 1 True if ordered (no nans)
@ FCMP_UNE
1 1 1 0 True if unordered or not equal
@ FCMP_UGE
1 0 1 1 True if unordered, greater than, or equal
@ FCMP_FALSE
0 0 0 0 Always false (always folded)
@ FCMP_UNO
1 0 0 0 True if unordered: isnan(X) | isnan(Y)
Predicate getInversePredicate() const
For example, EQ -> NE, UGT -> ULE, SLT -> SGE, OEQ -> UNE, UGT -> OLE, OLT -> UGE,...
This is the shared class of boolean and integer constants.
bool isOne() const
This is just a convenience method to make client code smaller for a common case.
int64_t getSExtValue() const
Return the constant as a 64-bit integer value after it has been sign extended as appropriate for the ...
unsigned getBitWidth() const
getBitWidth - Return the scalar bitwidth of this constant.
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
const APInt & getValue() const
Return the constant as an APInt value reference.
static LLVM_ABI Constant * getNullValue(Type *Ty)
Constructor to create a '0' constant of arbitrary type.
LLVM_ABI bool isNullValue() const
Return true if this is the value that would be returned by getNullValue.
bool isValidLocationForIntrinsic(const DILocation *DL) const
Check that a location is valid for this variable.
Value * getAddress() const
DILocalVariable * getVariable() const
DIExpression * getExpression() const
This is a fast-path instruction selection class that generates poor code and doesn't support illegal ...
FunctionLoweringInfo - This contains information that is global to a function that is used when lower...
Module * getParent()
Get the module that this global value is contained inside of...
LLVM_ABI bool isAtomic() const LLVM_READONLY
Return true if this instruction has an AtomicOrdering of unordered or higher.
Tracks which library functions to use for a particular subtarget.
Value * getPointerOperand()
Align getAlign() const
Return the alignment of the access that is being performed.
bool usesWindowsCFI() const
bool isVector() const
Return true if this is a vector value type.
bool isInteger() const
Return true if this is an integer or a vector integer type.
TypeSize getSizeInBits() const
Returns the size of the specified MVT in bits.
TypeSize getStoreSize() const
Return the number of bytes overwritten by a store of the specified value type.
MVT getVectorElementType() const
MachineInstrBundleIterator< MachineInstr > iterator
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
void setFrameAddressIsTaken(bool T)
void setStackProtectorIndex(int I)
const TargetSubtargetInfo & getSubtarget() const
getSubtarget - Return the subtarget for which this machine code is being compiled.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
Function & getFunction()
Return the LLVM function that this machine code represents.
void addCallSiteInfo(const MachineInstr *CallI, CallSiteInfo &&CallInfo)
Start tracking the arguments passed to the call CallI.
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
const MachineInstrBuilder & addReg(Register RegNo, RegState Flags={}, unsigned SubReg=0) const
Add a new virtual register operand.
const MachineInstrBuilder & addImm(int64_t Val) const
Add a new immediate operand.
const MachineInstrBuilder & addMetadata(const MDNode *MD) const
const MachineInstrBuilder & addSym(MCSymbol *Sym, unsigned char TargetFlags=0) const
const MachineInstrBuilder & addConstantPoolIndex(unsigned Idx, int Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addRegMask(const uint32_t *Mask) const
const MachineInstrBuilder & addGlobalAddress(const GlobalValue *GV, int64_t Offset=0, unsigned TargetFlags=0) const
const MachineInstrBuilder & addMBB(MachineBasicBlock *MBB, unsigned TargetFlags=0) const
unsigned getNumOperands() const
Retuns the total number of operands.
const MCInstrDesc & getDesc() const
Returns the target instruction descriptor of this MachineInstr.
MachineOperand * mop_iterator
iterator/begin/end - Iterate over all operands of a machine instruction.
LLVM_ABI void addMemOperand(MachineFunction &MF, MachineMemOperand *MO)
Add a MachineMemOperand to the machine instruction.
bool isNonTemporal() const
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
bool isReg() const
isReg - Tests if this is a MO_Register operand.
LLVM_ABI void setReg(Register Reg)
Change the register this operand corresponds to.
Register getReg() const
getReg - Returns the register number.
Value * getLength() const
Value * getRawDest() const
unsigned getDestAddressSpace() const
Value * getRawSource() const
Return the arguments to the instruction.
unsigned getSourceAddressSpace() const
Metadata * getModuleFlag(StringRef Key) const
Return the corresponding value if Key appears in module flags, otherwise return null.
Wrapper class representing virtual and physical registers.
void push_back(const T &Elt)
Value * getValueOperand()
Value * getPointerOperand()
TypeSize getElementOffset(unsigned Idx) const
Provides information about what library functions are available for the current target.
const MCAsmInfo * getMCAsmInfo() const
Return target specific asm information.
bool contains(Register Reg) const
Return true if the specified register is included in this register class.
bool isOSMSVCRT() const
Is this a "Windows" OS targeting a "MSVCRT.dll" environment.
bool isVectorTy() const
True if this is an instance of VectorType.
bool isArrayTy() const
True if this is an instance of ArrayType.
bool isStructTy() const
True if this is an instance of StructType.
LLVM_ABI unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
bool isIntegerTy() const
True if this is an instance of IntegerType.
const Use * const_op_iterator
Value * getOperand(unsigned i) const
unsigned getNumOperands() const
LLVM Value Representation.
Type * getType() const
All values are typed, get the type of this value.
bool hasOneUse() const
Return true if there is exactly one use of this value.
LLVMContext & getContext() const
All values hold a context through their type.
MachineInstr * foldMemoryOperandImpl(MachineFunction &MF, MachineInstr &MI, ArrayRef< unsigned > Ops, MachineBasicBlock::iterator InsertPt, int FrameIndex, LiveIntervals *LIS=nullptr, VirtRegMap *VRM=nullptr) const override
Fold a load or store of the specified stack slot into the specified machine instruction for the speci...
Register getSRetReturnReg() const
unsigned getBytesToPopOnReturn() const
Register getPtrSizedFrameRegister(const MachineFunction &MF) const
Register getStackRegister() const
const Triple & getTargetTriple() const
StructType * getStructTypeOrNull() const
TypeSize getSequentialElementStride(const DataLayout &DL) const
const ParentTy * getParent() const
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
constexpr char Align[]
Key for Kernel::Arg::Metadata::mAlign.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
@ HiPE
Used by the High-Performance Erlang Compiler (HiPE).
@ GHC
Used by the Glasgow Haskell Compiler (GHC).
@ Fast
Attempts to make calls as fast as possible (e.g.
@ Tail
Attemps to make calls as fast as possible while guaranteeing that tail call optimization can always b...
@ SwiftTail
This follows the Swift calling convention in how arguments are passed but guarantees tail calls will ...
@ C
The default llvm calling convention, compatible with C.
NodeType
ISD::NodeType enum - This enum defines the target-independent operators for a SelectionDAG.
@ ADD
Simple integer binary arithmetic operators.
@ ANY_EXTEND
ANY_EXTEND - Used for integer types. The high bits are undefined.
@ BITCAST
BITCAST - This operator converts between integer, vector and FP values, as if the value was stored to...
@ SIGN_EXTEND
Conversion operators.
@ BasicBlock
Various leaf nodes.
@ ZERO_EXTEND
ZERO_EXTEND - Used for integer types, zeroing the new bits.
@ AND
Bitwise operators - logical and, logical or, logical xor.
CondCode
ISD::CondCode enum - These are ordered carefully to make the bitfields below work out,...
Flag
These should be considered private to the implementation of the MCInstrDesc class.
Predicate
Predicate - These are "(BI << 5) | BO" for various predicates.
@ X86
Windows x64, Windows Itanium (IA-64)
@ MO_GOTPCREL_NORELAX
MO_GOTPCREL_NORELAX - Same as MO_GOTPCREL except that R_X86_64_GOTPCREL relocations are guaranteed to...
@ MO_GOTOFF
MO_GOTOFF - On a symbol operand this indicates that the immediate is the offset to the location of th...
@ MO_COFFSTUB
MO_COFFSTUB - On a symbol operand "FOO", this indicates that the reference is actually to the "....
@ MO_PLT
MO_PLT - On a symbol operand this indicates that the immediate is offset to the PLT entry of symbol n...
@ MO_NO_FLAG
MO_NO_FLAG - No flag for the operand.
@ MO_DLLIMPORT
MO_DLLIMPORT - On a symbol operand "FOO", this indicates that the reference is actually to the "__imp...
@ MO_PIC_BASE_OFFSET
MO_PIC_BASE_OFFSET - On a symbol operand this indicates that the immediate should get the value of th...
@ MO_GOTPCREL
MO_GOTPCREL - On a symbol operand this indicates that the immediate is offset to the GOT entry for th...
FastISel * createFastISel(FunctionLoweringInfo &funcInfo, const TargetLibraryInfo *libInfo, const LibcallLoweringInfo *libcallLowering)
std::pair< CondCode, bool > getX86ConditionCode(CmpInst::Predicate Predicate)
Return a pair of condition code for the given predicate and whether the instruction operands should b...
bool isCalleePop(CallingConv::ID CallingConv, bool is64Bit, bool IsVarArg, bool GuaranteeTCO)
Determines whether the callee is required to pop its own arguments.
unsigned getCMovOpcode(unsigned RegBytes, bool HasMemoryOperand=false, bool HasNDD=false)
Return a cmov opcode for the given register size in bytes, and operand type.
StringMapEntry< std::atomic< TypeEntryBody * > > TypeEntry
@ User
could "use" a pointer
@ Emitted
Assigned address, still materializing.
friend class Instruction
Iterator for Instructions in a `BasicBlock.
This is an optimization pass for GlobalISel generic memory operations.
static bool isGlobalStubReference(unsigned char TargetFlag)
isGlobalStubReference - Return true if the specified TargetFlag operand is a reference to a stub for ...
FunctionAddr VTableAddr Value
static bool isGlobalRelativeToPICBase(unsigned char TargetFlag)
isGlobalRelativeToPICBase - Return true if the specified global value reference is relative to a 32-b...
LLVM_ABI Register constrainOperandRegClass(const MachineFunction &MF, const TargetRegisterInfo &TRI, MachineRegisterInfo &MRI, const TargetInstrInfo &TII, const RegisterBankInfo &RBI, MachineInstr &InsertPt, const TargetRegisterClass &RegClass, MachineOperand &RegMO)
Constrain the Register operand OpIdx, so that it is now constrained to the TargetRegisterClass passed...
LLVM_ABI void GetReturnInfo(CallingConv::ID CC, Type *ReturnType, AttributeList attr, SmallVectorImpl< ISD::OutputArg > &Outs, const TargetLowering &TLI, const DataLayout &DL)
Given an LLVM IR type and return type attributes, compute the return value EVTs and flags,...
MachineInstrBuilder BuildMI(MachineFunction &MF, const MIMetadata &MIMD, const MCInstrDesc &MCID)
Builder interface. Specify how to create the initial instruction itself.
constexpr bool isInt(int64_t x)
Checks if an integer fits into the given bit width.
constexpr RegState getKillRegState(bool B)
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
auto successors(const MachineBasicBlock *BB)
static const MachineInstrBuilder & addConstantPoolReference(const MachineInstrBuilder &MIB, unsigned CPI, Register GlobalBaseReg, unsigned char OpFlags)
addConstantPoolReference - This function is used to add a reference to the base of a constant value s...
static const MachineInstrBuilder & addRegReg(const MachineInstrBuilder &MIB, Register Reg1, bool isKill1, unsigned SubReg1, Register Reg2, bool isKill2, unsigned SubReg2)
addRegReg - This function is used to add a memory reference of the form: [Reg + Reg].
static const MachineInstrBuilder & addFrameReference(const MachineInstrBuilder &MIB, int FI, int Offset=0, bool mem=true)
addFrameReference - This function is used to add a reference to the base of an abstract object on the...
static const MachineInstrBuilder & addFullAddress(const MachineInstrBuilder &MIB, const X86AddressMode &AM)
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
auto reverse(ContainerTy &&C)
LLVM_ABI void report_fatal_error(Error Err, bool gen_crash_diag=true)
generic_gep_type_iterator<> gep_type_iterator
constexpr bool isUInt(uint64_t x)
Checks if an unsigned integer fits into the given bit width.
class LLVM_GSL_OWNER SmallVector
Forward declaration of SmallVector so that calculateSmallVectorDefaultInlinedElements can reference s...
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
uint16_t MCPhysReg
An unsigned integer type large enough to represent all physical registers, but not necessarily virtua...
DWARFExpression::Operation Op
bool CC_X86(unsigned ValNo, MVT ValVT, MVT LocVT, CCValAssign::LocInfo LocInfo, ISD::ArgFlagsTy ArgFlags, Type *OrigTy, CCState &State)
bool RetCC_X86(unsigned ValNo, MVT ValVT, MVT LocVT, CCValAssign::LocInfo LocInfo, ISD::ArgFlagsTy ArgFlags, Type *OrigTy, CCState &State)
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
gep_type_iterator gep_type_begin(const User *GEP)
static const MachineInstrBuilder & addDirectMem(const MachineInstrBuilder &MIB, Register Reg)
addDirectMem - This function is used to add a direct memory reference to the current instruction – th...
void swap(llvm::BitVector &LHS, llvm::BitVector &RHS)
Implement std::swap in terms of BitVector swap.
constexpr uint64_t value() const
This is a hole in the type system and should not be abused.
bool isSimple() const
Test if the given EVT is simple (as opposed to being extended).
bool bitsGT(EVT VT) const
Return true if this has more bits than VT.
bool bitsLT(EVT VT) const
Return true if this has less bits than VT.
TypeSize getSizeInBits() const
Return the size of the specified value type in bits.
MVT getSimpleVT() const
Return the SimpleValueType held in the specified simple EVT.
static LLVM_ABI MachinePointerInfo getStack(MachineFunction &MF, int64_t Offset, uint8_t ID=0)
Stack pointer relative access.
static LLVM_ABI MachinePointerInfo getConstantPool(MachineFunction &MF)
Return a MachinePointerInfo record that refers to the constant pool.
X86AddressMode - This struct holds a generalized full x86 address mode.
void getFullAddress(SmallVectorImpl< MachineOperand > &MO)
union llvm::X86AddressMode::BaseUnion Base
enum llvm::X86AddressMode::@202116273335065351270200035056227005202106004277 BaseType