LLVM 22.0.0git
CallLowering.cpp
Go to the documentation of this file.
1//===-- lib/CodeGen/GlobalISel/CallLowering.cpp - Call lowering -----------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8///
9/// \file
10/// This file implements some simple delegations needed for call lowering.
11///
12//===----------------------------------------------------------------------===//
13
23#include "llvm/IR/DataLayout.h"
24#include "llvm/IR/LLVMContext.h"
25#include "llvm/IR/Module.h"
27
28#define DEBUG_TYPE "call-lowering"
29
30using namespace llvm;
31
32void CallLowering::anchor() {}
33
34/// Helper function which updates \p Flags when \p AttrFn returns true.
35static void
37 const std::function<bool(Attribute::AttrKind)> &AttrFn) {
38 // TODO: There are missing flags. Add them here.
39 if (AttrFn(Attribute::SExt))
40 Flags.setSExt();
41 if (AttrFn(Attribute::ZExt))
42 Flags.setZExt();
43 if (AttrFn(Attribute::InReg))
44 Flags.setInReg();
45 if (AttrFn(Attribute::StructRet))
46 Flags.setSRet();
47 if (AttrFn(Attribute::Nest))
48 Flags.setNest();
49 if (AttrFn(Attribute::ByVal))
50 Flags.setByVal();
51 if (AttrFn(Attribute::ByRef))
52 Flags.setByRef();
53 if (AttrFn(Attribute::Preallocated))
54 Flags.setPreallocated();
55 if (AttrFn(Attribute::InAlloca))
56 Flags.setInAlloca();
57 if (AttrFn(Attribute::Returned))
58 Flags.setReturned();
59 if (AttrFn(Attribute::SwiftSelf))
60 Flags.setSwiftSelf();
61 if (AttrFn(Attribute::SwiftAsync))
62 Flags.setSwiftAsync();
63 if (AttrFn(Attribute::SwiftError))
64 Flags.setSwiftError();
65}
66
68 unsigned ArgIdx) const {
69 ISD::ArgFlagsTy Flags;
70 addFlagsUsingAttrFn(Flags, [&Call, &ArgIdx](Attribute::AttrKind Attr) {
71 return Call.paramHasAttr(ArgIdx, Attr);
72 });
73 return Flags;
74}
75
78 ISD::ArgFlagsTy Flags;
80 return Call.hasRetAttr(Attr);
81 });
82 return Flags;
83}
84
86 const AttributeList &Attrs,
87 unsigned OpIdx) const {
88 addFlagsUsingAttrFn(Flags, [&Attrs, &OpIdx](Attribute::AttrKind Attr) {
89 return Attrs.hasAttributeAtIndex(OpIdx, Attr);
90 });
91}
92
94 ArrayRef<Register> ResRegs,
96 Register SwiftErrorVReg,
97 std::optional<PtrAuthInfo> PAI,
98 Register ConvergenceCtrlToken,
99 std::function<Register()> GetCalleeReg) const {
100 CallLoweringInfo Info;
101 const DataLayout &DL = MIRBuilder.getDataLayout();
102 MachineFunction &MF = MIRBuilder.getMF();
104 bool CanBeTailCalled = CB.isTailCall() &&
106 (MF.getFunction()
107 .getFnAttribute("disable-tail-calls")
108 .getValueAsString() != "true");
109
110 CallingConv::ID CallConv = CB.getCallingConv();
111 Type *RetTy = CB.getType();
112 bool IsVarArg = CB.getFunctionType()->isVarArg();
113
115 getReturnInfo(CallConv, RetTy, CB.getAttributes(), SplitArgs, DL);
116 Info.CanLowerReturn = canLowerReturn(MF, CallConv, SplitArgs, IsVarArg);
117
118 Info.IsConvergent = CB.isConvergent();
119
120 if (!Info.CanLowerReturn) {
121 // Callee requires sret demotion.
122 insertSRetOutgoingArgument(MIRBuilder, CB, Info);
123
124 // The sret demotion isn't compatible with tail-calls, since the sret
125 // argument points into the caller's stack frame.
126 CanBeTailCalled = false;
127 }
128
129 // First step is to marshall all the function's parameters into the correct
130 // physregs and memory locations. Gather the sequence of argument types that
131 // we'll pass to the assigner function.
132 unsigned i = 0;
133 unsigned NumFixedArgs = CB.getFunctionType()->getNumParams();
134 for (const auto &Arg : CB.args()) {
135 ArgInfo OrigArg{ArgRegs[i], *Arg.get(), i, getAttributesForArgIdx(CB, i)};
136 setArgFlags(OrigArg, i + AttributeList::FirstArgIndex, DL, CB);
137 if (i >= NumFixedArgs)
138 OrigArg.Flags[0].setVarArg();
139
140 // If we have an explicit sret argument that is an Instruction, (i.e., it
141 // might point to function-local memory), we can't meaningfully tail-call.
142 if (OrigArg.Flags[0].isSRet() && isa<Instruction>(&Arg))
143 CanBeTailCalled = false;
144
145 Info.OrigArgs.push_back(OrigArg);
146 ++i;
147 }
148
149 // Try looking through a bitcast from one function type to another.
150 // Commonly happens with calls to objc_msgSend().
151 const Value *CalleeV = CB.getCalledOperand()->stripPointerCasts();
152
153 // If IRTranslator chose to drop the ptrauth info, we can turn this into
154 // a direct call.
156 CalleeV = cast<ConstantPtrAuth>(CalleeV)->getPointer();
157 assert(isa<Function>(CalleeV));
158 }
159
160 if (const Function *F = dyn_cast<Function>(CalleeV)) {
161 if (F->hasFnAttribute(Attribute::NonLazyBind)) {
162 LLT Ty = getLLTForType(*F->getType(), DL);
163 Register Reg = MIRBuilder.buildGlobalValue(Ty, F).getReg(0);
164 Info.Callee = MachineOperand::CreateReg(Reg, false);
165 } else {
166 Info.Callee = MachineOperand::CreateGA(F, 0);
167 }
168 } else if (isa<GlobalIFunc>(CalleeV) || isa<GlobalAlias>(CalleeV)) {
169 // IR IFuncs and Aliases can't be forward declared (only defined), so the
170 // callee must be in the same TU and therefore we can direct-call it without
171 // worrying about it being out of range.
172 Info.Callee = MachineOperand::CreateGA(cast<GlobalValue>(CalleeV), 0);
173 } else
174 Info.Callee = MachineOperand::CreateReg(GetCalleeReg(), false);
175
176 Register ReturnHintAlignReg;
177 Align ReturnHintAlign;
178
179 Info.OrigRet = ArgInfo{ResRegs, RetTy, 0, getAttributesForReturn(CB)};
180
181 if (!Info.OrigRet.Ty->isVoidTy()) {
182 setArgFlags(Info.OrigRet, AttributeList::ReturnIndex, DL, CB);
183
184 if (MaybeAlign Alignment = CB.getRetAlign()) {
185 if (*Alignment > Align(1)) {
186 ReturnHintAlignReg = MRI.cloneVirtualRegister(ResRegs[0]);
187 Info.OrigRet.Regs[0] = ReturnHintAlignReg;
188 ReturnHintAlign = *Alignment;
189 }
190 }
191 }
192
193 auto Bundle = CB.getOperandBundle(LLVMContext::OB_kcfi);
194 if (Bundle && CB.isIndirectCall()) {
195 Info.CFIType = cast<ConstantInt>(Bundle->Inputs[0]);
196 assert(Info.CFIType->getType()->isIntegerTy(32) && "Invalid CFI type");
197 }
198
200 Info.DeactivationSymbol = cast<GlobalValue>(Bundle->Inputs[0]);
201 }
202
203 Info.CB = &CB;
204 Info.KnownCallees = CB.getMetadata(LLVMContext::MD_callees);
205 Info.CallConv = CallConv;
206 Info.SwiftErrorVReg = SwiftErrorVReg;
207 Info.PAI = PAI;
208 Info.ConvergenceCtrlToken = ConvergenceCtrlToken;
209 Info.IsMustTailCall = CB.isMustTailCall();
210 Info.IsTailCall = CanBeTailCalled;
211 Info.IsVarArg = IsVarArg;
212 if (!lowerCall(MIRBuilder, Info))
213 return false;
214
215 if (ReturnHintAlignReg && !Info.LoweredTailCall) {
216 MIRBuilder.buildAssertAlign(ResRegs[0], ReturnHintAlignReg,
217 ReturnHintAlign);
218 }
219
220 return true;
221}
222
223template <typename FuncInfoTy>
225 const DataLayout &DL,
226 const FuncInfoTy &FuncInfo) const {
227 auto &Flags = Arg.Flags[0];
228 const AttributeList &Attrs = FuncInfo.getAttributes();
229 addArgFlagsFromAttributes(Flags, Attrs, OpIdx);
230
232 if (PtrTy) {
233 Flags.setPointer();
234 Flags.setPointerAddrSpace(PtrTy->getPointerAddressSpace());
235 }
236
237 Align MemAlign = DL.getABITypeAlign(Arg.Ty);
238 if (Flags.isByVal() || Flags.isInAlloca() || Flags.isPreallocated() ||
239 Flags.isByRef()) {
240 assert(OpIdx >= AttributeList::FirstArgIndex);
241 unsigned ParamIdx = OpIdx - AttributeList::FirstArgIndex;
242
243 Type *ElementTy = FuncInfo.getParamByValType(ParamIdx);
244 if (!ElementTy)
245 ElementTy = FuncInfo.getParamByRefType(ParamIdx);
246 if (!ElementTy)
247 ElementTy = FuncInfo.getParamInAllocaType(ParamIdx);
248 if (!ElementTy)
249 ElementTy = FuncInfo.getParamPreallocatedType(ParamIdx);
250
251 assert(ElementTy && "Must have byval, inalloca or preallocated type");
252
253 uint64_t MemSize = DL.getTypeAllocSize(ElementTy);
254 if (Flags.isByRef())
255 Flags.setByRefSize(MemSize);
256 else
257 Flags.setByValSize(MemSize);
258
259 // For ByVal, alignment should be passed from FE. BE will guess if
260 // this info is not there but there are cases it cannot get right.
261 if (auto ParamAlign = FuncInfo.getParamStackAlign(ParamIdx))
262 MemAlign = *ParamAlign;
263 else if ((ParamAlign = FuncInfo.getParamAlign(ParamIdx)))
264 MemAlign = *ParamAlign;
265 else
266 MemAlign = getTLI()->getByValTypeAlignment(ElementTy, DL);
267 } else if (OpIdx >= AttributeList::FirstArgIndex) {
268 if (auto ParamAlign =
269 FuncInfo.getParamStackAlign(OpIdx - AttributeList::FirstArgIndex))
270 MemAlign = *ParamAlign;
271 }
272 Flags.setMemAlign(MemAlign);
273 Flags.setOrigAlign(DL.getABITypeAlign(Arg.Ty));
274
275 // Don't try to use the returned attribute if the argument is marked as
276 // swiftself, since it won't be passed in x0.
277 if (Flags.isSwiftSelf())
278 Flags.setReturned(false);
279}
280
281template void
283 const DataLayout &DL,
284 const Function &FuncInfo) const;
285
286template void
288 const DataLayout &DL,
289 const CallBase &FuncInfo) const;
290
292 SmallVectorImpl<ArgInfo> &SplitArgs,
293 const DataLayout &DL,
294 CallingConv::ID CallConv,
295 SmallVectorImpl<uint64_t> *Offsets) const {
296 LLVMContext &Ctx = OrigArg.Ty->getContext();
297
298 SmallVector<EVT, 4> SplitVTs;
299 ComputeValueVTs(*TLI, DL, OrigArg.Ty, SplitVTs, /*MemVTs=*/nullptr, Offsets,
300 0);
301
302 if (SplitVTs.size() == 0)
303 return;
304
305 if (SplitVTs.size() == 1) {
306 // No splitting to do, but we want to replace the original type (e.g. [1 x
307 // double] -> double).
308 SplitArgs.emplace_back(OrigArg.Regs[0], SplitVTs[0].getTypeForEVT(Ctx),
309 OrigArg.OrigArgIndex, OrigArg.Flags[0],
310 OrigArg.OrigValue);
311 return;
312 }
313
314 // Create one ArgInfo for each virtual register in the original ArgInfo.
315 assert(OrigArg.Regs.size() == SplitVTs.size() && "Regs / types mismatch");
316
317 bool NeedsRegBlock = TLI->functionArgumentNeedsConsecutiveRegisters(
318 OrigArg.Ty, CallConv, false, DL);
319 for (unsigned i = 0, e = SplitVTs.size(); i < e; ++i) {
320 Type *SplitTy = SplitVTs[i].getTypeForEVT(Ctx);
321 SplitArgs.emplace_back(OrigArg.Regs[i], SplitTy, OrigArg.OrigArgIndex,
322 OrigArg.Flags[0]);
323 if (NeedsRegBlock)
324 SplitArgs.back().Flags[0].setInConsecutiveRegs();
325 }
326
327 SplitArgs.back().Flags[0].setInConsecutiveRegsLast();
328}
329
330/// Pack values \p SrcRegs to cover the vector type result \p DstRegs.
333 ArrayRef<Register> SrcRegs) {
334 MachineRegisterInfo &MRI = *B.getMRI();
335 LLT LLTy = MRI.getType(DstRegs[0]);
336 LLT PartLLT = MRI.getType(SrcRegs[0]);
337
338 // Deal with v3s16 split into v2s16
339 LLT LCMTy = getCoverTy(LLTy, PartLLT);
340 if (LCMTy == LLTy) {
341 // Common case where no padding is needed.
342 assert(DstRegs.size() == 1);
343 return B.buildConcatVectors(DstRegs[0], SrcRegs);
344 }
345
346 // We need to create an unmerge to the result registers, which may require
347 // widening the original value.
348 Register UnmergeSrcReg;
349 if (LCMTy != PartLLT) {
350 assert(DstRegs.size() == 1);
351 return B.buildDeleteTrailingVectorElements(
352 DstRegs[0], B.buildMergeLikeInstr(LCMTy, SrcRegs));
353 } else {
354 // We don't need to widen anything if we're extracting a scalar which was
355 // promoted to a vector e.g. s8 -> v4s8 -> s8
356 assert(SrcRegs.size() == 1);
357 UnmergeSrcReg = SrcRegs[0];
358 }
359
360 int NumDst = LCMTy.getSizeInBits() / LLTy.getSizeInBits();
361
362 SmallVector<Register, 8> PadDstRegs(NumDst);
363 llvm::copy(DstRegs, PadDstRegs.begin());
364
365 // Create the excess dead defs for the unmerge.
366 for (int I = DstRegs.size(); I != NumDst; ++I)
367 PadDstRegs[I] = MRI.createGenericVirtualRegister(LLTy);
368
369 if (PadDstRegs.size() == 1)
370 return B.buildDeleteTrailingVectorElements(DstRegs[0], UnmergeSrcReg);
371 return B.buildUnmerge(PadDstRegs, UnmergeSrcReg);
372}
373
374/// Create a sequence of instructions to combine pieces split into register
375/// typed values to the original IR value. \p OrigRegs contains the destination
376/// value registers of type \p LLTy, and \p Regs contains the legalized pieces
377/// with type \p PartLLT. This is used for incoming values (physregs to vregs).
379 ArrayRef<Register> Regs, LLT LLTy, LLT PartLLT,
380 const ISD::ArgFlagsTy Flags) {
381 MachineRegisterInfo &MRI = *B.getMRI();
382
383 if (PartLLT == LLTy) {
384 // We should have avoided introducing a new virtual register, and just
385 // directly assigned here.
386 assert(OrigRegs[0] == Regs[0]);
387 return;
388 }
389
390 if (PartLLT.getSizeInBits() == LLTy.getSizeInBits() && OrigRegs.size() == 1 &&
391 Regs.size() == 1) {
392 B.buildBitcast(OrigRegs[0], Regs[0]);
393 return;
394 }
395
396 // A vector PartLLT needs extending to LLTy's element size.
397 // E.g. <2 x s64> = G_SEXT <2 x s32>.
398 if (PartLLT.isVector() == LLTy.isVector() &&
399 PartLLT.getScalarSizeInBits() > LLTy.getScalarSizeInBits() &&
400 (!PartLLT.isVector() ||
401 PartLLT.getElementCount() == LLTy.getElementCount()) &&
402 OrigRegs.size() == 1 && Regs.size() == 1) {
403 Register SrcReg = Regs[0];
404
405 LLT LocTy = MRI.getType(SrcReg);
406
407 if (Flags.isSExt()) {
408 SrcReg = B.buildAssertSExt(LocTy, SrcReg, LLTy.getScalarSizeInBits())
409 .getReg(0);
410 } else if (Flags.isZExt()) {
411 SrcReg = B.buildAssertZExt(LocTy, SrcReg, LLTy.getScalarSizeInBits())
412 .getReg(0);
413 }
414
415 // Sometimes pointers are passed zero extended.
416 LLT OrigTy = MRI.getType(OrigRegs[0]);
417 if (OrigTy.isPointer()) {
418 LLT IntPtrTy = LLT::scalar(OrigTy.getSizeInBits());
419 B.buildIntToPtr(OrigRegs[0], B.buildTrunc(IntPtrTy, SrcReg));
420 return;
421 }
422
423 B.buildTrunc(OrigRegs[0], SrcReg);
424 return;
425 }
426
427 if (!LLTy.isVector() && !PartLLT.isVector()) {
428 assert(OrigRegs.size() == 1);
429 LLT OrigTy = MRI.getType(OrigRegs[0]);
430
431 unsigned SrcSize = PartLLT.getSizeInBits().getFixedValue() * Regs.size();
432 if (SrcSize == OrigTy.getSizeInBits())
433 B.buildMergeValues(OrigRegs[0], Regs);
434 else {
435 auto Widened = B.buildMergeLikeInstr(LLT::scalar(SrcSize), Regs);
436 B.buildTrunc(OrigRegs[0], Widened);
437 }
438
439 return;
440 }
441
442 if (PartLLT.isVector()) {
443 assert(OrigRegs.size() == 1);
444 SmallVector<Register> CastRegs(Regs);
445
446 // If PartLLT is a mismatched vector in both number of elements and element
447 // size, e.g. PartLLT == v2s64 and LLTy is v3s32, then first coerce it to
448 // have the same elt type, i.e. v4s32.
449 // TODO: Extend this coersion to element multiples other than just 2.
450 if (TypeSize::isKnownGT(PartLLT.getSizeInBits(), LLTy.getSizeInBits()) &&
451 PartLLT.getScalarSizeInBits() == LLTy.getScalarSizeInBits() * 2 &&
452 Regs.size() == 1) {
453 LLT NewTy = PartLLT.changeElementType(LLTy.getElementType())
454 .changeElementCount(PartLLT.getElementCount() * 2);
455 CastRegs[0] = B.buildBitcast(NewTy, Regs[0]).getReg(0);
456 PartLLT = NewTy;
457 }
458
459 if (LLTy.getScalarType() == PartLLT.getElementType()) {
460 mergeVectorRegsToResultRegs(B, OrigRegs, CastRegs);
461 } else {
462 unsigned I = 0;
463 LLT GCDTy = getGCDType(LLTy, PartLLT);
464
465 // We are both splitting a vector, and bitcasting its element types. Cast
466 // the source pieces into the appropriate number of pieces with the result
467 // element type.
468 for (Register SrcReg : CastRegs)
469 CastRegs[I++] = B.buildBitcast(GCDTy, SrcReg).getReg(0);
470 mergeVectorRegsToResultRegs(B, OrigRegs, CastRegs);
471 }
472
473 return;
474 }
475
476 assert(LLTy.isVector() && !PartLLT.isVector());
477
478 LLT DstEltTy = LLTy.getElementType();
479
480 // Pointer information was discarded. We'll need to coerce some register types
481 // to avoid violating type constraints.
482 LLT RealDstEltTy = MRI.getType(OrigRegs[0]).getElementType();
483
484 assert(DstEltTy.getSizeInBits() == RealDstEltTy.getSizeInBits());
485
486 if (DstEltTy == PartLLT) {
487 // Vector was trivially scalarized.
488
489 if (RealDstEltTy.isPointer()) {
490 for (Register Reg : Regs)
491 MRI.setType(Reg, RealDstEltTy);
492 }
493
494 B.buildBuildVector(OrigRegs[0], Regs);
495 } else if (DstEltTy.getSizeInBits() > PartLLT.getSizeInBits()) {
496 // Deal with vector with 64-bit elements decomposed to 32-bit
497 // registers. Need to create intermediate 64-bit elements.
498 SmallVector<Register, 8> EltMerges;
499 int PartsPerElt =
500 divideCeil(DstEltTy.getSizeInBits(), PartLLT.getSizeInBits());
501 LLT ExtendedPartTy = LLT::scalar(PartLLT.getSizeInBits() * PartsPerElt);
502
503 for (int I = 0, NumElts = LLTy.getNumElements(); I != NumElts; ++I) {
504 auto Merge =
505 B.buildMergeLikeInstr(ExtendedPartTy, Regs.take_front(PartsPerElt));
506 if (ExtendedPartTy.getSizeInBits() > RealDstEltTy.getSizeInBits())
507 Merge = B.buildTrunc(RealDstEltTy, Merge);
508 // Fix the type in case this is really a vector of pointers.
509 MRI.setType(Merge.getReg(0), RealDstEltTy);
510 EltMerges.push_back(Merge.getReg(0));
511 Regs = Regs.drop_front(PartsPerElt);
512 }
513
514 B.buildBuildVector(OrigRegs[0], EltMerges);
515 } else {
516 // Vector was split, and elements promoted to a wider type.
517 // FIXME: Should handle floating point promotions.
518 unsigned NumElts = LLTy.getNumElements();
519 LLT BVType = LLT::fixed_vector(NumElts, PartLLT);
520
521 Register BuildVec;
522 if (NumElts == Regs.size())
523 BuildVec = B.buildBuildVector(BVType, Regs).getReg(0);
524 else {
525 // Vector elements are packed in the inputs.
526 // e.g. we have a <4 x s16> but 2 x s32 in regs.
527 assert(NumElts > Regs.size());
528 LLT SrcEltTy = MRI.getType(Regs[0]);
529
530 LLT OriginalEltTy = MRI.getType(OrigRegs[0]).getElementType();
531
532 // Input registers contain packed elements.
533 // Determine how many elements per reg.
534 assert((SrcEltTy.getSizeInBits() % OriginalEltTy.getSizeInBits()) == 0);
535 unsigned EltPerReg =
536 (SrcEltTy.getSizeInBits() / OriginalEltTy.getSizeInBits());
537
539 BVRegs.reserve(Regs.size() * EltPerReg);
540 for (Register R : Regs) {
541 auto Unmerge = B.buildUnmerge(OriginalEltTy, R);
542 for (unsigned K = 0; K < EltPerReg; ++K)
543 BVRegs.push_back(B.buildAnyExt(PartLLT, Unmerge.getReg(K)).getReg(0));
544 }
545
546 // We may have some more elements in BVRegs, e.g. if we have 2 s32 pieces
547 // for a <3 x s16> vector. We should have less than EltPerReg extra items.
548 if (BVRegs.size() > NumElts) {
549 assert((BVRegs.size() - NumElts) < EltPerReg);
550 BVRegs.truncate(NumElts);
551 }
552 BuildVec = B.buildBuildVector(BVType, BVRegs).getReg(0);
553 }
554 B.buildTrunc(OrigRegs[0], BuildVec);
555 }
556}
557
558/// Create a sequence of instructions to expand the value in \p SrcReg (of type
559/// \p SrcTy) to the types in \p DstRegs (of type \p PartTy). \p ExtendOp should
560/// contain the type of scalar value extension if necessary.
561///
562/// This is used for outgoing values (vregs to physregs)
564 Register SrcReg, LLT SrcTy, LLT PartTy,
565 unsigned ExtendOp = TargetOpcode::G_ANYEXT) {
566 // We could just insert a regular copy, but this is unreachable at the moment.
567 assert(SrcTy != PartTy && "identical part types shouldn't reach here");
568
569 const TypeSize PartSize = PartTy.getSizeInBits();
570
571 if (PartTy.isVector() == SrcTy.isVector() &&
572 PartTy.getScalarSizeInBits() > SrcTy.getScalarSizeInBits()) {
573 assert(DstRegs.size() == 1);
574 B.buildInstr(ExtendOp, {DstRegs[0]}, {SrcReg});
575 return;
576 }
577
578 if (SrcTy.isVector() && !PartTy.isVector() &&
579 TypeSize::isKnownGT(PartSize, SrcTy.getElementType().getSizeInBits())) {
580 // Vector was scalarized, and the elements extended.
581 auto UnmergeToEltTy = B.buildUnmerge(SrcTy.getElementType(), SrcReg);
582 for (int i = 0, e = DstRegs.size(); i != e; ++i)
583 B.buildAnyExt(DstRegs[i], UnmergeToEltTy.getReg(i));
584 return;
585 }
586
587 if (SrcTy.isVector() && PartTy.isVector() &&
588 PartTy.getSizeInBits() == SrcTy.getSizeInBits() &&
589 ElementCount::isKnownLT(SrcTy.getElementCount(),
590 PartTy.getElementCount())) {
591 // A coercion like: v2f32 -> v4f32 or nxv2f32 -> nxv4f32
592 Register DstReg = DstRegs.front();
593 B.buildPadVectorWithUndefElements(DstReg, SrcReg);
594 return;
595 }
596
597 LLT GCDTy = getGCDType(SrcTy, PartTy);
598 if (GCDTy == PartTy) {
599 // If this already evenly divisible, we can create a simple unmerge.
600 B.buildUnmerge(DstRegs, SrcReg);
601 return;
602 }
603
604 if (SrcTy.isVector() && !PartTy.isVector() &&
605 SrcTy.getScalarSizeInBits() > PartTy.getSizeInBits()) {
606 LLT ExtTy =
607 LLT::vector(SrcTy.getElementCount(),
608 LLT::scalar(PartTy.getScalarSizeInBits() * DstRegs.size() /
609 SrcTy.getNumElements()));
610 auto Ext = B.buildAnyExt(ExtTy, SrcReg);
611 B.buildUnmerge(DstRegs, Ext);
612 return;
613 }
614
615 MachineRegisterInfo &MRI = *B.getMRI();
616 LLT DstTy = MRI.getType(DstRegs[0]);
617 LLT LCMTy = getCoverTy(SrcTy, PartTy);
618
619 if (PartTy.isVector() && LCMTy == PartTy) {
620 assert(DstRegs.size() == 1);
621 B.buildPadVectorWithUndefElements(DstRegs[0], SrcReg);
622 return;
623 }
624
625 const unsigned DstSize = DstTy.getSizeInBits();
626 const unsigned SrcSize = SrcTy.getSizeInBits();
627 unsigned CoveringSize = LCMTy.getSizeInBits();
628
629 Register UnmergeSrc = SrcReg;
630
631 if (!LCMTy.isVector() && CoveringSize != SrcSize) {
632 // For scalars, it's common to be able to use a simple extension.
633 if (SrcTy.isScalar() && DstTy.isScalar()) {
634 CoveringSize = alignTo(SrcSize, DstSize);
635 LLT CoverTy = LLT::scalar(CoveringSize);
636 UnmergeSrc = B.buildInstr(ExtendOp, {CoverTy}, {SrcReg}).getReg(0);
637 } else {
638 // Widen to the common type.
639 // FIXME: This should respect the extend type
640 Register Undef = B.buildUndef(SrcTy).getReg(0);
641 SmallVector<Register, 8> MergeParts(1, SrcReg);
642 for (unsigned Size = SrcSize; Size != CoveringSize; Size += SrcSize)
643 MergeParts.push_back(Undef);
644 UnmergeSrc = B.buildMergeLikeInstr(LCMTy, MergeParts).getReg(0);
645 }
646 }
647
648 if (LCMTy.isVector() && CoveringSize != SrcSize)
649 UnmergeSrc = B.buildPadVectorWithUndefElements(LCMTy, SrcReg).getReg(0);
650
651 B.buildUnmerge(DstRegs, UnmergeSrc);
652}
653
655 ValueHandler &Handler, ValueAssigner &Assigner,
657 CallingConv::ID CallConv, bool IsVarArg,
658 ArrayRef<Register> ThisReturnRegs) const {
659 MachineFunction &MF = MIRBuilder.getMF();
660 const Function &F = MF.getFunction();
662
663 CCState CCInfo(CallConv, IsVarArg, MF, ArgLocs, F.getContext());
664 if (!determineAssignments(Assigner, Args, CCInfo))
665 return false;
666
667 return handleAssignments(Handler, Args, CCInfo, ArgLocs, MIRBuilder,
668 ThisReturnRegs);
669}
670
672 if (Flags.isSExt())
673 return TargetOpcode::G_SEXT;
674 if (Flags.isZExt())
675 return TargetOpcode::G_ZEXT;
676 return TargetOpcode::G_ANYEXT;
677}
678
681 CCState &CCInfo) const {
682 LLVMContext &Ctx = CCInfo.getContext();
683 const CallingConv::ID CallConv = CCInfo.getCallingConv();
684
685 unsigned NumArgs = Args.size();
686 for (unsigned i = 0; i != NumArgs; ++i) {
687 EVT CurVT = EVT::getEVT(Args[i].Ty);
688
689 MVT NewVT = TLI->getRegisterTypeForCallingConv(Ctx, CallConv, CurVT);
690
691 // If we need to split the type over multiple regs, check it's a scenario
692 // we currently support.
693 unsigned NumParts =
694 TLI->getNumRegistersForCallingConv(Ctx, CallConv, CurVT);
695
696 if (NumParts == 1) {
697 // Try to use the register type if we couldn't assign the VT.
698 if (Assigner.assignArg(i, CurVT, NewVT, NewVT, CCValAssign::Full, Args[i],
699 Args[i].Flags[0], CCInfo))
700 return false;
701 continue;
702 }
703
704 // For incoming arguments (physregs to vregs), we could have values in
705 // physregs (or memlocs) which we want to extract and copy to vregs.
706 // During this, we might have to deal with the LLT being split across
707 // multiple regs, so we have to record this information for later.
708 //
709 // If we have outgoing args, then we have the opposite case. We have a
710 // vreg with an LLT which we want to assign to a physical location, and
711 // we might have to record that the value has to be split later.
712
713 // We're handling an incoming arg which is split over multiple regs.
714 // E.g. passing an s128 on AArch64.
715 ISD::ArgFlagsTy OrigFlags = Args[i].Flags[0];
716 Args[i].Flags.clear();
717
718 for (unsigned Part = 0; Part < NumParts; ++Part) {
719 ISD::ArgFlagsTy Flags = OrigFlags;
720 if (Part == 0) {
721 Flags.setSplit();
722 } else {
723 Flags.setOrigAlign(Align(1));
724 if (Part == NumParts - 1)
725 Flags.setSplitEnd();
726 }
727
728 Args[i].Flags.push_back(Flags);
729 if (Assigner.assignArg(i, CurVT, NewVT, NewVT, CCValAssign::Full, Args[i],
730 Args[i].Flags[Part], CCInfo)) {
731 // Still couldn't assign this smaller part type for some reason.
732 return false;
733 }
734 }
735 }
736
737 return true;
738}
739
742 CCState &CCInfo,
744 MachineIRBuilder &MIRBuilder,
745 ArrayRef<Register> ThisReturnRegs) const {
746 MachineFunction &MF = MIRBuilder.getMF();
748 const Function &F = MF.getFunction();
749 const DataLayout &DL = F.getDataLayout();
750
751 const unsigned NumArgs = Args.size();
752
753 // Stores thunks for outgoing register assignments. This is used so we delay
754 // generating register copies until mem loc assignments are done. We do this
755 // so that if the target is using the delayed stack protector feature, we can
756 // find the split point of the block accurately. E.g. if we have:
757 // G_STORE %val, %memloc
758 // $x0 = COPY %foo
759 // $x1 = COPY %bar
760 // CALL func
761 // ... then the split point for the block will correctly be at, and including,
762 // the copy to $x0. If instead the G_STORE instruction immediately precedes
763 // the CALL, then we'd prematurely choose the CALL as the split point, thus
764 // generating a split block with a CALL that uses undefined physregs.
765 SmallVector<std::function<void()>> DelayedOutgoingRegAssignments;
766
767 for (unsigned i = 0, j = 0; i != NumArgs; ++i, ++j) {
768 assert(j < ArgLocs.size() && "Skipped too many arg locs");
769 CCValAssign &VA = ArgLocs[j];
770 assert(VA.getValNo() == i && "Location doesn't correspond to current arg");
771
772 if (VA.needsCustom()) {
773 std::function<void()> Thunk;
774 unsigned NumArgRegs = Handler.assignCustomValue(
775 Args[i], ArrayRef(ArgLocs).slice(j), &Thunk);
776 if (Thunk)
777 DelayedOutgoingRegAssignments.emplace_back(Thunk);
778 if (!NumArgRegs)
779 return false;
780 j += (NumArgRegs - 1);
781 continue;
782 }
783
784 auto AllocaAddressSpace = MF.getDataLayout().getAllocaAddrSpace();
785
786 const MVT ValVT = VA.getValVT();
787 const MVT LocVT = VA.getLocVT();
788
789 const LLT LocTy(LocVT);
790 const LLT ValTy(ValVT);
791 const LLT NewLLT = Handler.isIncomingArgumentHandler() ? LocTy : ValTy;
792 const EVT OrigVT = EVT::getEVT(Args[i].Ty);
793 const LLT OrigTy = getLLTForType(*Args[i].Ty, DL);
794 const LLT PointerTy = LLT::pointer(
795 AllocaAddressSpace, DL.getPointerSizeInBits(AllocaAddressSpace));
796
797 // Expected to be multiple regs for a single incoming arg.
798 // There should be Regs.size() ArgLocs per argument.
799 // This should be the same as getNumRegistersForCallingConv
800 const unsigned NumParts = Args[i].Flags.size();
801
802 // Now split the registers into the assigned types.
803 Args[i].OrigRegs.assign(Args[i].Regs.begin(), Args[i].Regs.end());
804
805 if (NumParts != 1 || NewLLT != OrigTy) {
806 // If we can't directly assign the register, we need one or more
807 // intermediate values.
808 Args[i].Regs.resize(NumParts);
809
810 // When we have indirect parameter passing we are receiving a pointer,
811 // that points to the actual value, so we need one "temporary" pointer.
812 if (VA.getLocInfo() == CCValAssign::Indirect) {
813 if (Handler.isIncomingArgumentHandler())
814 Args[i].Regs[0] = MRI.createGenericVirtualRegister(PointerTy);
815 } else {
816 // For each split register, create and assign a vreg that will store
817 // the incoming component of the larger value. These will later be
818 // merged to form the final vreg.
819 for (unsigned Part = 0; Part < NumParts; ++Part)
820 Args[i].Regs[Part] = MRI.createGenericVirtualRegister(NewLLT);
821 }
822 }
823
824 assert((j + (NumParts - 1)) < ArgLocs.size() &&
825 "Too many regs for number of args");
826
827 // Coerce into outgoing value types before register assignment.
828 if (!Handler.isIncomingArgumentHandler() && OrigTy != ValTy &&
830 assert(Args[i].OrigRegs.size() == 1);
831 buildCopyToRegs(MIRBuilder, Args[i].Regs, Args[i].OrigRegs[0], OrigTy,
832 ValTy, extendOpFromFlags(Args[i].Flags[0]));
833 }
834
835 bool IndirectParameterPassingHandled = false;
836 bool BigEndianPartOrdering = TLI->hasBigEndianPartOrdering(OrigVT, DL);
837 for (unsigned Part = 0; Part < NumParts; ++Part) {
838 assert((VA.getLocInfo() != CCValAssign::Indirect || Part == 0) &&
839 "Only the first parameter should be processed when "
840 "handling indirect passing!");
841 Register ArgReg = Args[i].Regs[Part];
842 // There should be Regs.size() ArgLocs per argument.
843 unsigned Idx = BigEndianPartOrdering ? NumParts - 1 - Part : Part;
844 CCValAssign &VA = ArgLocs[j + Idx];
845 const ISD::ArgFlagsTy Flags = Args[i].Flags[Part];
846
847 // We found an indirect parameter passing, and we have an
848 // OutgoingValueHandler as our handler (so we are at the call site or the
849 // return value). In this case, start the construction of the following
850 // GMIR, that is responsible for the preparation of indirect parameter
851 // passing:
852 //
853 // %1(indirectly passed type) = The value to pass
854 // %3(pointer) = G_FRAME_INDEX %stack.0
855 // G_STORE %1, %3 :: (store (s128), align 8)
856 //
857 // After this GMIR, the remaining part of the loop body will decide how
858 // to get the value to the caller and we break out of the loop.
859 if (VA.getLocInfo() == CCValAssign::Indirect &&
860 !Handler.isIncomingArgumentHandler()) {
861 Align AlignmentForStored = DL.getPrefTypeAlign(Args[i].Ty);
862 MachineFrameInfo &MFI = MF.getFrameInfo();
863 // Get some space on the stack for the value, so later we can pass it
864 // as a reference.
865 int FrameIdx = MFI.CreateStackObject(OrigTy.getScalarSizeInBits(),
866 AlignmentForStored, false);
867 Register PointerToStackReg =
868 MIRBuilder.buildFrameIndex(PointerTy, FrameIdx).getReg(0);
869 MachinePointerInfo StackPointerMPO =
871 // Store the value in the previously created stack space.
872 MIRBuilder.buildStore(Args[i].OrigRegs[Part], PointerToStackReg,
873 StackPointerMPO,
874 inferAlignFromPtrInfo(MF, StackPointerMPO));
875
876 ArgReg = PointerToStackReg;
877 IndirectParameterPassingHandled = true;
878 }
879
880 if (VA.isMemLoc() && !Flags.isByVal()) {
881 // Individual pieces may have been spilled to the stack and others
882 // passed in registers.
883
884 // TODO: The memory size may be larger than the value we need to
885 // store. We may need to adjust the offset for big endian targets.
886 LLT MemTy = Handler.getStackValueStoreType(DL, VA, Flags);
887
889 Register StackAddr =
891 ? PointerTy.getSizeInBytes()
892 : MemTy.getSizeInBytes(),
893 VA.getLocMemOffset(), MPO, Flags);
894
895 // Finish the handling of indirect passing from the passers
896 // (OutgoingParameterHandler) side.
897 // This branch is needed, so the pointer to the value is loaded onto the
898 // stack.
900 Handler.assignValueToAddress(ArgReg, StackAddr, PointerTy, MPO, VA);
901 else
902 Handler.assignValueToAddress(Args[i], Part, StackAddr, MemTy, MPO,
903 VA);
904 } else if (VA.isMemLoc() && Flags.isByVal()) {
905 assert(Args[i].Regs.size() == 1 && "didn't expect split byval pointer");
906
907 if (Handler.isIncomingArgumentHandler()) {
908 // We just need to copy the frame index value to the pointer.
910 Register StackAddr = Handler.getStackAddress(
911 Flags.getByValSize(), VA.getLocMemOffset(), MPO, Flags);
912 MIRBuilder.buildCopy(Args[i].Regs[0], StackAddr);
913 } else {
914 // For outgoing byval arguments, insert the implicit copy byval
915 // implies, such that writes in the callee do not modify the caller's
916 // value.
917 uint64_t MemSize = Flags.getByValSize();
918 int64_t Offset = VA.getLocMemOffset();
919
920 MachinePointerInfo DstMPO;
921 Register StackAddr =
922 Handler.getStackAddress(MemSize, Offset, DstMPO, Flags);
923
924 MachinePointerInfo SrcMPO(Args[i].OrigValue);
925 if (!Args[i].OrigValue) {
926 // We still need to accurately track the stack address space if we
927 // don't know the underlying value.
928 const LLT PtrTy = MRI.getType(StackAddr);
929 SrcMPO = MachinePointerInfo(PtrTy.getAddressSpace());
930 }
931
932 Align DstAlign = std::max(Flags.getNonZeroByValAlign(),
933 inferAlignFromPtrInfo(MF, DstMPO));
934
935 Align SrcAlign = std::max(Flags.getNonZeroByValAlign(),
936 inferAlignFromPtrInfo(MF, SrcMPO));
937
938 Handler.copyArgumentMemory(Args[i], StackAddr, Args[i].Regs[0],
939 DstMPO, DstAlign, SrcMPO, SrcAlign,
940 MemSize, VA);
941 }
942 } else if (i == 0 && !ThisReturnRegs.empty() &&
943 Handler.isIncomingArgumentHandler() &&
945 Handler.assignValueToReg(ArgReg, ThisReturnRegs[Part], VA);
946 } else if (Handler.isIncomingArgumentHandler()) {
947 Handler.assignValueToReg(ArgReg, VA.getLocReg(), VA);
948 } else {
949 DelayedOutgoingRegAssignments.emplace_back([=, &Handler]() {
950 Handler.assignValueToReg(ArgReg, VA.getLocReg(), VA);
951 });
952 }
953
954 // Finish the handling of indirect parameter passing when receiving
955 // the value (we are in the called function or the caller when receiving
956 // the return value).
957 if (VA.getLocInfo() == CCValAssign::Indirect &&
958 Handler.isIncomingArgumentHandler()) {
959 Align Alignment = DL.getABITypeAlign(Args[i].Ty);
961
962 // Since we are doing indirect parameter passing, we know that the value
963 // in the temporary register is not the value passed to the function,
964 // but rather a pointer to that value. Let's load that value into the
965 // virtual register where the parameter should go.
966 MIRBuilder.buildLoad(Args[i].OrigRegs[0], Args[i].Regs[0], MPO,
967 Alignment);
968
969 IndirectParameterPassingHandled = true;
970 }
971
972 if (IndirectParameterPassingHandled)
973 break;
974 }
975
976 // Now that all pieces have been assigned, re-pack the register typed values
977 // into the original value typed registers. This is only necessary, when
978 // the value was passed in multiple registers, not indirectly.
979 if (Handler.isIncomingArgumentHandler() && OrigVT != LocVT &&
980 !IndirectParameterPassingHandled) {
981 // Merge the split registers into the expected larger result vregs of
982 // the original call.
983 buildCopyFromRegs(MIRBuilder, Args[i].OrigRegs, Args[i].Regs, OrigTy,
984 LocTy, Args[i].Flags[0]);
985 }
986
987 j += NumParts - 1;
988 }
989 for (auto &Fn : DelayedOutgoingRegAssignments)
990 Fn();
991
992 return true;
993}
994
996 ArrayRef<Register> VRegs, Register DemoteReg,
997 int FI) const {
998 MachineFunction &MF = MIRBuilder.getMF();
1000 const DataLayout &DL = MF.getDataLayout();
1001
1002 SmallVector<EVT, 4> SplitVTs;
1004 ComputeValueVTs(*TLI, DL, RetTy, SplitVTs, /*MemVTs=*/nullptr, &Offsets, 0);
1005
1006 assert(VRegs.size() == SplitVTs.size());
1007
1008 unsigned NumValues = SplitVTs.size();
1009 Align BaseAlign = DL.getPrefTypeAlign(RetTy);
1010 Type *RetPtrTy =
1011 PointerType::get(RetTy->getContext(), DL.getAllocaAddrSpace());
1012 LLT OffsetLLTy = getLLTForType(*DL.getIndexType(RetPtrTy), DL);
1013
1015
1016 for (unsigned I = 0; I < NumValues; ++I) {
1017 Register Addr;
1018 MIRBuilder.materializeObjectPtrOffset(Addr, DemoteReg, OffsetLLTy,
1019 Offsets[I]);
1020 auto *MMO = MF.getMachineMemOperand(PtrInfo, MachineMemOperand::MOLoad,
1021 MRI.getType(VRegs[I]),
1022 commonAlignment(BaseAlign, Offsets[I]));
1023 MIRBuilder.buildLoad(VRegs[I], Addr, *MMO);
1024 }
1025}
1026
1028 ArrayRef<Register> VRegs,
1029 Register DemoteReg) const {
1030 MachineFunction &MF = MIRBuilder.getMF();
1032 const DataLayout &DL = MF.getDataLayout();
1033
1034 SmallVector<EVT, 4> SplitVTs;
1036 ComputeValueVTs(*TLI, DL, RetTy, SplitVTs, /*MemVTs=*/nullptr, &Offsets, 0);
1037
1038 assert(VRegs.size() == SplitVTs.size());
1039
1040 unsigned NumValues = SplitVTs.size();
1041 Align BaseAlign = DL.getPrefTypeAlign(RetTy);
1042 unsigned AS = DL.getAllocaAddrSpace();
1043 LLT OffsetLLTy = getLLTForType(*DL.getIndexType(RetTy->getContext(), AS), DL);
1044
1045 MachinePointerInfo PtrInfo(AS);
1046
1047 for (unsigned I = 0; I < NumValues; ++I) {
1048 Register Addr;
1049 MIRBuilder.materializeObjectPtrOffset(Addr, DemoteReg, OffsetLLTy,
1050 Offsets[I]);
1051 auto *MMO = MF.getMachineMemOperand(PtrInfo, MachineMemOperand::MOStore,
1052 MRI.getType(VRegs[I]),
1053 commonAlignment(BaseAlign, Offsets[I]));
1054 MIRBuilder.buildStore(VRegs[I], Addr, *MMO);
1055 }
1056}
1057
1059 const Function &F, SmallVectorImpl<ArgInfo> &SplitArgs, Register &DemoteReg,
1060 MachineRegisterInfo &MRI, const DataLayout &DL) const {
1061 unsigned AS = DL.getAllocaAddrSpace();
1062 DemoteReg = MRI.createGenericVirtualRegister(
1063 LLT::pointer(AS, DL.getPointerSizeInBits(AS)));
1064
1065 Type *PtrTy = PointerType::get(F.getContext(), AS);
1066
1067 SmallVector<EVT, 1> ValueVTs;
1068 ComputeValueVTs(*TLI, DL, PtrTy, ValueVTs);
1069
1070 // NOTE: Assume that a pointer won't get split into more than one VT.
1071 assert(ValueVTs.size() == 1);
1072
1073 ArgInfo DemoteArg(DemoteReg, ValueVTs[0].getTypeForEVT(PtrTy->getContext()),
1075 setArgFlags(DemoteArg, AttributeList::ReturnIndex, DL, F);
1076 DemoteArg.Flags[0].setSRet();
1077 SplitArgs.insert(SplitArgs.begin(), DemoteArg);
1078}
1079
1081 const CallBase &CB,
1082 CallLoweringInfo &Info) const {
1083 const DataLayout &DL = MIRBuilder.getDataLayout();
1084 Type *RetTy = CB.getType();
1085 unsigned AS = DL.getAllocaAddrSpace();
1086 LLT FramePtrTy = LLT::pointer(AS, DL.getPointerSizeInBits(AS));
1087
1088 int FI = MIRBuilder.getMF().getFrameInfo().CreateStackObject(
1089 DL.getTypeAllocSize(RetTy), DL.getPrefTypeAlign(RetTy), false);
1090
1091 Register DemoteReg = MIRBuilder.buildFrameIndex(FramePtrTy, FI).getReg(0);
1092 ArgInfo DemoteArg(DemoteReg, PointerType::get(RetTy->getContext(), AS),
1094 setArgFlags(DemoteArg, AttributeList::ReturnIndex, DL, CB);
1095 DemoteArg.Flags[0].setSRet();
1096
1097 Info.OrigArgs.insert(Info.OrigArgs.begin(), DemoteArg);
1098 Info.DemoteStackIndex = FI;
1099 Info.DemoteRegister = DemoteReg;
1100}
1101
1104 CCAssignFn *Fn) const {
1105 for (unsigned I = 0, E = Outs.size(); I < E; ++I) {
1106 MVT VT = MVT::getVT(Outs[I].Ty);
1107 if (Fn(I, VT, VT, CCValAssign::Full, Outs[I].Flags[0], Outs[I].Ty, CCInfo))
1108 return false;
1109 }
1110 return true;
1111}
1112
1114 AttributeList Attrs,
1116 const DataLayout &DL) const {
1117 LLVMContext &Context = RetTy->getContext();
1119
1120 SmallVector<EVT, 4> SplitVTs;
1121 ComputeValueVTs(*TLI, DL, RetTy, SplitVTs);
1122 addArgFlagsFromAttributes(Flags, Attrs, AttributeList::ReturnIndex);
1123
1124 for (EVT VT : SplitVTs) {
1125 unsigned NumParts =
1126 TLI->getNumRegistersForCallingConv(Context, CallConv, VT);
1127 MVT RegVT = TLI->getRegisterTypeForCallingConv(Context, CallConv, VT);
1128 Type *PartTy = EVT(RegVT).getTypeForEVT(Context);
1129
1130 for (unsigned I = 0; I < NumParts; ++I) {
1131 Outs.emplace_back(PartTy, Flags);
1132 }
1133 }
1134}
1135
1137 const auto &F = MF.getFunction();
1138 Type *ReturnType = F.getReturnType();
1139 CallingConv::ID CallConv = F.getCallingConv();
1140
1142 getReturnInfo(CallConv, ReturnType, F.getAttributes(), SplitArgs,
1143 MF.getDataLayout());
1144 return canLowerReturn(MF, CallConv, SplitArgs, F.isVarArg());
1145}
1146
1148 const MachineRegisterInfo &MRI, const uint32_t *CallerPreservedMask,
1149 const SmallVectorImpl<CCValAssign> &OutLocs,
1150 const SmallVectorImpl<ArgInfo> &OutArgs) const {
1151 for (unsigned i = 0; i < OutLocs.size(); ++i) {
1152 const auto &ArgLoc = OutLocs[i];
1153 // If it's not a register, it's fine.
1154 if (!ArgLoc.isRegLoc())
1155 continue;
1156
1157 MCRegister PhysReg = ArgLoc.getLocReg();
1158
1159 // Only look at callee-saved registers.
1160 if (MachineOperand::clobbersPhysReg(CallerPreservedMask, PhysReg))
1161 continue;
1162
1163 LLVM_DEBUG(
1164 dbgs()
1165 << "... Call has an argument passed in a callee-saved register.\n");
1166
1167 // Check if it was copied from.
1168 const ArgInfo &OutInfo = OutArgs[i];
1169
1170 if (OutInfo.Regs.size() > 1) {
1171 LLVM_DEBUG(
1172 dbgs() << "... Cannot handle arguments in multiple registers.\n");
1173 return false;
1174 }
1175
1176 // Check if we copy the register, walking through copies from virtual
1177 // registers. Note that getDefIgnoringCopies does not ignore copies from
1178 // physical registers.
1179 MachineInstr *RegDef = getDefIgnoringCopies(OutInfo.Regs[0], MRI);
1180 if (!RegDef || RegDef->getOpcode() != TargetOpcode::COPY) {
1181 LLVM_DEBUG(
1182 dbgs()
1183 << "... Parameter was not copied into a VReg, cannot tail call.\n");
1184 return false;
1185 }
1186
1187 // Got a copy. Verify that it's the same as the register we want.
1188 Register CopyRHS = RegDef->getOperand(1).getReg();
1189 if (CopyRHS != PhysReg) {
1190 LLVM_DEBUG(dbgs() << "... Callee-saved register was not copied into "
1191 "VReg, cannot tail call.\n");
1192 return false;
1193 }
1194 }
1195
1196 return true;
1197}
1198
1200 MachineFunction &MF,
1202 ValueAssigner &CalleeAssigner,
1203 ValueAssigner &CallerAssigner) const {
1204 const Function &F = MF.getFunction();
1205 CallingConv::ID CalleeCC = Info.CallConv;
1206 CallingConv::ID CallerCC = F.getCallingConv();
1207
1208 if (CallerCC == CalleeCC)
1209 return true;
1210
1212 CCState CCInfo1(CalleeCC, Info.IsVarArg, MF, ArgLocs1, F.getContext());
1213 if (!determineAssignments(CalleeAssigner, InArgs, CCInfo1))
1214 return false;
1215
1217 CCState CCInfo2(CallerCC, F.isVarArg(), MF, ArgLocs2, F.getContext());
1218 if (!determineAssignments(CallerAssigner, InArgs, CCInfo2))
1219 return false;
1220
1221 // We need the argument locations to match up exactly. If there's more in
1222 // one than the other, then we are done.
1223 if (ArgLocs1.size() != ArgLocs2.size())
1224 return false;
1225
1226 // Make sure that each location is passed in exactly the same way.
1227 for (unsigned i = 0, e = ArgLocs1.size(); i < e; ++i) {
1228 const CCValAssign &Loc1 = ArgLocs1[i];
1229 const CCValAssign &Loc2 = ArgLocs2[i];
1230
1231 // We need both of them to be the same. So if one is a register and one
1232 // isn't, we're done.
1233 if (Loc1.isRegLoc() != Loc2.isRegLoc())
1234 return false;
1235
1236 if (Loc1.isRegLoc()) {
1237 // If they don't have the same register location, we're done.
1238 if (Loc1.getLocReg() != Loc2.getLocReg())
1239 return false;
1240
1241 // They matched, so we can move to the next ArgLoc.
1242 continue;
1243 }
1244
1245 // Loc1 wasn't a RegLoc, so they both must be MemLocs. Check if they match.
1246 if (Loc1.getLocMemOffset() != Loc2.getLocMemOffset())
1247 return false;
1248 }
1249
1250 return true;
1251}
1252
1254 const DataLayout &DL, const CCValAssign &VA, ISD::ArgFlagsTy Flags) const {
1255 const MVT ValVT = VA.getValVT();
1256 if (ValVT != MVT::iPTR) {
1257 LLT ValTy(ValVT);
1258
1259 // We lost the pointeriness going through CCValAssign, so try to restore it
1260 // based on the flags.
1261 if (Flags.isPointer()) {
1262 LLT PtrTy = LLT::pointer(Flags.getPointerAddrSpace(),
1263 ValTy.getScalarSizeInBits());
1264 if (ValVT.isVector() && ValVT.getVectorNumElements() != 1)
1265 return LLT::vector(ValTy.getElementCount(), PtrTy);
1266 return PtrTy;
1267 }
1268
1269 return ValTy;
1270 }
1271
1272 unsigned AddrSpace = Flags.getPointerAddrSpace();
1273 return LLT::pointer(AddrSpace, DL.getPointerSize(AddrSpace));
1274}
1275
1277 const ArgInfo &Arg, Register DstPtr, Register SrcPtr,
1278 const MachinePointerInfo &DstPtrInfo, Align DstAlign,
1279 const MachinePointerInfo &SrcPtrInfo, Align SrcAlign, uint64_t MemSize,
1280 CCValAssign &VA) const {
1281 MachineFunction &MF = MIRBuilder.getMF();
1283 SrcPtrInfo,
1285 SrcAlign);
1286
1288 DstPtrInfo,
1290 MemSize, DstAlign);
1291
1292 const LLT PtrTy = MRI.getType(DstPtr);
1293 const LLT SizeTy = LLT::scalar(PtrTy.getSizeInBits());
1294
1295 auto SizeConst = MIRBuilder.buildConstant(SizeTy, MemSize);
1296 MIRBuilder.buildMemCpy(DstPtr, SrcPtr, SizeConst, *DstMMO, *SrcMMO);
1297}
1298
1300 const CCValAssign &VA,
1301 unsigned MaxSizeBits) {
1302 LLT LocTy{VA.getLocVT()};
1303 LLT ValTy{VA.getValVT()};
1304
1305 if (LocTy.getSizeInBits() == ValTy.getSizeInBits())
1306 return ValReg;
1307
1308 if (LocTy.isScalar() && MaxSizeBits && MaxSizeBits < LocTy.getSizeInBits()) {
1309 if (MaxSizeBits <= ValTy.getSizeInBits())
1310 return ValReg;
1311 LocTy = LLT::scalar(MaxSizeBits);
1312 }
1313
1314 const LLT ValRegTy = MRI.getType(ValReg);
1315 if (ValRegTy.isPointer()) {
1316 // The x32 ABI wants to zero extend 32-bit pointers to 64-bit registers, so
1317 // we have to cast to do the extension.
1318 LLT IntPtrTy = LLT::scalar(ValRegTy.getSizeInBits());
1319 ValReg = MIRBuilder.buildPtrToInt(IntPtrTy, ValReg).getReg(0);
1320 }
1321
1322 switch (VA.getLocInfo()) {
1323 default:
1324 break;
1325 case CCValAssign::Full:
1326 case CCValAssign::BCvt:
1327 // FIXME: bitconverting between vector types may or may not be a
1328 // nop in big-endian situations.
1329 return ValReg;
1330 case CCValAssign::AExt: {
1331 auto MIB = MIRBuilder.buildAnyExt(LocTy, ValReg);
1332 return MIB.getReg(0);
1333 }
1334 case CCValAssign::SExt: {
1335 Register NewReg = MRI.createGenericVirtualRegister(LocTy);
1336 MIRBuilder.buildSExt(NewReg, ValReg);
1337 return NewReg;
1338 }
1339 case CCValAssign::ZExt: {
1340 Register NewReg = MRI.createGenericVirtualRegister(LocTy);
1341 MIRBuilder.buildZExt(NewReg, ValReg);
1342 return NewReg;
1343 }
1344 }
1345 llvm_unreachable("unable to extend register");
1346}
1347
1348void CallLowering::ValueAssigner::anchor() {}
1349
1351 const CCValAssign &VA, Register SrcReg, LLT NarrowTy) {
1352 switch (VA.getLocInfo()) {
1354 return MIRBuilder
1355 .buildAssertZExt(MRI.cloneVirtualRegister(SrcReg), SrcReg,
1356 NarrowTy.getScalarSizeInBits())
1357 .getReg(0);
1358 }
1360 return MIRBuilder
1361 .buildAssertSExt(MRI.cloneVirtualRegister(SrcReg), SrcReg,
1362 NarrowTy.getScalarSizeInBits())
1363 .getReg(0);
1364 break;
1365 }
1366 default:
1367 return SrcReg;
1368 }
1369}
1370
1371/// Check if we can use a basic COPY instruction between the two types.
1372///
1373/// We're currently building on top of the infrastructure using MVT, which loses
1374/// pointer information in the CCValAssign. We accept copies from physical
1375/// registers that have been reported as integers if it's to an equivalent sized
1376/// pointer LLT.
1377static bool isCopyCompatibleType(LLT SrcTy, LLT DstTy) {
1378 if (SrcTy == DstTy)
1379 return true;
1380
1381 if (SrcTy.getSizeInBits() != DstTy.getSizeInBits())
1382 return false;
1383
1384 SrcTy = SrcTy.getScalarType();
1385 DstTy = DstTy.getScalarType();
1386
1387 return (SrcTy.isPointer() && DstTy.isScalar()) ||
1388 (DstTy.isPointer() && SrcTy.isScalar());
1389}
1390
1392 Register ValVReg, Register PhysReg, const CCValAssign &VA) {
1393 const MVT LocVT = VA.getLocVT();
1394 const LLT LocTy(LocVT);
1395 const LLT RegTy = MRI.getType(ValVReg);
1396
1397 if (isCopyCompatibleType(RegTy, LocTy)) {
1398 MIRBuilder.buildCopy(ValVReg, PhysReg);
1399 return;
1400 }
1401
1402 auto Copy = MIRBuilder.buildCopy(LocTy, PhysReg);
1403 auto Hint = buildExtensionHint(VA, Copy.getReg(0), RegTy);
1404 MIRBuilder.buildTrunc(ValVReg, Hint);
1405}
unsigned const MachineRegisterInfo * MRI
assert(UImm &&(UImm !=~static_cast< T >(0)) &&"Invalid immediate!")
MachineBasicBlock MachineBasicBlock::iterator DebugLoc DL
static GCRegistry::Add< OcamlGC > B("ocaml", "ocaml 3.10-compatible GC")
static void addFlagsUsingAttrFn(ISD::ArgFlagsTy &Flags, const std::function< bool(Attribute::AttrKind)> &AttrFn)
Helper function which updates Flags when AttrFn returns true.
static void buildCopyToRegs(MachineIRBuilder &B, ArrayRef< Register > DstRegs, Register SrcReg, LLT SrcTy, LLT PartTy, unsigned ExtendOp=TargetOpcode::G_ANYEXT)
Create a sequence of instructions to expand the value in SrcReg (of type SrcTy) to the types in DstRe...
static MachineInstrBuilder mergeVectorRegsToResultRegs(MachineIRBuilder &B, ArrayRef< Register > DstRegs, ArrayRef< Register > SrcRegs)
Pack values SrcRegs to cover the vector type result DstRegs.
static void buildCopyFromRegs(MachineIRBuilder &B, ArrayRef< Register > OrigRegs, ArrayRef< Register > Regs, LLT LLTy, LLT PartLLT, const ISD::ArgFlagsTy Flags)
Create a sequence of instructions to combine pieces split into register typed values to the original ...
static bool isCopyCompatibleType(LLT SrcTy, LLT DstTy)
Check if we can use a basic COPY instruction between the two types.
static unsigned extendOpFromFlags(llvm::ISD::ArgFlagsTy Flags)
This file describes how to lower LLVM calls to machine code calls.
Module.h This file contains the declarations for the Module class.
#define F(x, y, z)
Definition MD5.cpp:54
#define I(x, y, z)
Definition MD5.cpp:57
This file declares the MachineIRBuilder class.
Register Reg
Promote Memory to Register
Definition Mem2Reg.cpp:110
static MCRegister getReg(const MCDisassembler *D, unsigned RC, unsigned RegNo)
MachineInstr unsigned OpIdx
R600 Clause Merge
#define LLVM_DEBUG(...)
Definition Debug.h:114
This file describes how to lower LLVM code to machine code.
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition ArrayRef.h:40
ArrayRef< T > take_front(size_t N=1) const
Return a copy of *this with only the first N elements.
Definition ArrayRef.h:219
ArrayRef< T > drop_front(size_t N=1) const
Drop the first N elements of the array.
Definition ArrayRef.h:195
const T & front() const
front - Get the first element.
Definition ArrayRef.h:145
size_t size() const
size - Get the array size.
Definition ArrayRef.h:142
bool empty() const
empty - Check if the array is empty.
Definition ArrayRef.h:137
AttrKind
This enumeration lists the attributes that can be associated with parameters, function results,...
Definition Attributes.h:88
CCState - This class holds information needed while lowering arguments and return values.
CallingConv::ID getCallingConv() const
LLVMContext & getContext() const
CCValAssign - Represent assignment of one arg/retval to a location.
Register getLocReg() const
LocInfo getLocInfo() const
bool needsCustom() const
int64_t getLocMemOffset() const
unsigned getValNo() const
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
MaybeAlign getRetAlign() const
Extract the alignment of the return value.
std::optional< OperandBundleUse > getOperandBundle(StringRef Name) const
Return an operand bundle by name, if present.
CallingConv::ID getCallingConv() const
LLVM_ABI bool isMustTailCall() const
Tests if this call site must be tail call optimized.
LLVM_ABI bool isIndirectCall() const
Return true if the callsite is an indirect call.
unsigned countOperandBundlesOfType(StringRef Name) const
Return the number of operand bundles with the tag Name attached to this instruction.
Value * getCalledOperand() const
bool isConvergent() const
Determine if the invoke is convergent.
FunctionType * getFunctionType() const
iterator_range< User::op_iterator > args()
Iteration adapter for range-for loops.
AttributeList getAttributes() const
Return the attributes for this call.
LLVM_ABI bool isTailCall() const
Tests if this call site is marked as a tail call.
void insertSRetOutgoingArgument(MachineIRBuilder &MIRBuilder, const CallBase &CB, CallLoweringInfo &Info) const
For the call-base described by CB, insert the hidden sret ArgInfo to the OrigArgs field of Info.
void insertSRetLoads(MachineIRBuilder &MIRBuilder, Type *RetTy, ArrayRef< Register > VRegs, Register DemoteReg, int FI) const
Load the returned value from the stack into virtual registers in VRegs.
bool checkReturnTypeForCallConv(MachineFunction &MF) const
Toplevel function to check the return type based on the target calling convention.
bool handleAssignments(ValueHandler &Handler, SmallVectorImpl< ArgInfo > &Args, CCState &CCState, SmallVectorImpl< CCValAssign > &ArgLocs, MachineIRBuilder &MIRBuilder, ArrayRef< Register > ThisReturnRegs={}) const
Use Handler to insert code to handle the argument/return values represented by Args.
bool resultsCompatible(CallLoweringInfo &Info, MachineFunction &MF, SmallVectorImpl< ArgInfo > &InArgs, ValueAssigner &CalleeAssigner, ValueAssigner &CallerAssigner) const
void splitToValueTypes(const ArgInfo &OrigArgInfo, SmallVectorImpl< ArgInfo > &SplitArgs, const DataLayout &DL, CallingConv::ID CallConv, SmallVectorImpl< uint64_t > *Offsets=nullptr) const
Break OrigArgInfo into one or more pieces the calling convention can process, returned in SplitArgs.
virtual bool canLowerReturn(MachineFunction &MF, CallingConv::ID CallConv, SmallVectorImpl< BaseArgInfo > &Outs, bool IsVarArg) const
This hook must be implemented to check whether the return values described by Outs can fit into the r...
virtual bool isTypeIsValidForThisReturn(EVT Ty) const
For targets which support the "returned" parameter attribute, returns true if the given type is a val...
void insertSRetIncomingArgument(const Function &F, SmallVectorImpl< ArgInfo > &SplitArgs, Register &DemoteReg, MachineRegisterInfo &MRI, const DataLayout &DL) const
Insert the hidden sret ArgInfo to the beginning of SplitArgs.
ISD::ArgFlagsTy getAttributesForArgIdx(const CallBase &Call, unsigned ArgIdx) const
bool determineAndHandleAssignments(ValueHandler &Handler, ValueAssigner &Assigner, SmallVectorImpl< ArgInfo > &Args, MachineIRBuilder &MIRBuilder, CallingConv::ID CallConv, bool IsVarArg, ArrayRef< Register > ThisReturnRegs={}) const
Invoke ValueAssigner::assignArg on each of the given Args and then use Handler to move them to the as...
void insertSRetStores(MachineIRBuilder &MIRBuilder, Type *RetTy, ArrayRef< Register > VRegs, Register DemoteReg) const
Store the return value given by VRegs into stack starting at the offset specified in DemoteReg.
void addArgFlagsFromAttributes(ISD::ArgFlagsTy &Flags, const AttributeList &Attrs, unsigned OpIdx) const
Adds flags to Flags based off of the attributes in Attrs.
bool parametersInCSRMatch(const MachineRegisterInfo &MRI, const uint32_t *CallerPreservedMask, const SmallVectorImpl< CCValAssign > &ArgLocs, const SmallVectorImpl< ArgInfo > &OutVals) const
Check whether parameters to a call that are passed in callee saved registers are the same as from the...
void getReturnInfo(CallingConv::ID CallConv, Type *RetTy, AttributeList Attrs, SmallVectorImpl< BaseArgInfo > &Outs, const DataLayout &DL) const
Get the type and the ArgFlags for the split components of RetTy as returned by ComputeValueVTs.
bool determineAssignments(ValueAssigner &Assigner, SmallVectorImpl< ArgInfo > &Args, CCState &CCInfo) const
Analyze the argument list in Args, using Assigner to populate CCInfo.
bool checkReturn(CCState &CCInfo, SmallVectorImpl< BaseArgInfo > &Outs, CCAssignFn *Fn) const
const TargetLowering * getTLI() const
Getter for generic TargetLowering class.
virtual bool lowerCall(MachineIRBuilder &MIRBuilder, CallLoweringInfo &Info) const
This hook must be implemented to lower the given call instruction, including argument and return valu...
void setArgFlags(ArgInfo &Arg, unsigned OpIdx, const DataLayout &DL, const FuncInfoTy &FuncInfo) const
ISD::ArgFlagsTy getAttributesForReturn(const CallBase &Call) const
A parsed version of the target data layout string in and methods for querying it.
Definition DataLayout.h:63
unsigned getAllocaAddrSpace() const
Definition DataLayout.h:239
unsigned getNumParams() const
Return the number of fixed parameters this function type requires.
bool isVarArg() const
MDNode * getMetadata(unsigned KindID) const
Get the metadata of given kind attached to this Instruction.
constexpr unsigned getScalarSizeInBits() const
constexpr bool isScalar() const
constexpr LLT changeElementType(LLT NewEltTy) const
If this type is a vector, return a vector with the same number of elements but the new element type.
static constexpr LLT vector(ElementCount EC, unsigned ScalarSizeInBits)
Get a low-level vector of some number of elements and element width.
static constexpr LLT scalar(unsigned SizeInBits)
Get a low-level scalar or aggregate "bag of bits".
constexpr uint16_t getNumElements() const
Returns the number of elements in a vector LLT.
constexpr bool isVector() const
static constexpr LLT pointer(unsigned AddressSpace, unsigned SizeInBits)
Get a low-level pointer in the given address space.
constexpr TypeSize getSizeInBits() const
Returns the total size of the type. Must only be called on sized types.
constexpr bool isPointer() const
constexpr LLT getElementType() const
Returns the vector's element type. Only valid for vector types.
constexpr ElementCount getElementCount() const
constexpr unsigned getAddressSpace() const
static constexpr LLT fixed_vector(unsigned NumElements, unsigned ScalarSizeInBits)
Get a low-level fixed-width vector of some number of elements and element width.
constexpr LLT changeElementCount(ElementCount EC) const
Return a vector or scalar with the same element type and the new element count.
constexpr LLT getScalarType() const
constexpr TypeSize getSizeInBytes() const
Returns the total size of the type in bytes, i.e.
This is an important class for using LLVM in a threaded context.
Definition LLVMContext.h:68
Wrapper class representing physical registers. Should be passed by value.
Definition MCRegister.h:41
Machine Value Type.
unsigned getVectorNumElements() const
bool isVector() const
Return true if this is a vector value type.
static LLVM_ABI MVT getVT(Type *Ty, bool HandleUnknown=false)
Return the value type corresponding to the specified type.
The MachineFrameInfo class represents an abstract stack frame until prolog/epilog code is inserted.
LLVM_ABI int CreateStackObject(uint64_t Size, Align Alignment, bool isSpillSlot, const AllocaInst *Alloca=nullptr, uint8_t ID=0)
Create a new statically sized stack object, returning a nonnegative identifier to represent it.
MachineMemOperand * getMachineMemOperand(MachinePointerInfo PtrInfo, MachineMemOperand::Flags f, LLT MemTy, Align base_alignment, const AAMDNodes &AAInfo=AAMDNodes(), const MDNode *Ranges=nullptr, SyncScope::ID SSID=SyncScope::System, AtomicOrdering Ordering=AtomicOrdering::NotAtomic, AtomicOrdering FailureOrdering=AtomicOrdering::NotAtomic)
getMachineMemOperand - Allocate a new MachineMemOperand.
MachineFrameInfo & getFrameInfo()
getFrameInfo - Return the frame info object for the current function.
MachineRegisterInfo & getRegInfo()
getRegInfo - Return information about the registers currently in use.
const DataLayout & getDataLayout() const
Return the DataLayout attached to the Module associated to this MF.
Function & getFunction()
Return the LLVM function that this machine code represents.
const TargetMachine & getTarget() const
getTarget - Return the target machine this machine code is compiled with
Helper class to build MachineInstr.
MachineInstrBuilder buildGlobalValue(const DstOp &Res, const GlobalValue *GV)
Build and insert Res = G_GLOBAL_VALUE GV.
std::optional< MachineInstrBuilder > materializeObjectPtrOffset(Register &Res, Register Op0, const LLT ValueTy, uint64_t Value)
Materialize and insert an instruction with appropriate flags for addressing some offset of an object,...
MachineInstrBuilder buildLoad(const DstOp &Res, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert Res = G_LOAD Addr, MMO.
MachineInstrBuilder buildStore(const SrcOp &Val, const SrcOp &Addr, MachineMemOperand &MMO)
Build and insert G_STORE Val, Addr, MMO.
MachineInstrBuilder buildFrameIndex(const DstOp &Res, int Idx)
Build and insert Res = G_FRAME_INDEX Idx.
MachineFunction & getMF()
Getter for the function we currently build.
MachineInstrBuilder buildAssertAlign(const DstOp &Res, const SrcOp &Op, Align AlignVal)
Build and insert Res = G_ASSERT_ALIGN Op, AlignVal.
MachineInstrBuilder buildCopy(const DstOp &Res, const SrcOp &Op)
Build and insert Res = COPY Op.
const DataLayout & getDataLayout() const
Register getReg(unsigned Idx) const
Get the register for the operand index.
Representation of each machine instruction.
unsigned getOpcode() const
Returns the opcode of this MachineInstr.
const MachineOperand & getOperand(unsigned i) const
A description of a memory reference used in the backend.
@ MODereferenceable
The memory access is dereferenceable (i.e., doesn't trap).
@ MOLoad
The memory access reads data.
@ MOStore
The memory access writes data.
Register getReg() const
getReg - Returns the register number.
static MachineOperand CreateGA(const GlobalValue *GV, int64_t Offset, unsigned TargetFlags=0)
static bool clobbersPhysReg(const uint32_t *RegMask, MCRegister PhysReg)
clobbersPhysReg - Returns true if this RegMask clobbers PhysReg.
static MachineOperand CreateReg(Register Reg, bool isDef, bool isImp=false, bool isKill=false, bool isDead=false, bool isUndef=false, bool isEarlyClobber=false, unsigned SubReg=0, bool isDebug=false, bool isInternalRead=false, bool isRenamable=false)
MachineRegisterInfo - Keep track of information for virtual and physical registers,...
Class to represent pointers.
static LLVM_ABI PointerType * get(Type *ElementType, unsigned AddressSpace)
This constructs a pointer to an object of the specified type in a numbered address space.
Wrapper class representing virtual and physical registers.
Definition Register.h:20
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
reference emplace_back(ArgTypes &&... Args)
void reserve(size_type N)
iterator insert(iterator I, T &&Elt)
void truncate(size_type N)
Like resize, but requires that N is less than size().
void push_back(const T &Elt)
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
virtual Align getByValTypeAlignment(Type *Ty, const DataLayout &DL) const
Returns the desired alignment for ByVal or InAlloca aggregate function arguments in the caller parame...
The instances of the Type class are immutable: once they are created, they are never changed.
Definition Type.h:45
LLVM_ABI unsigned getPointerAddressSpace() const
Get the address space of this pointer or pointer vector type.
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition Type.h:352
LLVMContext & getContext() const
Return the LLVMContext in which this type was uniqued.
Definition Type.h:128
LLVM Value Representation.
Definition Value.h:75
Type * getType() const
All values are typed, get the type of this value.
Definition Value.h:256
LLVM_ABI const Value * stripPointerCasts() const
Strip off pointer casts, all-zero GEPs and address space casts.
Definition Value.cpp:701
constexpr ScalarTy getFixedValue() const
Definition TypeSize.h:200
static constexpr bool isKnownLT(const FixedOrScalableQuantity &LHS, const FixedOrScalableQuantity &RHS)
Definition TypeSize.h:216
static constexpr bool isKnownGT(const FixedOrScalableQuantity &LHS, const FixedOrScalableQuantity &RHS)
Definition TypeSize.h:223
CallInst * Call
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
unsigned ID
LLVM IR allows to use arbitrary numbers as calling convention identifiers.
Definition CallingConv.h:24
This is an optimization pass for GlobalISel generic memory operations.
@ Offset
Definition DWP.cpp:532
void ComputeValueVTs(const TargetLowering &TLI, const DataLayout &DL, Type *Ty, SmallVectorImpl< EVT > &ValueVTs, SmallVectorImpl< EVT > *MemVTs=nullptr, SmallVectorImpl< TypeSize > *Offsets=nullptr, TypeSize StartingOffset=TypeSize::getZero())
ComputeValueVTs - Given an LLVM IR type, compute a sequence of EVTs that represent all the individual...
Definition Analysis.cpp:119
decltype(auto) dyn_cast(const From &Val)
dyn_cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:643
bool CCAssignFn(unsigned ValNo, MVT ValVT, MVT LocVT, CCValAssign::LocInfo LocInfo, ISD::ArgFlagsTy ArgFlags, Type *OrigTy, CCState &State)
CCAssignFn - This function assigns a location for Val, updating State to reflect the change.
void * PointerTy
LLVM_ABI MachineInstr * getDefIgnoringCopies(Register Reg, const MachineRegisterInfo &MRI)
Find the def instruction for Reg, folding away any trivial copies.
Definition Utils.cpp:492
LLVM_ABI raw_ostream & dbgs()
dbgs() - This returns a reference to a raw_ostream for debugging messages.
Definition Debug.cpp:207
bool isa(const From &Val)
isa<X> - Return true if the parameter to the template is an instance of one of the template type argu...
Definition Casting.h:547
constexpr T divideCeil(U Numerator, V Denominator)
Returns the integer ceil(Numerator / Denominator).
Definition MathExtras.h:394
LLVM_ABI LLVM_READNONE LLT getCoverTy(LLT OrigTy, LLT TargetTy)
Return smallest type that covers both OrigTy and TargetTy and is multiple of TargetTy.
Definition Utils.cpp:1255
uint64_t alignTo(uint64_t Size, Align A)
Returns a multiple of A needed to store Size bytes.
Definition Alignment.h:144
bool isInTailCallPosition(const CallBase &Call, const TargetMachine &TM, bool ReturnsFirstArg=false)
Test if the given instruction is in a position to be optimized with a tail-call.
Definition Analysis.cpp:543
ArrayRef(const T &OneElt) -> ArrayRef< T >
OutputIt copy(R &&Range, OutputIt Out)
Definition STLExtras.h:1835
decltype(auto) cast(const From &Val)
cast<X> - Return the argument parameter cast to the specified type.
Definition Casting.h:559
Align commonAlignment(Align A, uint64_t Offset)
Returns the alignment that satisfies both alignments.
Definition Alignment.h:201
LLVM_ABI LLVM_READNONE LLT getGCDType(LLT OrigTy, LLT TargetTy)
Return a type where the total size is the greatest common divisor of OrigTy and TargetTy.
Definition Utils.cpp:1276
LLVM_ABI LLT getLLTForType(Type &Ty, const DataLayout &DL)
Construct a low-level type based on an LLVM type.
LLVM_ABI Align inferAlignFromPtrInfo(MachineFunction &MF, const MachinePointerInfo &MPO)
Definition Utils.cpp:898
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition Alignment.h:39
const Value * OrigValue
Optionally track the original IR value for the argument.
SmallVector< Register, 4 > Regs
unsigned OrigArgIndex
Index original Function's argument.
static const unsigned NoArgIndex
Sentinel value for implicit machine-level input arguments.
SmallVector< ISD::ArgFlagsTy, 4 > Flags
void assignValueToReg(Register ValVReg, Register PhysReg, const CCValAssign &VA) override
Provides a default implementation for argument handling.
Register buildExtensionHint(const CCValAssign &VA, Register SrcReg, LLT NarrowTy)
Insert G_ASSERT_ZEXT/G_ASSERT_SEXT or other hint instruction based on VA, returning the new register ...
Argument handling is mostly uniform between the four places that make these decisions: function forma...
virtual bool assignArg(unsigned ValNo, EVT OrigVT, MVT ValVT, MVT LocVT, CCValAssign::LocInfo LocInfo, const ArgInfo &Info, ISD::ArgFlagsTy Flags, CCState &State)
Wrap call to (typically tablegenerated CCAssignFn).
void copyArgumentMemory(const ArgInfo &Arg, Register DstPtr, Register SrcPtr, const MachinePointerInfo &DstPtrInfo, Align DstAlign, const MachinePointerInfo &SrcPtrInfo, Align SrcAlign, uint64_t MemSize, CCValAssign &VA) const
Do a memory copy of MemSize bytes from SrcPtr to DstPtr.
virtual Register getStackAddress(uint64_t MemSize, int64_t Offset, MachinePointerInfo &MPO, ISD::ArgFlagsTy Flags)=0
Materialize a VReg containing the address of the specified stack-based object.
virtual LLT getStackValueStoreType(const DataLayout &DL, const CCValAssign &VA, ISD::ArgFlagsTy Flags) const
Return the in-memory size to write for the argument at VA.
bool isIncomingArgumentHandler() const
Returns true if the handler is dealing with incoming arguments, i.e.
virtual void assignValueToAddress(Register ValVReg, Register Addr, LLT MemTy, const MachinePointerInfo &MPO, const CCValAssign &VA)=0
The specified value has been assigned to a stack location.
Register extendRegister(Register ValReg, const CCValAssign &VA, unsigned MaxSizeBits=0)
Extend a register to the location type given in VA, capped at extending to at most MaxSize bits.
virtual unsigned assignCustomValue(ArgInfo &Arg, ArrayRef< CCValAssign > VAs, std::function< void()> *Thunk=nullptr)
Handle custom values, which may be passed into one or more of VAs.
virtual void assignValueToReg(Register ValVReg, Register PhysReg, const CCValAssign &VA)=0
The specified value has been assigned to a physical register, handle the appropriate COPY (either to ...
Extended Value Type.
Definition ValueTypes.h:35
static LLVM_ABI EVT getEVT(Type *Ty, bool HandleUnknown=false)
Return the value type corresponding to the specified type.
LLVM_ABI Type * getTypeForEVT(LLVMContext &Context) const
This method returns an LLVM type corresponding to the specified EVT.
This class contains a discriminated union of information about pointers in memory operands,...
static LLVM_ABI MachinePointerInfo getUnknownStack(MachineFunction &MF)
Stack memory without other information.
static LLVM_ABI MachinePointerInfo getFixedStack(MachineFunction &MF, int FI, int64_t Offset=0)
Return a MachinePointerInfo record that refers to the specified FrameIndex.
This struct is a compact representation of a valid (power of two) or undefined (0) alignment.
Definition Alignment.h:106