LLVM 19.0.0git
TargetTransformInfoImpl.h
Go to the documentation of this file.
1//===- TargetTransformInfoImpl.h --------------------------------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8/// \file
9/// This file provides helpers for the implementation of
10/// a TargetTransformInfo-conforming class.
11///
12//===----------------------------------------------------------------------===//
13
14#ifndef LLVM_ANALYSIS_TARGETTRANSFORMINFOIMPL_H
15#define LLVM_ANALYSIS_TARGETTRANSFORMINFOIMPL_H
16
20#include "llvm/IR/DataLayout.h"
23#include "llvm/IR/Operator.h"
25#include <optional>
26#include <utility>
27
28namespace llvm {
29
30class Function;
31
32/// Base class for use as a mix-in that aids implementing
33/// a TargetTransformInfo-compatible class.
35protected:
37
38 const DataLayout &DL;
39
41
42public:
43 // Provide value semantics. MSVC requires that we spell all of these out.
46
47 const DataLayout &getDataLayout() const { return DL; }
48
49 InstructionCost getGEPCost(Type *PointeeType, const Value *Ptr,
52 // In the basic model, we just assume that all-constant GEPs will be folded
53 // into their uses via addressing modes.
54 for (const Value *Operand : Operands)
55 if (!isa<Constant>(Operand))
56 return TTI::TCC_Basic;
57
58 return TTI::TCC_Free;
59 }
60
62 unsigned &JTSize,
64 BlockFrequencyInfo *BFI) const {
65 (void)PSI;
66 (void)BFI;
67 JTSize = 0;
68 return SI.getNumCases();
69 }
70
71 unsigned getInliningThresholdMultiplier() const { return 1; }
74 return 8;
75 }
76 unsigned adjustInliningThreshold(const CallBase *CB) const { return 0; }
77 unsigned getCallerAllocaCost(const CallBase *CB, const AllocaInst *AI) const {
78 return 0;
79 };
80
81 int getInlinerVectorBonusPercent() const { return 150; }
82
84 return TTI::TCC_Expensive;
85 }
86
88 return 64;
89 }
90
91 // Although this default value is arbitrary, it is not random. It is assumed
92 // that a condition that evaluates the same way by a higher percentage than
93 // this is best represented as control flow. Therefore, the default value N
94 // should be set such that the win from N% correct executions is greater than
95 // the loss from (100 - N)% mispredicted executions for the majority of
96 // intended targets.
98 return BranchProbability(99, 100);
99 }
100
101 bool hasBranchDivergence(const Function *F = nullptr) const { return false; }
102
103 bool isSourceOfDivergence(const Value *V) const { return false; }
104
105 bool isAlwaysUniform(const Value *V) const { return false; }
106
107 bool isValidAddrSpaceCast(unsigned FromAS, unsigned ToAS) const {
108 return false;
109 }
110
111 bool addrspacesMayAlias(unsigned AS0, unsigned AS1) const {
112 return true;
113 }
114
115 unsigned getFlatAddressSpace() const { return -1; }
116
118 Intrinsic::ID IID) const {
119 return false;
120 }
121
122 bool isNoopAddrSpaceCast(unsigned, unsigned) const { return false; }
124 return AS == 0;
125 };
126
127 unsigned getAssumedAddrSpace(const Value *V) const { return -1; }
128
129 bool isSingleThreaded() const { return false; }
130
131 std::pair<const Value *, unsigned>
133 return std::make_pair(nullptr, -1);
134 }
135
137 Value *NewV) const {
138 return nullptr;
139 }
140
141 bool isLoweredToCall(const Function *F) const {
142 assert(F && "A concrete function must be provided to this routine.");
143
144 // FIXME: These should almost certainly not be handled here, and instead
145 // handled with the help of TLI or the target itself. This was largely
146 // ported from existing analysis heuristics here so that such refactorings
147 // can take place in the future.
148
149 if (F->isIntrinsic())
150 return false;
151
152 if (F->hasLocalLinkage() || !F->hasName())
153 return true;
154
155 StringRef Name = F->getName();
156
157 // These will all likely lower to a single selection DAG node.
158 if (Name == "copysign" || Name == "copysignf" || Name == "copysignl" ||
159 Name == "fabs" || Name == "fabsf" || Name == "fabsl" || Name == "sin" ||
160 Name == "fmin" || Name == "fminf" || Name == "fminl" ||
161 Name == "fmax" || Name == "fmaxf" || Name == "fmaxl" ||
162 Name == "sinf" || Name == "sinl" || Name == "cos" || Name == "cosf" ||
163 Name == "cosl" || Name == "sqrt" || Name == "sqrtf" || Name == "sqrtl")
164 return false;
165
166 // These are all likely to be optimized into something smaller.
167 if (Name == "pow" || Name == "powf" || Name == "powl" || Name == "exp2" ||
168 Name == "exp2l" || Name == "exp2f" || Name == "floor" ||
169 Name == "floorf" || Name == "ceil" || Name == "round" ||
170 Name == "ffs" || Name == "ffsl" || Name == "abs" || Name == "labs" ||
171 Name == "llabs")
172 return false;
173
174 return true;
175 }
176
179 HardwareLoopInfo &HWLoopInfo) const {
180 return false;
181 }
182
183 bool preferPredicateOverEpilogue(TailFoldingInfo *TFI) const { return false; }
184
186 getPreferredTailFoldingStyle(bool IVUpdateMayOverflow = true) const {
188 }
189
190 std::optional<Instruction *> instCombineIntrinsic(InstCombiner &IC,
191 IntrinsicInst &II) const {
192 return std::nullopt;
193 }
194
195 std::optional<Value *>
197 APInt DemandedMask, KnownBits &Known,
198 bool &KnownBitsComputed) const {
199 return std::nullopt;
200 }
201
203 InstCombiner &IC, IntrinsicInst &II, APInt DemandedElts, APInt &UndefElts,
204 APInt &UndefElts2, APInt &UndefElts3,
205 std::function<void(Instruction *, unsigned, APInt, APInt &)>
206 SimplifyAndSetOp) const {
207 return std::nullopt;
208 }
209
212 OptimizationRemarkEmitter *) const {}
213
215 TTI::PeelingPreferences &) const {}
216
217 bool isLegalAddImmediate(int64_t Imm) const { return false; }
218
219 bool isLegalAddScalableImmediate(int64_t Imm) const { return false; }
220
221 bool isLegalICmpImmediate(int64_t Imm) const { return false; }
222
223 bool isLegalAddressingMode(Type *Ty, GlobalValue *BaseGV, int64_t BaseOffset,
224 bool HasBaseReg, int64_t Scale, unsigned AddrSpace,
225 Instruction *I = nullptr,
226 int64_t ScalableOffset = 0) const {
227 // Guess that only reg and reg+reg addressing is allowed. This heuristic is
228 // taken from the implementation of LSR.
229 return !BaseGV && BaseOffset == 0 && (Scale == 0 || Scale == 1);
230 }
231
232 bool isLSRCostLess(const TTI::LSRCost &C1, const TTI::LSRCost &C2) const {
233 return std::tie(C1.NumRegs, C1.AddRecCost, C1.NumIVMuls, C1.NumBaseAdds,
234 C1.ScaleCost, C1.ImmCost, C1.SetupCost) <
235 std::tie(C2.NumRegs, C2.AddRecCost, C2.NumIVMuls, C2.NumBaseAdds,
236 C2.ScaleCost, C2.ImmCost, C2.SetupCost);
237 }
238
239 bool isNumRegsMajorCostOfLSR() const { return true; }
240
241 bool shouldFoldTerminatingConditionAfterLSR() const { return false; }
242
243 bool isProfitableLSRChainElement(Instruction *I) const { return false; }
244
245 bool canMacroFuseCmp() const { return false; }
246
249 TargetLibraryInfo *LibInfo) const {
250 return false;
251 }
252
255 return TTI::AMK_None;
256 }
257
258 bool isLegalMaskedStore(Type *DataType, Align Alignment) const {
259 return false;
260 }
261
262 bool isLegalMaskedLoad(Type *DataType, Align Alignment) const {
263 return false;
264 }
265
266 bool isLegalNTStore(Type *DataType, Align Alignment) const {
267 // By default, assume nontemporal memory stores are available for stores
268 // that are aligned and have a size that is a power of 2.
269 unsigned DataSize = DL.getTypeStoreSize(DataType);
270 return Alignment >= DataSize && isPowerOf2_32(DataSize);
271 }
272
273 bool isLegalNTLoad(Type *DataType, Align Alignment) const {
274 // By default, assume nontemporal memory loads are available for loads that
275 // are aligned and have a size that is a power of 2.
276 unsigned DataSize = DL.getTypeStoreSize(DataType);
277 return Alignment >= DataSize && isPowerOf2_32(DataSize);
278 }
279
280 bool isLegalBroadcastLoad(Type *ElementTy, ElementCount NumElements) const {
281 return false;
282 }
283
284 bool isLegalMaskedScatter(Type *DataType, Align Alignment) const {
285 return false;
286 }
287
288 bool isLegalMaskedGather(Type *DataType, Align Alignment) const {
289 return false;
290 }
291
292 bool forceScalarizeMaskedGather(VectorType *DataType, Align Alignment) const {
293 return false;
294 }
295
297 Align Alignment) const {
298 return false;
299 }
300
301 bool isLegalMaskedCompressStore(Type *DataType, Align Alignment) const {
302 return false;
303 }
304
305 bool isLegalAltInstr(VectorType *VecTy, unsigned Opcode0, unsigned Opcode1,
306 const SmallBitVector &OpcodeMask) const {
307 return false;
308 }
309
310 bool isLegalMaskedExpandLoad(Type *DataType, Align Alignment) const {
311 return false;
312 }
313
314 bool isLegalStridedLoadStore(Type *DataType, Align Alignment) const {
315 return false;
316 }
317
318 bool enableOrderedReductions() const { return false; }
319
320 bool hasDivRemOp(Type *DataType, bool IsSigned) const { return false; }
321
322 bool hasVolatileVariant(Instruction *I, unsigned AddrSpace) const {
323 return false;
324 }
325
326 bool prefersVectorizedAddressing() const { return true; }
327
329 int64_t BaseOffset, bool HasBaseReg,
330 int64_t Scale,
331 unsigned AddrSpace) const {
332 // Guess that all legal addressing mode are free.
333 if (isLegalAddressingMode(Ty, BaseGV, BaseOffset, HasBaseReg, Scale,
334 AddrSpace))
335 return 0;
336 return -1;
337 }
338
339 bool LSRWithInstrQueries() const { return false; }
340
341 bool isTruncateFree(Type *Ty1, Type *Ty2) const { return false; }
342
343 bool isProfitableToHoist(Instruction *I) const { return true; }
344
345 bool useAA() const { return false; }
346
347 bool isTypeLegal(Type *Ty) const { return false; }
348
349 unsigned getRegUsageForType(Type *Ty) const { return 1; }
350
351 bool shouldBuildLookupTables() const { return true; }
352
353 bool shouldBuildLookupTablesForConstant(Constant *C) const { return true; }
354
355 bool shouldBuildRelLookupTables() const { return false; }
356
357 bool useColdCCForColdCall(Function &F) const { return false; }
358
360 const APInt &DemandedElts,
361 bool Insert, bool Extract,
363 return 0;
364 }
365
370 return 0;
371 }
372
373 bool supportsEfficientVectorElementLoadStore() const { return false; }
374
375 bool supportsTailCalls() const { return true; }
376
377 bool enableAggressiveInterleaving(bool LoopHasReductions) const {
378 return false;
379 }
380
382 bool IsZeroCmp) const {
383 return {};
384 }
385
386 bool enableSelectOptimize() const { return true; }
387
389 // If the select is a logical-and/logical-or then it is better treated as a
390 // and/or by the backend.
391 using namespace llvm::PatternMatch;
392 return isa<SelectInst>(I) &&
395 }
396
397 bool enableInterleavedAccessVectorization() const { return false; }
398
399 bool enableMaskedInterleavedAccessVectorization() const { return false; }
400
401 bool isFPVectorizationPotentiallyUnsafe() const { return false; }
402
404 unsigned AddressSpace, Align Alignment,
405 unsigned *Fast) const {
406 return false;
407 }
408
409 TTI::PopcntSupportKind getPopcntSupport(unsigned IntTyWidthInBit) const {
410 return TTI::PSK_Software;
411 }
412
413 bool haveFastSqrt(Type *Ty) const { return false; }
414
415 bool isExpensiveToSpeculativelyExecute(const Instruction *I) { return true; }
416
417 bool isFCmpOrdCheaperThanFCmpZero(Type *Ty) const { return true; }
418
421 }
422
423 InstructionCost getIntImmCodeSizeCost(unsigned Opcode, unsigned Idx,
424 const APInt &Imm, Type *Ty) const {
425 return 0;
426 }
427
430 return TTI::TCC_Basic;
431 }
432
433 InstructionCost getIntImmCostInst(unsigned Opcode, unsigned Idx,
434 const APInt &Imm, Type *Ty,
436 Instruction *Inst = nullptr) const {
437 return TTI::TCC_Free;
438 }
439
441 const APInt &Imm, Type *Ty,
443 return TTI::TCC_Free;
444 }
445
447 const Function &Fn) const {
448 return false;
449 }
450
451 unsigned getNumberOfRegisters(unsigned ClassID) const { return 8; }
452
453 unsigned getRegisterClassForType(bool Vector, Type *Ty = nullptr) const {
454 return Vector ? 1 : 0;
455 };
456
457 const char *getRegisterClassName(unsigned ClassID) const {
458 switch (ClassID) {
459 default:
460 return "Generic::Unknown Register Class";
461 case 0:
462 return "Generic::ScalarRC";
463 case 1:
464 return "Generic::VectorRC";
465 }
466 }
467
469 return TypeSize::getFixed(32);
470 }
471
472 unsigned getMinVectorRegisterBitWidth() const { return 128; }
473
474 std::optional<unsigned> getMaxVScale() const { return std::nullopt; }
475 std::optional<unsigned> getVScaleForTuning() const { return std::nullopt; }
476 bool isVScaleKnownToBeAPowerOfTwo() const { return false; }
477
478 bool
480 return false;
481 }
482
483 ElementCount getMinimumVF(unsigned ElemWidth, bool IsScalable) const {
484 return ElementCount::get(0, IsScalable);
485 }
486
487 unsigned getMaximumVF(unsigned ElemWidth, unsigned Opcode) const { return 0; }
488 unsigned getStoreMinimumVF(unsigned VF, Type *, Type *) const { return VF; }
489
491 const Instruction &I, bool &AllowPromotionWithoutCommonHeader) const {
492 AllowPromotionWithoutCommonHeader = false;
493 return false;
494 }
495
496 unsigned getCacheLineSize() const { return 0; }
497 std::optional<unsigned>
499 switch (Level) {
501 [[fallthrough]];
503 return std::nullopt;
504 }
505 llvm_unreachable("Unknown TargetTransformInfo::CacheLevel");
506 }
507
508 std::optional<unsigned>
510 switch (Level) {
512 [[fallthrough]];
514 return std::nullopt;
515 }
516
517 llvm_unreachable("Unknown TargetTransformInfo::CacheLevel");
518 }
519
520 std::optional<unsigned> getMinPageSize() const { return {}; }
521
522 unsigned getPrefetchDistance() const { return 0; }
523 unsigned getMinPrefetchStride(unsigned NumMemAccesses,
524 unsigned NumStridedMemAccesses,
525 unsigned NumPrefetches, bool HasCall) const {
526 return 1;
527 }
528 unsigned getMaxPrefetchIterationsAhead() const { return UINT_MAX; }
529 bool enableWritePrefetching() const { return false; }
530 bool shouldPrefetchAddressSpace(unsigned AS) const { return !AS; }
531
532 unsigned getMaxInterleaveFactor(ElementCount VF) const { return 1; }
533
535 unsigned Opcode, Type *Ty, TTI::TargetCostKind CostKind,
538 const Instruction *CxtI = nullptr) const {
539 // Widenable conditions will eventually lower into constants, so some
540 // operations with them will be trivially optimized away.
541 auto IsWidenableCondition = [](const Value *V) {
542 if (auto *II = dyn_cast<IntrinsicInst>(V))
543 if (II->getIntrinsicID() == Intrinsic::experimental_widenable_condition)
544 return true;
545 return false;
546 };
547 // FIXME: A number of transformation tests seem to require these values
548 // which seems a little odd for how arbitary there are.
549 switch (Opcode) {
550 default:
551 break;
552 case Instruction::FDiv:
553 case Instruction::FRem:
554 case Instruction::SDiv:
555 case Instruction::SRem:
556 case Instruction::UDiv:
557 case Instruction::URem:
558 // FIXME: Unlikely to be true for CodeSize.
559 return TTI::TCC_Expensive;
560 case Instruction::And:
561 case Instruction::Or:
562 if (any_of(Args, IsWidenableCondition))
563 return TTI::TCC_Free;
564 break;
565 }
566
567 // Assume a 3cy latency for fp arithmetic ops.
569 if (Ty->getScalarType()->isFloatingPointTy())
570 return 3;
571
572 return 1;
573 }
574
576 unsigned Opcode1,
577 const SmallBitVector &OpcodeMask,
580 }
581
583 ArrayRef<int> Mask,
585 VectorType *SubTp,
586 ArrayRef<const Value *> Args = std::nullopt,
587 const Instruction *CxtI = nullptr) const {
588 return 1;
589 }
590
591 InstructionCost getCastInstrCost(unsigned Opcode, Type *Dst, Type *Src,
594 const Instruction *I) const {
595 switch (Opcode) {
596 default:
597 break;
598 case Instruction::IntToPtr: {
599 unsigned SrcSize = Src->getScalarSizeInBits();
600 if (DL.isLegalInteger(SrcSize) &&
601 SrcSize <= DL.getPointerTypeSizeInBits(Dst))
602 return 0;
603 break;
604 }
605 case Instruction::PtrToInt: {
606 unsigned DstSize = Dst->getScalarSizeInBits();
607 if (DL.isLegalInteger(DstSize) &&
608 DstSize >= DL.getPointerTypeSizeInBits(Src))
609 return 0;
610 break;
611 }
612 case Instruction::BitCast:
613 if (Dst == Src || (Dst->isPointerTy() && Src->isPointerTy()))
614 // Identity and pointer-to-pointer casts are free.
615 return 0;
616 break;
617 case Instruction::Trunc: {
618 // trunc to a native type is free (assuming the target has compare and
619 // shift-right of the same width).
620 TypeSize DstSize = DL.getTypeSizeInBits(Dst);
621 if (!DstSize.isScalable() && DL.isLegalInteger(DstSize.getFixedValue()))
622 return 0;
623 break;
624 }
625 }
626 return 1;
627 }
628
630 VectorType *VecTy,
631 unsigned Index) const {
632 return 1;
633 }
634
636 const Instruction *I = nullptr) const {
637 // A phi would be free, unless we're costing the throughput because it
638 // will require a register.
639 if (Opcode == Instruction::PHI && CostKind != TTI::TCK_RecipThroughput)
640 return 0;
641 return 1;
642 }
643
644 InstructionCost getCmpSelInstrCost(unsigned Opcode, Type *ValTy, Type *CondTy,
645 CmpInst::Predicate VecPred,
647 const Instruction *I) const {
648 return 1;
649 }
650
653 unsigned Index, Value *Op0,
654 Value *Op1) const {
655 return 1;
656 }
657
660 unsigned Index) const {
661 return 1;
662 }
663
664 unsigned getReplicationShuffleCost(Type *EltTy, int ReplicationFactor, int VF,
665 const APInt &DemandedDstElts,
667 return 1;
668 }
669
670 InstructionCost getMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment,
671 unsigned AddressSpace,
674 const Instruction *I) const {
675 return 1;
676 }
677
678 InstructionCost getVPMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment,
679 unsigned AddressSpace,
681 const Instruction *I) const {
682 return 1;
683 }
684
686 Align Alignment, unsigned AddressSpace,
688 return 1;
689 }
690
692 const Value *Ptr, bool VariableMask,
693 Align Alignment,
695 const Instruction *I = nullptr) const {
696 return 1;
697 }
698
700 const Value *Ptr, bool VariableMask,
701 Align Alignment,
703 const Instruction *I = nullptr) const {
705 }
706
708 unsigned Opcode, Type *VecTy, unsigned Factor, ArrayRef<unsigned> Indices,
709 Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind,
710 bool UseMaskForCond, bool UseMaskForGaps) const {
711 return 1;
712 }
713
716 switch (ICA.getID()) {
717 default:
718 break;
719 case Intrinsic::allow_runtime_check:
720 case Intrinsic::allow_ubsan_check:
721 case Intrinsic::annotation:
722 case Intrinsic::assume:
723 case Intrinsic::sideeffect:
724 case Intrinsic::pseudoprobe:
725 case Intrinsic::arithmetic_fence:
726 case Intrinsic::dbg_assign:
727 case Intrinsic::dbg_declare:
728 case Intrinsic::dbg_value:
729 case Intrinsic::dbg_label:
730 case Intrinsic::invariant_start:
731 case Intrinsic::invariant_end:
732 case Intrinsic::launder_invariant_group:
733 case Intrinsic::strip_invariant_group:
734 case Intrinsic::is_constant:
735 case Intrinsic::lifetime_start:
736 case Intrinsic::lifetime_end:
737 case Intrinsic::experimental_noalias_scope_decl:
738 case Intrinsic::objectsize:
739 case Intrinsic::ptr_annotation:
740 case Intrinsic::var_annotation:
741 case Intrinsic::experimental_gc_result:
742 case Intrinsic::experimental_gc_relocate:
743 case Intrinsic::coro_alloc:
744 case Intrinsic::coro_begin:
745 case Intrinsic::coro_free:
746 case Intrinsic::coro_end:
747 case Intrinsic::coro_frame:
748 case Intrinsic::coro_size:
749 case Intrinsic::coro_align:
750 case Intrinsic::coro_suspend:
751 case Intrinsic::coro_subfn_addr:
752 case Intrinsic::threadlocal_address:
753 case Intrinsic::experimental_widenable_condition:
754 case Intrinsic::ssa_copy:
755 // These intrinsics don't actually represent code after lowering.
756 return 0;
757 }
758 return 1;
759 }
760
764 return 1;
765 }
766
767 // Assume that we have a register of the right size for the type.
768 unsigned getNumberOfParts(Type *Tp) const { return 1; }
769
771 const SCEV *) const {
772 return 0;
773 }
774
776 std::optional<FastMathFlags> FMF,
777 TTI::TargetCostKind) const {
778 return 1;
779 }
780
783 TTI::TargetCostKind) const {
784 return 1;
785 }
786
787 InstructionCost getExtendedReductionCost(unsigned Opcode, bool IsUnsigned,
788 Type *ResTy, VectorType *Ty,
789 FastMathFlags FMF,
791 return 1;
792 }
793
795 VectorType *Ty,
797 return 1;
798 }
799
801 return 0;
802 }
803
805 return false;
806 }
807
809 // Note for overrides: You must ensure for all element unordered-atomic
810 // memory intrinsics that all power-of-2 element sizes up to, and
811 // including, the return value of this method have a corresponding
812 // runtime lib call. These runtime lib call definitions can be found
813 // in RuntimeLibcalls.h
814 return 0;
815 }
816
818 Type *ExpectedType) const {
819 return nullptr;
820 }
821
822 Type *
824 unsigned SrcAddrSpace, unsigned DestAddrSpace,
825 unsigned SrcAlign, unsigned DestAlign,
826 std::optional<uint32_t> AtomicElementSize) const {
827 return AtomicElementSize ? Type::getIntNTy(Context, *AtomicElementSize * 8)
829 }
830
832 SmallVectorImpl<Type *> &OpsOut, LLVMContext &Context,
833 unsigned RemainingBytes, unsigned SrcAddrSpace, unsigned DestAddrSpace,
834 unsigned SrcAlign, unsigned DestAlign,
835 std::optional<uint32_t> AtomicCpySize) const {
836 unsigned OpSizeInBytes = AtomicCpySize ? *AtomicCpySize : 1;
837 Type *OpType = Type::getIntNTy(Context, OpSizeInBytes * 8);
838 for (unsigned i = 0; i != RemainingBytes; i += OpSizeInBytes)
839 OpsOut.push_back(OpType);
840 }
841
842 bool areInlineCompatible(const Function *Caller,
843 const Function *Callee) const {
844 return (Caller->getFnAttribute("target-cpu") ==
845 Callee->getFnAttribute("target-cpu")) &&
846 (Caller->getFnAttribute("target-features") ==
847 Callee->getFnAttribute("target-features"));
848 }
849
850 unsigned getInlineCallPenalty(const Function *F, const CallBase &Call,
851 unsigned DefaultCallPenalty) const {
852 return DefaultCallPenalty;
853 }
854
855 bool areTypesABICompatible(const Function *Caller, const Function *Callee,
856 const ArrayRef<Type *> &Types) const {
857 return (Caller->getFnAttribute("target-cpu") ==
858 Callee->getFnAttribute("target-cpu")) &&
859 (Caller->getFnAttribute("target-features") ==
860 Callee->getFnAttribute("target-features"));
861 }
862
864 const DataLayout &DL) const {
865 return false;
866 }
867
869 const DataLayout &DL) const {
870 return false;
871 }
872
873 unsigned getLoadStoreVecRegBitWidth(unsigned AddrSpace) const { return 128; }
874
875 bool isLegalToVectorizeLoad(LoadInst *LI) const { return true; }
876
877 bool isLegalToVectorizeStore(StoreInst *SI) const { return true; }
878
879 bool isLegalToVectorizeLoadChain(unsigned ChainSizeInBytes, Align Alignment,
880 unsigned AddrSpace) const {
881 return true;
882 }
883
884 bool isLegalToVectorizeStoreChain(unsigned ChainSizeInBytes, Align Alignment,
885 unsigned AddrSpace) const {
886 return true;
887 }
888
890 ElementCount VF) const {
891 return true;
892 }
893
894 bool isElementTypeLegalForScalableVector(Type *Ty) const { return true; }
895
896 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize,
897 unsigned ChainSizeInBytes,
898 VectorType *VecTy) const {
899 return VF;
900 }
901
902 unsigned getStoreVectorFactor(unsigned VF, unsigned StoreSize,
903 unsigned ChainSizeInBytes,
904 VectorType *VecTy) const {
905 return VF;
906 }
907
908 bool preferInLoopReduction(unsigned Opcode, Type *Ty,
909 TTI::ReductionFlags Flags) const {
910 return false;
911 }
912
913 bool preferPredicatedReductionSelect(unsigned Opcode, Type *Ty,
914 TTI::ReductionFlags Flags) const {
915 return false;
916 }
917
919 return true;
920 }
921
922 bool shouldExpandReduction(const IntrinsicInst *II) const { return true; }
923
924 unsigned getGISelRematGlobalCost() const { return 1; }
925
926 unsigned getMinTripCountTailFoldingThreshold() const { return 0; }
927
928 bool supportsScalableVectors() const { return false; }
929
930 bool enableScalableVectorization() const { return false; }
931
932 bool hasActiveVectorLength(unsigned Opcode, Type *DataType,
933 Align Alignment) const {
934 return false;
935 }
936
941 /* OperatorStrategy */ TargetTransformInfo::VPLegalization::Convert);
942 }
943
944 bool hasArmWideBranch(bool) const { return false; }
945
946 unsigned getMaxNumArgs() const { return UINT_MAX; }
947
948protected:
949 // Obtain the minimum required size to hold the value (without the sign)
950 // In case of a vector it returns the min required size for one element.
951 unsigned minRequiredElementSize(const Value *Val, bool &isSigned) const {
952 if (isa<ConstantDataVector>(Val) || isa<ConstantVector>(Val)) {
953 const auto *VectorValue = cast<Constant>(Val);
954
955 // In case of a vector need to pick the max between the min
956 // required size for each element
957 auto *VT = cast<FixedVectorType>(Val->getType());
958
959 // Assume unsigned elements
960 isSigned = false;
961
962 // The max required size is the size of the vector element type
963 unsigned MaxRequiredSize =
964 VT->getElementType()->getPrimitiveSizeInBits().getFixedValue();
965
966 unsigned MinRequiredSize = 0;
967 for (unsigned i = 0, e = VT->getNumElements(); i < e; ++i) {
968 if (auto *IntElement =
969 dyn_cast<ConstantInt>(VectorValue->getAggregateElement(i))) {
970 bool signedElement = IntElement->getValue().isNegative();
971 // Get the element min required size.
972 unsigned ElementMinRequiredSize =
973 IntElement->getValue().getSignificantBits() - 1;
974 // In case one element is signed then all the vector is signed.
975 isSigned |= signedElement;
976 // Save the max required bit size between all the elements.
977 MinRequiredSize = std::max(MinRequiredSize, ElementMinRequiredSize);
978 } else {
979 // not an int constant element
980 return MaxRequiredSize;
981 }
982 }
983 return MinRequiredSize;
984 }
985
986 if (const auto *CI = dyn_cast<ConstantInt>(Val)) {
987 isSigned = CI->getValue().isNegative();
988 return CI->getValue().getSignificantBits() - 1;
989 }
990
991 if (const auto *Cast = dyn_cast<SExtInst>(Val)) {
992 isSigned = true;
993 return Cast->getSrcTy()->getScalarSizeInBits() - 1;
994 }
995
996 if (const auto *Cast = dyn_cast<ZExtInst>(Val)) {
997 isSigned = false;
998 return Cast->getSrcTy()->getScalarSizeInBits();
999 }
1000
1001 isSigned = false;
1002 return Val->getType()->getScalarSizeInBits();
1003 }
1004
1005 bool isStridedAccess(const SCEV *Ptr) const {
1006 return Ptr && isa<SCEVAddRecExpr>(Ptr);
1007 }
1008
1010 const SCEV *Ptr) const {
1011 if (!isStridedAccess(Ptr))
1012 return nullptr;
1013 const SCEVAddRecExpr *AddRec = cast<SCEVAddRecExpr>(Ptr);
1014 return dyn_cast<SCEVConstant>(AddRec->getStepRecurrence(*SE));
1015 }
1016
1018 int64_t MergeDistance) const {
1019 const SCEVConstant *Step = getConstantStrideStep(SE, Ptr);
1020 if (!Step)
1021 return false;
1022 APInt StrideVal = Step->getAPInt();
1023 if (StrideVal.getBitWidth() > 64)
1024 return false;
1025 // FIXME: Need to take absolute value for negative stride case.
1026 return StrideVal.getSExtValue() < MergeDistance;
1027 }
1028};
1029
1030/// CRTP base class for use as a mix-in that aids implementing
1031/// a TargetTransformInfo-compatible class.
1032template <typename T>
1034private:
1036
1037protected:
1039
1040public:
1042
1046 assert(PointeeType && Ptr && "can't get GEPCost of nullptr");
1047 auto *BaseGV = dyn_cast<GlobalValue>(Ptr->stripPointerCasts());
1048 bool HasBaseReg = (BaseGV == nullptr);
1049
1050 auto PtrSizeBits = DL.getPointerTypeSizeInBits(Ptr->getType());
1051 APInt BaseOffset(PtrSizeBits, 0);
1052 int64_t Scale = 0;
1053
1054 auto GTI = gep_type_begin(PointeeType, Operands);
1055 Type *TargetType = nullptr;
1056
1057 // Handle the case where the GEP instruction has a single operand,
1058 // the basis, therefore TargetType is a nullptr.
1059 if (Operands.empty())
1060 return !BaseGV ? TTI::TCC_Free : TTI::TCC_Basic;
1061
1062 for (auto I = Operands.begin(); I != Operands.end(); ++I, ++GTI) {
1063 TargetType = GTI.getIndexedType();
1064 // We assume that the cost of Scalar GEP with constant index and the
1065 // cost of Vector GEP with splat constant index are the same.
1066 const ConstantInt *ConstIdx = dyn_cast<ConstantInt>(*I);
1067 if (!ConstIdx)
1068 if (auto Splat = getSplatValue(*I))
1069 ConstIdx = dyn_cast<ConstantInt>(Splat);
1070 if (StructType *STy = GTI.getStructTypeOrNull()) {
1071 // For structures the index is always splat or scalar constant
1072 assert(ConstIdx && "Unexpected GEP index");
1073 uint64_t Field = ConstIdx->getZExtValue();
1074 BaseOffset += DL.getStructLayout(STy)->getElementOffset(Field);
1075 } else {
1076 // If this operand is a scalable type, bail out early.
1077 // TODO: Make isLegalAddressingMode TypeSize aware.
1078 if (TargetType->isScalableTy())
1079 return TTI::TCC_Basic;
1080 int64_t ElementSize =
1081 GTI.getSequentialElementStride(DL).getFixedValue();
1082 if (ConstIdx) {
1083 BaseOffset +=
1084 ConstIdx->getValue().sextOrTrunc(PtrSizeBits) * ElementSize;
1085 } else {
1086 // Needs scale register.
1087 if (Scale != 0)
1088 // No addressing mode takes two scale registers.
1089 return TTI::TCC_Basic;
1090 Scale = ElementSize;
1091 }
1092 }
1093 }
1094
1095 // If we haven't been provided a hint, use the target type for now.
1096 //
1097 // TODO: Take a look at potentially removing this: This is *slightly* wrong
1098 // as it's possible to have a GEP with a foldable target type but a memory
1099 // access that isn't foldable. For example, this load isn't foldable on
1100 // RISC-V:
1101 //
1102 // %p = getelementptr i32, ptr %base, i32 42
1103 // %x = load <2 x i32>, ptr %p
1104 if (!AccessType)
1105 AccessType = TargetType;
1106
1107 // If the final address of the GEP is a legal addressing mode for the given
1108 // access type, then we can fold it into its users.
1109 if (static_cast<T *>(this)->isLegalAddressingMode(
1110 AccessType, const_cast<GlobalValue *>(BaseGV),
1111 BaseOffset.sextOrTrunc(64).getSExtValue(), HasBaseReg, Scale,
1112 Ptr->getType()->getPointerAddressSpace()))
1113 return TTI::TCC_Free;
1114
1115 // TODO: Instead of returning TCC_Basic here, we should use
1116 // getArithmeticInstrCost. Or better yet, provide a hook to let the target
1117 // model it.
1118 return TTI::TCC_Basic;
1119 }
1120
1122 const Value *Base,
1124 Type *AccessTy,
1127 // In the basic model we take into account GEP instructions only
1128 // (although here can come alloca instruction, a value, constants and/or
1129 // constant expressions, PHIs, bitcasts ... whatever allowed to be used as a
1130 // pointer). Typically, if Base is a not a GEP-instruction and all the
1131 // pointers are relative to the same base address, all the rest are
1132 // either GEP instructions, PHIs, bitcasts or constants. When we have same
1133 // base, we just calculate cost of each non-Base GEP as an ADD operation if
1134 // any their index is a non-const.
1135 // If no known dependecies between the pointers cost is calculated as a sum
1136 // of costs of GEP instructions.
1137 for (const Value *V : Ptrs) {
1138 const auto *GEP = dyn_cast<GetElementPtrInst>(V);
1139 if (!GEP)
1140 continue;
1141 if (Info.isSameBase() && V != Base) {
1142 if (GEP->hasAllConstantIndices())
1143 continue;
1144 Cost += static_cast<T *>(this)->getArithmeticInstrCost(
1145 Instruction::Add, GEP->getType(), CostKind,
1147 std::nullopt);
1148 } else {
1149 SmallVector<const Value *> Indices(GEP->indices());
1150 Cost += static_cast<T *>(this)->getGEPCost(GEP->getSourceElementType(),
1151 GEP->getPointerOperand(),
1152 Indices, AccessTy, CostKind);
1153 }
1154 }
1155 return Cost;
1156 }
1157
1161 using namespace llvm::PatternMatch;
1162
1163 auto *TargetTTI = static_cast<T *>(this);
1164 // Handle non-intrinsic calls, invokes, and callbr.
1165 // FIXME: Unlikely to be true for anything but CodeSize.
1166 auto *CB = dyn_cast<CallBase>(U);
1167 if (CB && !isa<IntrinsicInst>(U)) {
1168 if (const Function *F = CB->getCalledFunction()) {
1169 if (!TargetTTI->isLoweredToCall(F))
1170 return TTI::TCC_Basic; // Give a basic cost if it will be lowered
1171
1172 return TTI::TCC_Basic * (F->getFunctionType()->getNumParams() + 1);
1173 }
1174 // For indirect or other calls, scale cost by number of arguments.
1175 return TTI::TCC_Basic * (CB->arg_size() + 1);
1176 }
1177
1178 Type *Ty = U->getType();
1179 unsigned Opcode = Operator::getOpcode(U);
1180 auto *I = dyn_cast<Instruction>(U);
1181 switch (Opcode) {
1182 default:
1183 break;
1184 case Instruction::Call: {
1185 assert(isa<IntrinsicInst>(U) && "Unexpected non-intrinsic call");
1186 auto *Intrinsic = cast<IntrinsicInst>(U);
1187 IntrinsicCostAttributes CostAttrs(Intrinsic->getIntrinsicID(), *CB);
1188 return TargetTTI->getIntrinsicInstrCost(CostAttrs, CostKind);
1189 }
1190 case Instruction::Br:
1191 case Instruction::Ret:
1192 case Instruction::PHI:
1193 case Instruction::Switch:
1194 return TargetTTI->getCFInstrCost(Opcode, CostKind, I);
1195 case Instruction::ExtractValue:
1196 case Instruction::Freeze:
1197 return TTI::TCC_Free;
1198 case Instruction::Alloca:
1199 if (cast<AllocaInst>(U)->isStaticAlloca())
1200 return TTI::TCC_Free;
1201 break;
1202 case Instruction::GetElementPtr: {
1203 const auto *GEP = cast<GEPOperator>(U);
1204 Type *AccessType = nullptr;
1205 // For now, only provide the AccessType in the simple case where the GEP
1206 // only has one user.
1207 if (GEP->hasOneUser() && I)
1208 AccessType = I->user_back()->getAccessType();
1209
1210 return TargetTTI->getGEPCost(GEP->getSourceElementType(),
1211 Operands.front(), Operands.drop_front(),
1212 AccessType, CostKind);
1213 }
1214 case Instruction::Add:
1215 case Instruction::FAdd:
1216 case Instruction::Sub:
1217 case Instruction::FSub:
1218 case Instruction::Mul:
1219 case Instruction::FMul:
1220 case Instruction::UDiv:
1221 case Instruction::SDiv:
1222 case Instruction::FDiv:
1223 case Instruction::URem:
1224 case Instruction::SRem:
1225 case Instruction::FRem:
1226 case Instruction::Shl:
1227 case Instruction::LShr:
1228 case Instruction::AShr:
1229 case Instruction::And:
1230 case Instruction::Or:
1231 case Instruction::Xor:
1232 case Instruction::FNeg: {
1234 TTI::OperandValueInfo Op2Info;
1235 if (Opcode != Instruction::FNeg)
1236 Op2Info = TTI::getOperandInfo(Operands[1]);
1237 return TargetTTI->getArithmeticInstrCost(Opcode, Ty, CostKind, Op1Info,
1238 Op2Info, Operands, I);
1239 }
1240 case Instruction::IntToPtr:
1241 case Instruction::PtrToInt:
1242 case Instruction::SIToFP:
1243 case Instruction::UIToFP:
1244 case Instruction::FPToUI:
1245 case Instruction::FPToSI:
1246 case Instruction::Trunc:
1247 case Instruction::FPTrunc:
1248 case Instruction::BitCast:
1249 case Instruction::FPExt:
1250 case Instruction::SExt:
1251 case Instruction::ZExt:
1252 case Instruction::AddrSpaceCast: {
1253 Type *OpTy = Operands[0]->getType();
1254 return TargetTTI->getCastInstrCost(
1255 Opcode, Ty, OpTy, TTI::getCastContextHint(I), CostKind, I);
1256 }
1257 case Instruction::Store: {
1258 auto *SI = cast<StoreInst>(U);
1259 Type *ValTy = Operands[0]->getType();
1261 return TargetTTI->getMemoryOpCost(Opcode, ValTy, SI->getAlign(),
1262 SI->getPointerAddressSpace(), CostKind,
1263 OpInfo, I);
1264 }
1265 case Instruction::Load: {
1266 // FIXME: Arbitary cost which could come from the backend.
1268 return 4;
1269 auto *LI = cast<LoadInst>(U);
1270 Type *LoadType = U->getType();
1271 // If there is a non-register sized type, the cost estimation may expand
1272 // it to be several instructions to load into multiple registers on the
1273 // target. But, if the only use of the load is a trunc instruction to a
1274 // register sized type, the instruction selector can combine these
1275 // instructions to be a single load. So, in this case, we use the
1276 // destination type of the trunc instruction rather than the load to
1277 // accurately estimate the cost of this load instruction.
1278 if (CostKind == TTI::TCK_CodeSize && LI->hasOneUse() &&
1279 !LoadType->isVectorTy()) {
1280 if (const TruncInst *TI = dyn_cast<TruncInst>(*LI->user_begin()))
1281 LoadType = TI->getDestTy();
1282 }
1283 return TargetTTI->getMemoryOpCost(Opcode, LoadType, LI->getAlign(),
1285 {TTI::OK_AnyValue, TTI::OP_None}, I);
1286 }
1287 case Instruction::Select: {
1288 const Value *Op0, *Op1;
1289 if (match(U, m_LogicalAnd(m_Value(Op0), m_Value(Op1))) ||
1290 match(U, m_LogicalOr(m_Value(Op0), m_Value(Op1)))) {
1291 // select x, y, false --> x & y
1292 // select x, true, y --> x | y
1293 const auto Op1Info = TTI::getOperandInfo(Op0);
1294 const auto Op2Info = TTI::getOperandInfo(Op1);
1295 assert(Op0->getType()->getScalarSizeInBits() == 1 &&
1296 Op1->getType()->getScalarSizeInBits() == 1);
1297
1299 return TargetTTI->getArithmeticInstrCost(
1300 match(U, m_LogicalOr()) ? Instruction::Or : Instruction::And, Ty,
1301 CostKind, Op1Info, Op2Info, Operands, I);
1302 }
1303 Type *CondTy = Operands[0]->getType();
1304 return TargetTTI->getCmpSelInstrCost(Opcode, U->getType(), CondTy,
1306 CostKind, I);
1307 }
1308 case Instruction::ICmp:
1309 case Instruction::FCmp: {
1310 Type *ValTy = Operands[0]->getType();
1311 // TODO: Also handle ICmp/FCmp constant expressions.
1312 return TargetTTI->getCmpSelInstrCost(Opcode, ValTy, U->getType(),
1313 I ? cast<CmpInst>(I)->getPredicate()
1315 CostKind, I);
1316 }
1317 case Instruction::InsertElement: {
1318 auto *IE = dyn_cast<InsertElementInst>(U);
1319 if (!IE)
1320 return TTI::TCC_Basic; // FIXME
1321 unsigned Idx = -1;
1322 if (auto *CI = dyn_cast<ConstantInt>(Operands[2]))
1323 if (CI->getValue().getActiveBits() <= 32)
1324 Idx = CI->getZExtValue();
1325 return TargetTTI->getVectorInstrCost(*IE, Ty, CostKind, Idx);
1326 }
1327 case Instruction::ShuffleVector: {
1328 auto *Shuffle = dyn_cast<ShuffleVectorInst>(U);
1329 if (!Shuffle)
1330 return TTI::TCC_Basic; // FIXME
1331
1332 auto *VecTy = cast<VectorType>(U->getType());
1333 auto *VecSrcTy = cast<VectorType>(Operands[0]->getType());
1334 ArrayRef<int> Mask = Shuffle->getShuffleMask();
1335 int NumSubElts, SubIndex;
1336
1337 // TODO: move more of this inside improveShuffleKindFromMask.
1338 if (Shuffle->changesLength()) {
1339 // Treat a 'subvector widening' as a free shuffle.
1340 if (Shuffle->increasesLength() && Shuffle->isIdentityWithPadding())
1341 return 0;
1342
1343 if (Shuffle->isExtractSubvectorMask(SubIndex))
1344 return TargetTTI->getShuffleCost(TTI::SK_ExtractSubvector, VecSrcTy,
1345 Mask, CostKind, SubIndex, VecTy,
1346 Operands, Shuffle);
1347
1348 if (Shuffle->isInsertSubvectorMask(NumSubElts, SubIndex))
1349 return TargetTTI->getShuffleCost(
1350 TTI::SK_InsertSubvector, VecTy, Mask, CostKind, SubIndex,
1351 FixedVectorType::get(VecTy->getScalarType(), NumSubElts),
1352 Operands, Shuffle);
1353
1354 int ReplicationFactor, VF;
1355 if (Shuffle->isReplicationMask(ReplicationFactor, VF)) {
1356 APInt DemandedDstElts = APInt::getZero(Mask.size());
1357 for (auto I : enumerate(Mask)) {
1358 if (I.value() != PoisonMaskElem)
1359 DemandedDstElts.setBit(I.index());
1360 }
1361 return TargetTTI->getReplicationShuffleCost(
1362 VecSrcTy->getElementType(), ReplicationFactor, VF,
1363 DemandedDstElts, CostKind);
1364 }
1365
1366 bool IsUnary = isa<UndefValue>(Operands[1]);
1367 NumSubElts = VecSrcTy->getElementCount().getKnownMinValue();
1368 SmallVector<int, 16> AdjustMask(Mask.begin(), Mask.end());
1369
1370 // Widening shuffle - widening the source(s) to the new length
1371 // (treated as free - see above), and then perform the adjusted
1372 // shuffle at that width.
1373 if (Shuffle->increasesLength()) {
1374 for (int &M : AdjustMask)
1375 M = M >= NumSubElts ? (M + (Mask.size() - NumSubElts)) : M;
1376
1377 return TargetTTI->getShuffleCost(
1379 AdjustMask, CostKind, 0, nullptr, Operands, Shuffle);
1380 }
1381
1382 // Narrowing shuffle - perform shuffle at original wider width and
1383 // then extract the lower elements.
1384 AdjustMask.append(NumSubElts - Mask.size(), PoisonMaskElem);
1385
1386 InstructionCost ShuffleCost = TargetTTI->getShuffleCost(
1388 VecSrcTy, AdjustMask, CostKind, 0, nullptr, Operands, Shuffle);
1389
1390 SmallVector<int, 16> ExtractMask(Mask.size());
1391 std::iota(ExtractMask.begin(), ExtractMask.end(), 0);
1392 return ShuffleCost + TargetTTI->getShuffleCost(
1393 TTI::SK_ExtractSubvector, VecSrcTy,
1394 ExtractMask, CostKind, 0, VecTy, {}, Shuffle);
1395 }
1396
1397 if (Shuffle->isIdentity())
1398 return 0;
1399
1400 if (Shuffle->isReverse())
1401 return TargetTTI->getShuffleCost(TTI::SK_Reverse, VecTy, Mask, CostKind,
1402 0, nullptr, Operands, Shuffle);
1403
1404 if (Shuffle->isSelect())
1405 return TargetTTI->getShuffleCost(TTI::SK_Select, VecTy, Mask, CostKind,
1406 0, nullptr, Operands, Shuffle);
1407
1408 if (Shuffle->isTranspose())
1409 return TargetTTI->getShuffleCost(TTI::SK_Transpose, VecTy, Mask,
1410 CostKind, 0, nullptr, Operands,
1411 Shuffle);
1412
1413 if (Shuffle->isZeroEltSplat())
1414 return TargetTTI->getShuffleCost(TTI::SK_Broadcast, VecTy, Mask,
1415 CostKind, 0, nullptr, Operands,
1416 Shuffle);
1417
1418 if (Shuffle->isSingleSource())
1419 return TargetTTI->getShuffleCost(TTI::SK_PermuteSingleSrc, VecTy, Mask,
1420 CostKind, 0, nullptr, Operands,
1421 Shuffle);
1422
1423 if (Shuffle->isInsertSubvectorMask(NumSubElts, SubIndex))
1424 return TargetTTI->getShuffleCost(
1425 TTI::SK_InsertSubvector, VecTy, Mask, CostKind, SubIndex,
1426 FixedVectorType::get(VecTy->getScalarType(), NumSubElts), Operands,
1427 Shuffle);
1428
1429 if (Shuffle->isSplice(SubIndex))
1430 return TargetTTI->getShuffleCost(TTI::SK_Splice, VecTy, Mask, CostKind,
1431 SubIndex, nullptr, Operands, Shuffle);
1432
1433 return TargetTTI->getShuffleCost(TTI::SK_PermuteTwoSrc, VecTy, Mask,
1434 CostKind, 0, nullptr, Operands, Shuffle);
1435 }
1436 case Instruction::ExtractElement: {
1437 auto *EEI = dyn_cast<ExtractElementInst>(U);
1438 if (!EEI)
1439 return TTI::TCC_Basic; // FIXME
1440 unsigned Idx = -1;
1441 if (auto *CI = dyn_cast<ConstantInt>(Operands[1]))
1442 if (CI->getValue().getActiveBits() <= 32)
1443 Idx = CI->getZExtValue();
1444 Type *DstTy = Operands[0]->getType();
1445 return TargetTTI->getVectorInstrCost(*EEI, DstTy, CostKind, Idx);
1446 }
1447 }
1448
1449 // By default, just classify everything as 'basic' or -1 to represent that
1450 // don't know the throughput cost.
1452 }
1453
1455 auto *TargetTTI = static_cast<T *>(this);
1456 SmallVector<const Value *, 4> Ops(I->operand_values());
1457 InstructionCost Cost = TargetTTI->getInstructionCost(
1460 }
1461
1462 bool supportsTailCallFor(const CallBase *CB) const {
1463 return static_cast<const T *>(this)->supportsTailCalls();
1464 }
1465};
1466} // namespace llvm
1467
1468#endif
Analysis containing CSE Info
Definition: CSEInfo.cpp:27
static cl::opt< TargetTransformInfo::TargetCostKind > CostKind("cost-kind", cl::desc("Target cost kind"), cl::init(TargetTransformInfo::TCK_RecipThroughput), cl::values(clEnumValN(TargetTransformInfo::TCK_RecipThroughput, "throughput", "Reciprocal throughput"), clEnumValN(TargetTransformInfo::TCK_Latency, "latency", "Instruction latency"), clEnumValN(TargetTransformInfo::TCK_CodeSize, "code-size", "Code size"), clEnumValN(TargetTransformInfo::TCK_SizeAndLatency, "size-latency", "Code size and latency")))
return RetTy
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
std::string Name
static bool isSigned(unsigned int Opcode)
Hexagon Common GEP
#define F(x, y, z)
Definition: MD5.cpp:55
#define I(x, y, z)
Definition: MD5.cpp:58
mir Rename Register Operands
LLVMContext & Context
static cl::opt< RegAllocEvictionAdvisorAnalysis::AdvisorMode > Mode("regalloc-enable-advisor", cl::Hidden, cl::init(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Default), cl::desc("Enable regalloc advisor mode"), cl::values(clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Default, "default", "Default"), clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Release, "release", "precompiled"), clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Development, "development", "for training")))
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
static SymbolRef::Type getType(const Symbol *Sym)
Definition: TapiFile.cpp:40
This pass exposes codegen information to IR-level passes.
Class for arbitrary precision integers.
Definition: APInt.h:76
void setBit(unsigned BitPosition)
Set the given bit to 1 whose position is given as "bitPosition".
Definition: APInt.h:1308
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition: APInt.h:1439
APInt sextOrTrunc(unsigned width) const
Sign extend or truncate to width.
Definition: APInt.cpp:1010
static APInt getZero(unsigned numBits)
Get the '0' value for the specified bit-width.
Definition: APInt.h:178
int64_t getSExtValue() const
Get sign extended value.
Definition: APInt.h:1513
an instruction to allocate memory on the stack
Definition: Instructions.h:59
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition: ArrayRef.h:41
A cache of @llvm.assume calls within a function.
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
Conditional or Unconditional Branch instruction.
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
Definition: InstrTypes.h:1494
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition: InstrTypes.h:993
This is the shared class of boolean and integer constants.
Definition: Constants.h:80
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
Definition: Constants.h:154
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition: Constants.h:145
This is an important base class in LLVM.
Definition: Constant.h:41
A parsed version of the target data layout string in and methods for querying it.
Definition: DataLayout.h:110
bool isLegalInteger(uint64_t Width) const
Returns true if the specified type is known to be a native integer type supported by the CPU.
Definition: DataLayout.h:260
const StructLayout * getStructLayout(StructType *Ty) const
Returns a StructLayout object, indicating the alignment of the struct, its size, and the offsets of i...
Definition: DataLayout.cpp:720
unsigned getPointerTypeSizeInBits(Type *) const
Layout pointer size, in bits, based on the type.
Definition: DataLayout.cpp:763
TypeSize getTypeSizeInBits(Type *Ty) const
Size examples:
Definition: DataLayout.h:672
TypeSize getTypeStoreSize(Type *Ty) const
Returns the maximum number of bytes that may be overwritten by storing the specified type.
Definition: DataLayout.h:472
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
Definition: Dominators.h:162
static constexpr ElementCount get(ScalarTy MinVal, bool Scalable)
Definition: TypeSize.h:302
Convenience struct for specifying and reasoning about fast-math flags.
Definition: FMF.h:20
static FixedVectorType * get(Type *ElementType, unsigned NumElts)
Definition: Type.cpp:692
The core instruction combiner logic.
Definition: InstCombiner.h:47
static InstructionCost getInvalid(CostType Val=0)
A wrapper class for inspecting calls to intrinsic functions.
Definition: IntrinsicInst.h:47
Intrinsic::ID getIntrinsicID() const
Return the intrinsic ID of this intrinsic.
Definition: IntrinsicInst.h:54
This is an important class for using LLVM in a threaded context.
Definition: LLVMContext.h:67
An instruction for reading from memory.
Definition: Instructions.h:184
Represents a single loop in the control flow graph.
Definition: LoopInfo.h:44
unsigned getOpcode() const
Return the opcode for this Instruction or ConstantExpr.
Definition: Operator.h:41
The optimization diagnostic interface.
Analysis providing profile information.
The RecurrenceDescriptor is used to identify recurrences variables in a loop.
Definition: IVDescriptors.h:71
This node represents a polynomial recurrence on the trip count of the specified loop.
const SCEV * getStepRecurrence(ScalarEvolution &SE) const
Constructs and returns the recurrence indicating how much this expression steps by.
This class represents a constant integer value.
const APInt & getAPInt() const
This class represents an analyzed expression in the program.
The main scalar evolution driver.
This is a 'bitvector' (really, a variable-sized bit array), optimized for the case when the array is ...
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
Definition: SmallVector.h:586
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
Definition: SmallVector.h:696
void push_back(const T &Elt)
Definition: SmallVector.h:426
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1209
An instruction for storing to memory.
Definition: Instructions.h:317
StringRef - Represent a constant reference to a string, i.e.
Definition: StringRef.h:50
TypeSize getElementOffset(unsigned Idx) const
Definition: DataLayout.h:651
Class to represent struct types.
Definition: DerivedTypes.h:216
Multiway switch.
Provides information about what library functions are available for the current target.
Base class for use as a mix-in that aids implementing a TargetTransformInfo-compatible class.
const DataLayout & getDataLayout() const
void getMemcpyLoopResidualLoweringType(SmallVectorImpl< Type * > &OpsOut, LLVMContext &Context, unsigned RemainingBytes, unsigned SrcAddrSpace, unsigned DestAddrSpace, unsigned SrcAlign, unsigned DestAlign, std::optional< uint32_t > AtomicCpySize) const
bool isLegalToVectorizeStore(StoreInst *SI) const
bool addrspacesMayAlias(unsigned AS0, unsigned AS1) const
bool shouldTreatInstructionLikeSelect(const Instruction *I)
bool isLegalToVectorizeLoad(LoadInst *LI) const
bool isLegalBroadcastLoad(Type *ElementTy, ElementCount NumElements) const
std::optional< unsigned > getVScaleForTuning() const
bool shouldMaximizeVectorBandwidth(TargetTransformInfo::RegisterKind K) const
bool canSaveCmp(Loop *L, BranchInst **BI, ScalarEvolution *SE, LoopInfo *LI, DominatorTree *DT, AssumptionCache *AC, TargetLibraryInfo *LibInfo) const
bool isLegalStridedLoadStore(Type *DataType, Align Alignment) const
std::optional< Value * > simplifyDemandedVectorEltsIntrinsic(InstCombiner &IC, IntrinsicInst &II, APInt DemandedElts, APInt &UndefElts, APInt &UndefElts2, APInt &UndefElts3, std::function< void(Instruction *, unsigned, APInt, APInt &)> SimplifyAndSetOp) const
bool isLegalICmpImmediate(int64_t Imm) const
unsigned getRegUsageForType(Type *Ty) const
unsigned getInterleavedMemoryOpCost(unsigned Opcode, Type *VecTy, unsigned Factor, ArrayRef< unsigned > Indices, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind, bool UseMaskForCond, bool UseMaskForGaps) const
bool areTypesABICompatible(const Function *Caller, const Function *Callee, const ArrayRef< Type * > &Types) const
void getPeelingPreferences(Loop *, ScalarEvolution &, TTI::PeelingPreferences &) const
bool isAlwaysUniform(const Value *V) const
bool isProfitableToHoist(Instruction *I) const
unsigned getEstimatedNumberOfCaseClusters(const SwitchInst &SI, unsigned &JTSize, ProfileSummaryInfo *PSI, BlockFrequencyInfo *BFI) const
InstructionCost getIntImmCostInst(unsigned Opcode, unsigned Idx, const APInt &Imm, Type *Ty, TTI::TargetCostKind CostKind, Instruction *Inst=nullptr) const
bool isLSRCostLess(const TTI::LSRCost &C1, const TTI::LSRCost &C2) const
bool isExpensiveToSpeculativelyExecute(const Instruction *I)
bool isTruncateFree(Type *Ty1, Type *Ty2) const
bool isStridedAccess(const SCEV *Ptr) const
InstructionCost getCallInstrCost(Function *F, Type *RetTy, ArrayRef< Type * > Tys, TTI::TargetCostKind CostKind) const
InstructionCost getArithmeticReductionCost(unsigned, VectorType *, std::optional< FastMathFlags > FMF, TTI::TargetCostKind) const
InstructionCost getFPOpCost(Type *Ty) const
bool shouldConsiderAddressTypePromotion(const Instruction &I, bool &AllowPromotionWithoutCommonHeader) const
bool areInlineCompatible(const Function *Caller, const Function *Callee) const
std::pair< const Value *, unsigned > getPredicatedAddrSpace(const Value *V) const
InstructionCost getMemcpyCost(const Instruction *I) const
unsigned getInliningCostBenefitAnalysisProfitableMultiplier() const
unsigned getStoreVectorFactor(unsigned VF, unsigned StoreSize, unsigned ChainSizeInBytes, VectorType *VecTy) const
std::optional< unsigned > getMaxVScale() const
TTI::PopcntSupportKind getPopcntSupport(unsigned IntTyWidthInBit) const
unsigned getCallerAllocaCost(const CallBase *CB, const AllocaInst *AI) const
bool isProfitableLSRChainElement(Instruction *I) const
InstructionCost getShuffleCost(TTI::ShuffleKind Kind, VectorType *Ty, ArrayRef< int > Mask, TTI::TargetCostKind CostKind, int Index, VectorType *SubTp, ArrayRef< const Value * > Args=std::nullopt, const Instruction *CxtI=nullptr) const
InstructionCost getIntImmCodeSizeCost(unsigned Opcode, unsigned Idx, const APInt &Imm, Type *Ty) const
bool preferToKeepConstantsAttached(const Instruction &Inst, const Function &Fn) const
Value * getOrCreateResultFromMemIntrinsic(IntrinsicInst *Inst, Type *ExpectedType) const
InstructionCost getIntrinsicInstrCost(const IntrinsicCostAttributes &ICA, TTI::TargetCostKind CostKind) const
bool isLegalMaskedStore(Type *DataType, Align Alignment) const
InstructionCost getGatherScatterOpCost(unsigned Opcode, Type *DataTy, const Value *Ptr, bool VariableMask, Align Alignment, TTI::TargetCostKind CostKind, const Instruction *I=nullptr) const
unsigned getInlineCallPenalty(const Function *F, const CallBase &Call, unsigned DefaultCallPenalty) const
InstructionCost getVPMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind, const Instruction *I) const
InstructionCost getStridedMemoryOpCost(unsigned Opcode, Type *DataTy, const Value *Ptr, bool VariableMask, Align Alignment, TTI::TargetCostKind CostKind, const Instruction *I=nullptr) const
bool isNoopAddrSpaceCast(unsigned, unsigned) const
unsigned getStoreMinimumVF(unsigned VF, Type *, Type *) const
InstructionCost getVectorInstrCost(const Instruction &I, Type *Val, TTI::TargetCostKind CostKind, unsigned Index) const
bool preferPredicatedReductionSelect(unsigned Opcode, Type *Ty, TTI::ReductionFlags Flags) const
TargetTransformInfoImplBase(TargetTransformInfoImplBase &&Arg)
TargetTransformInfo::VPLegalization getVPLegalizationStrategy(const VPIntrinsic &PI) const
void getUnrollingPreferences(Loop *, ScalarEvolution &, TTI::UnrollingPreferences &, OptimizationRemarkEmitter *) const
bool isLegalToVectorizeLoadChain(unsigned ChainSizeInBytes, Align Alignment, unsigned AddrSpace) const
bool isLegalAddressingMode(Type *Ty, GlobalValue *BaseGV, int64_t BaseOffset, bool HasBaseReg, int64_t Scale, unsigned AddrSpace, Instruction *I=nullptr, int64_t ScalableOffset=0) const
bool isHardwareLoopProfitable(Loop *L, ScalarEvolution &SE, AssumptionCache &AC, TargetLibraryInfo *LibInfo, HardwareLoopInfo &HWLoopInfo) const
unsigned minRequiredElementSize(const Value *Val, bool &isSigned) const
Type * getMemcpyLoopLoweringType(LLVMContext &Context, Value *Length, unsigned SrcAddrSpace, unsigned DestAddrSpace, unsigned SrcAlign, unsigned DestAlign, std::optional< uint32_t > AtomicElementSize) const
std::optional< unsigned > getCacheSize(TargetTransformInfo::CacheLevel Level) const
unsigned getAssumedAddrSpace(const Value *V) const
bool isLegalMaskedExpandLoad(Type *DataType, Align Alignment) const
bool isLegalNTStore(Type *DataType, Align Alignment) const
unsigned getRegisterClassForType(bool Vector, Type *Ty=nullptr) const
bool isLegalMaskedGather(Type *DataType, Align Alignment) const
unsigned adjustInliningThreshold(const CallBase *CB) const
BranchProbability getPredictableBranchThreshold() const
std::optional< unsigned > getMinPageSize() const
bool collectFlatAddressOperands(SmallVectorImpl< int > &OpIndexes, Intrinsic::ID IID) const
InstructionCost getExtendedReductionCost(unsigned Opcode, bool IsUnsigned, Type *ResTy, VectorType *Ty, FastMathFlags FMF, TTI::TargetCostKind CostKind) const
bool allowsMisalignedMemoryAccesses(LLVMContext &Context, unsigned BitWidth, unsigned AddressSpace, Align Alignment, unsigned *Fast) const
const SCEVConstant * getConstantStrideStep(ScalarEvolution *SE, const SCEV *Ptr) const
unsigned getMinPrefetchStride(unsigned NumMemAccesses, unsigned NumStridedMemAccesses, unsigned NumPrefetches, bool HasCall) const
InstructionCost getIntImmCostIntrin(Intrinsic::ID IID, unsigned Idx, const APInt &Imm, Type *Ty, TTI::TargetCostKind CostKind) const
unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, unsigned ChainSizeInBytes, VectorType *VecTy) const
bool isIndexedLoadLegal(TTI::MemIndexedMode Mode, Type *Ty, const DataLayout &DL) const
bool shouldPrefetchAddressSpace(unsigned AS) const
InstructionCost getIntImmCost(const APInt &Imm, Type *Ty, TTI::TargetCostKind CostKind) const
unsigned getReplicationShuffleCost(Type *EltTy, int ReplicationFactor, int VF, const APInt &DemandedDstElts, TTI::TargetCostKind CostKind)
bool isSourceOfDivergence(const Value *V) const
bool enableAggressiveInterleaving(bool LoopHasReductions) const
unsigned getMaxInterleaveFactor(ElementCount VF) const
InstructionCost getCastInstrCost(unsigned Opcode, Type *Dst, Type *Src, TTI::CastContextHint CCH, TTI::TargetCostKind CostKind, const Instruction *I) const
std::optional< unsigned > getCacheAssociativity(TargetTransformInfo::CacheLevel Level) const
bool hasVolatileVariant(Instruction *I, unsigned AddrSpace) const
InstructionCost getArithmeticInstrCost(unsigned Opcode, Type *Ty, TTI::TargetCostKind CostKind, TTI::OperandValueInfo Opd1Info, TTI::OperandValueInfo Opd2Info, ArrayRef< const Value * > Args, const Instruction *CxtI=nullptr) const
TTI::AddressingModeKind getPreferredAddressingMode(const Loop *L, ScalarEvolution *SE) const
bool forceScalarizeMaskedGather(VectorType *DataType, Align Alignment) const
bool getTgtMemIntrinsic(IntrinsicInst *Inst, MemIntrinsicInfo &Info) const
unsigned getInliningCostBenefitAnalysisSavingsMultiplier() const
bool canHaveNonUndefGlobalInitializerInAddressSpace(unsigned AS) const
InstructionCost getMulAccReductionCost(bool IsUnsigned, Type *ResTy, VectorType *Ty, TTI::TargetCostKind CostKind) const
bool isIndexedStoreLegal(TTI::MemIndexedMode Mode, Type *Ty, const DataLayout &DL) const
bool hasDivRemOp(Type *DataType, bool IsSigned) const
InstructionCost getAddressComputationCost(Type *Tp, ScalarEvolution *, const SCEV *) const
InstructionCost getCmpSelInstrCost(unsigned Opcode, Type *ValTy, Type *CondTy, CmpInst::Predicate VecPred, TTI::TargetCostKind CostKind, const Instruction *I) const
TTI::MemCmpExpansionOptions enableMemCmpExpansion(bool OptSize, bool IsZeroCmp) const
bool preferInLoopReduction(unsigned Opcode, Type *Ty, TTI::ReductionFlags Flags) const
InstructionCost getOperandsScalarizationOverhead(ArrayRef< const Value * > Args, ArrayRef< Type * > Tys, TTI::TargetCostKind CostKind) const
TypeSize getRegisterBitWidth(TargetTransformInfo::RegisterKind K) const
bool isConstantStridedAccessLessThan(ScalarEvolution *SE, const SCEV *Ptr, int64_t MergeDistance) const
InstructionCost getMaskedMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind) const
bool isLoweredToCall(const Function *F) const
bool hasBranchDivergence(const Function *F=nullptr) const
TargetTransformInfoImplBase(const DataLayout &DL)
const char * getRegisterClassName(unsigned ClassID) const
bool isElementTypeLegalForScalableVector(Type *Ty) const
bool preferPredicateOverEpilogue(TailFoldingInfo *TFI) const
bool isLegalToVectorizeReduction(const RecurrenceDescriptor &RdxDesc, ElementCount VF) const
InstructionCost getScalarizationOverhead(VectorType *Ty, const APInt &DemandedElts, bool Insert, bool Extract, TTI::TargetCostKind CostKind) const
unsigned getMaximumVF(unsigned ElemWidth, unsigned Opcode) const
TailFoldingStyle getPreferredTailFoldingStyle(bool IVUpdateMayOverflow=true) const
InstructionCost getCFInstrCost(unsigned Opcode, TTI::TargetCostKind CostKind, const Instruction *I=nullptr) const
InstructionCost getCostOfKeepingLiveOverCall(ArrayRef< Type * > Tys) const
InstructionCost getMinMaxReductionCost(Intrinsic::ID IID, VectorType *, FastMathFlags, TTI::TargetCostKind) const
InstructionCost getVectorInstrCost(unsigned Opcode, Type *Val, TTI::TargetCostKind CostKind, unsigned Index, Value *Op0, Value *Op1) const
unsigned getLoadStoreVecRegBitWidth(unsigned AddrSpace) const
InstructionCost getMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind, TTI::OperandValueInfo OpInfo, const Instruction *I) const
bool useColdCCForColdCall(Function &F) const
bool shouldExpandReduction(const IntrinsicInst *II) const
bool isLegalMaskedScatter(Type *DataType, Align Alignment) const
unsigned getNumberOfRegisters(unsigned ClassID) const
InstructionCost getExtractWithExtendCost(unsigned Opcode, Type *Dst, VectorType *VecTy, unsigned Index) const
InstructionCost getGEPCost(Type *PointeeType, const Value *Ptr, ArrayRef< const Value * > Operands, Type *AccessType, TTI::TargetCostKind CostKind) const
bool isLegalNTLoad(Type *DataType, Align Alignment) const
std::optional< Value * > simplifyDemandedUseBitsIntrinsic(InstCombiner &IC, IntrinsicInst &II, APInt DemandedMask, KnownBits &Known, bool &KnownBitsComputed) const
bool forceScalarizeMaskedScatter(VectorType *DataType, Align Alignment) const
bool hasActiveVectorLength(unsigned Opcode, Type *DataType, Align Alignment) const
bool isLegalAltInstr(VectorType *VecTy, unsigned Opcode0, unsigned Opcode1, const SmallBitVector &OpcodeMask) const
bool isLegalMaskedLoad(Type *DataType, Align Alignment) const
bool isValidAddrSpaceCast(unsigned FromAS, unsigned ToAS) const
std::optional< Instruction * > instCombineIntrinsic(InstCombiner &IC, IntrinsicInst &II) const
InstructionCost getScalingFactorCost(Type *Ty, GlobalValue *BaseGV, int64_t BaseOffset, bool HasBaseReg, int64_t Scale, unsigned AddrSpace) const
bool isLegalAddScalableImmediate(int64_t Imm) const
bool isLegalMaskedCompressStore(Type *DataType, Align Alignment) const
InstructionCost getAltInstrCost(VectorType *VecTy, unsigned Opcode0, unsigned Opcode1, const SmallBitVector &OpcodeMask, TTI::TargetCostKind CostKind) const
TargetTransformInfoImplBase(const TargetTransformInfoImplBase &Arg)=default
bool isLegalToVectorizeStoreChain(unsigned ChainSizeInBytes, Align Alignment, unsigned AddrSpace) const
ElementCount getMinimumVF(unsigned ElemWidth, bool IsScalable) const
bool shouldBuildLookupTablesForConstant(Constant *C) const
Value * rewriteIntrinsicWithAddressSpace(IntrinsicInst *II, Value *OldV, Value *NewV) const
CRTP base class for use as a mix-in that aids implementing a TargetTransformInfo-compatible class.
bool supportsTailCallFor(const CallBase *CB) const
InstructionCost getGEPCost(Type *PointeeType, const Value *Ptr, ArrayRef< const Value * > Operands, Type *AccessType, TTI::TargetCostKind CostKind)
InstructionCost getPointersChainCost(ArrayRef< const Value * > Ptrs, const Value *Base, const TTI::PointersChainInfo &Info, Type *AccessTy, TTI::TargetCostKind CostKind)
InstructionCost getInstructionCost(const User *U, ArrayRef< const Value * > Operands, TTI::TargetCostKind CostKind)
bool isExpensiveToSpeculativelyExecute(const Instruction *I)
This pass provides access to the codegen interfaces that are needed for IR-level transformations.
static CastContextHint getCastContextHint(const Instruction *I)
Calculates a CastContextHint from I.
static OperandValueInfo getOperandInfo(const Value *V)
Collect properties of V used in cost analysis, e.g. OP_PowerOf2.
TargetCostKind
The kind of cost model.
@ TCK_RecipThroughput
Reciprocal throughput.
@ TCK_CodeSize
Instruction code size.
@ TCK_SizeAndLatency
The weighted sum of size and latency.
@ TCK_Latency
The latency of instruction.
PopcntSupportKind
Flags indicating the kind of support for population count.
@ TCC_Expensive
The cost of a 'div' instruction on x86.
@ TCC_Free
Expected to fold away in lowering.
@ TCC_Basic
The cost of a typical 'add' instruction.
MemIndexedMode
The type of load/store indexing.
ShuffleKind
The various kinds of shuffle patterns for vector queries.
@ SK_InsertSubvector
InsertSubvector. Index indicates start offset.
@ SK_Select
Selects elements from the corresponding lane of either source operand.
@ SK_PermuteSingleSrc
Shuffle elements of single source vector with any shuffle mask.
@ SK_Transpose
Transpose two vectors.
@ SK_Splice
Concatenates elements from the first input vector with elements of the second input vector.
@ SK_Broadcast
Broadcast element 0 to all other elements.
@ SK_PermuteTwoSrc
Merge elements from two source vectors into one with any shuffle mask.
@ SK_Reverse
Reverse the order of the vector.
@ SK_ExtractSubvector
ExtractSubvector Index indicates start offset.
CastContextHint
Represents a hint about the context in which a cast is used.
CacheLevel
The possible cache levels.
This class represents a truncation of integer types.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition: TypeSize.h:330
The instances of the Type class are immutable: once they are created, they are never changed.
Definition: Type.h:45
bool isVectorTy() const
True if this is an instance of VectorType.
Definition: Type.h:265
unsigned getPointerAddressSpace() const
Get the address space of this pointer or pointer vector type.
static IntegerType * getIntNTy(LLVMContext &C, unsigned N)
unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
static IntegerType * getInt8Ty(LLVMContext &C)
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
Definition: Type.h:185
bool isScalableTy() const
Return true if this is a type whose size is a known multiple of vscale.
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition: Type.h:348
This is the common base class for vector predication intrinsics.
LLVM Value Representation.
Definition: Value.h:74
Type * getType() const
All values are typed, get the type of this value.
Definition: Value.h:255
Base class of all SIMD vector types.
Definition: DerivedTypes.h:403
constexpr ScalarTy getFixedValue() const
Definition: TypeSize.h:187
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition: TypeSize.h:171
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ Fast
Attempts to make calls as fast as possible (e.g.
Definition: CallingConv.h:41
@ C
The default llvm calling convention, compatible with C.
Definition: CallingConv.h:34
bool match(Val *V, const Pattern &P)
Definition: PatternMatch.h:49
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
class_match< Value > m_Value()
Match an arbitrary value and ignore it.
Definition: PatternMatch.h:92
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
match_combine_or< LTy, RTy > m_CombineOr(const LTy &L, const RTy &R)
Combine two pattern matchers matching L || R.
Definition: PatternMatch.h:239
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
@ Length
Definition: DWP.cpp:456
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are are tuples (A,...
Definition: STLExtras.h:2406
AddressSpace
Definition: NVPTXBaseInfo.h:21
Value * getSplatValue(const Value *V)
Get splat value if the input is a splat vector or return nullptr.
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1729
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition: MathExtras.h:275
constexpr int PoisonMaskElem
constexpr unsigned BitWidth
Definition: BitmaskEnum.h:191
gep_type_iterator gep_type_begin(const User *GEP)
InstructionCost Cost
@ DataWithoutLaneMask
Same as Data, but avoids using the get.active.lane.mask intrinsic to calculate the mask and instead i...
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition: Alignment.h:39
Attributes of a target dependent hardware loop.
Information about a load/store intrinsic defined by the target.
Returns options for expansion of memcmp. IsZeroCmp is.
Describe known properties for a set of pointers.
Flags describing the kind of vector reduction.
Parameters that control the generic loop unrolling transformation.