LLVM 19.0.0git
TargetTransformInfoImpl.h
Go to the documentation of this file.
1//===- TargetTransformInfoImpl.h --------------------------------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8/// \file
9/// This file provides helpers for the implementation of
10/// a TargetTransformInfo-conforming class.
11///
12//===----------------------------------------------------------------------===//
13
14#ifndef LLVM_ANALYSIS_TARGETTRANSFORMINFOIMPL_H
15#define LLVM_ANALYSIS_TARGETTRANSFORMINFOIMPL_H
16
20#include "llvm/IR/DataLayout.h"
23#include "llvm/IR/Operator.h"
25#include <optional>
26#include <utility>
27
28namespace llvm {
29
30class Function;
31
32/// Base class for use as a mix-in that aids implementing
33/// a TargetTransformInfo-compatible class.
35
36protected:
38
39 const DataLayout &DL;
40
42
43public:
44 // Provide value semantics. MSVC requires that we spell all of these out.
47
48 const DataLayout &getDataLayout() const { return DL; }
49
50 InstructionCost getGEPCost(Type *PointeeType, const Value *Ptr,
53 // In the basic model, we just assume that all-constant GEPs will be folded
54 // into their uses via addressing modes.
55 for (const Value *Operand : Operands)
56 if (!isa<Constant>(Operand))
57 return TTI::TCC_Basic;
58
59 return TTI::TCC_Free;
60 }
61
63 unsigned &JTSize,
65 BlockFrequencyInfo *BFI) const {
66 (void)PSI;
67 (void)BFI;
68 JTSize = 0;
69 return SI.getNumCases();
70 }
71
72 unsigned getInliningThresholdMultiplier() const { return 1; }
75 return 8;
76 }
77 unsigned adjustInliningThreshold(const CallBase *CB) const { return 0; }
78 unsigned getCallerAllocaCost(const CallBase *CB, const AllocaInst *AI) const {
79 return 0;
80 };
81
82 int getInlinerVectorBonusPercent() const { return 150; }
83
85 return TTI::TCC_Expensive;
86 }
87
89 return 64;
90 }
91
92 // Although this default value is arbitrary, it is not random. It is assumed
93 // that a condition that evaluates the same way by a higher percentage than
94 // this is best represented as control flow. Therefore, the default value N
95 // should be set such that the win from N% correct executions is greater than
96 // the loss from (100 - N)% mispredicted executions for the majority of
97 // intended targets.
99 return BranchProbability(99, 100);
100 }
101
102 bool hasBranchDivergence(const Function *F = nullptr) const { return false; }
103
104 bool isSourceOfDivergence(const Value *V) const { return false; }
105
106 bool isAlwaysUniform(const Value *V) const { return false; }
107
108 bool isValidAddrSpaceCast(unsigned FromAS, unsigned ToAS) const {
109 return false;
110 }
111
112 bool addrspacesMayAlias(unsigned AS0, unsigned AS1) const {
113 return true;
114 }
115
116 unsigned getFlatAddressSpace() const { return -1; }
117
119 Intrinsic::ID IID) const {
120 return false;
121 }
122
123 bool isNoopAddrSpaceCast(unsigned, unsigned) const { return false; }
125 return AS == 0;
126 };
127
128 unsigned getAssumedAddrSpace(const Value *V) const { return -1; }
129
130 bool isSingleThreaded() const { return false; }
131
132 std::pair<const Value *, unsigned>
134 return std::make_pair(nullptr, -1);
135 }
136
138 Value *NewV) const {
139 return nullptr;
140 }
141
142 bool isLoweredToCall(const Function *F) const {
143 assert(F && "A concrete function must be provided to this routine.");
144
145 // FIXME: These should almost certainly not be handled here, and instead
146 // handled with the help of TLI or the target itself. This was largely
147 // ported from existing analysis heuristics here so that such refactorings
148 // can take place in the future.
149
150 if (F->isIntrinsic())
151 return false;
152
153 if (F->hasLocalLinkage() || !F->hasName())
154 return true;
155
156 StringRef Name = F->getName();
157
158 // These will all likely lower to a single selection DAG node.
159 if (Name == "copysign" || Name == "copysignf" || Name == "copysignl" ||
160 Name == "fabs" || Name == "fabsf" || Name == "fabsl" || Name == "sin" ||
161 Name == "fmin" || Name == "fminf" || Name == "fminl" ||
162 Name == "fmax" || Name == "fmaxf" || Name == "fmaxl" ||
163 Name == "sinf" || Name == "sinl" || Name == "cos" || Name == "cosf" ||
164 Name == "cosl" || Name == "sqrt" || Name == "sqrtf" || Name == "sqrtl")
165 return false;
166
167 // These are all likely to be optimized into something smaller.
168 if (Name == "pow" || Name == "powf" || Name == "powl" || Name == "exp2" ||
169 Name == "exp2l" || Name == "exp2f" || Name == "floor" ||
170 Name == "floorf" || Name == "ceil" || Name == "round" ||
171 Name == "ffs" || Name == "ffsl" || Name == "abs" || Name == "labs" ||
172 Name == "llabs")
173 return false;
174
175 return true;
176 }
177
180 HardwareLoopInfo &HWLoopInfo) const {
181 return false;
182 }
183
184 bool preferPredicateOverEpilogue(TailFoldingInfo *TFI) const { return false; }
185
187 getPreferredTailFoldingStyle(bool IVUpdateMayOverflow = true) const {
189 }
190
191 std::optional<Instruction *> instCombineIntrinsic(InstCombiner &IC,
192 IntrinsicInst &II) const {
193 return std::nullopt;
194 }
195
196 std::optional<Value *>
198 APInt DemandedMask, KnownBits &Known,
199 bool &KnownBitsComputed) const {
200 return std::nullopt;
201 }
202
204 InstCombiner &IC, IntrinsicInst &II, APInt DemandedElts, APInt &UndefElts,
205 APInt &UndefElts2, APInt &UndefElts3,
206 std::function<void(Instruction *, unsigned, APInt, APInt &)>
207 SimplifyAndSetOp) const {
208 return std::nullopt;
209 }
210
213 OptimizationRemarkEmitter *) const {}
214
216 TTI::PeelingPreferences &) const {}
217
218 bool isLegalAddImmediate(int64_t Imm) const { return false; }
219
220 bool isLegalAddScalableImmediate(int64_t Imm) const { return false; }
221
222 bool isLegalICmpImmediate(int64_t Imm) const { return false; }
223
224 bool isLegalAddressingMode(Type *Ty, GlobalValue *BaseGV, int64_t BaseOffset,
225 bool HasBaseReg, int64_t Scale, unsigned AddrSpace,
226 Instruction *I = nullptr,
227 int64_t ScalableOffset = 0) const {
228 // Guess that only reg and reg+reg addressing is allowed. This heuristic is
229 // taken from the implementation of LSR.
230 return !BaseGV && BaseOffset == 0 && (Scale == 0 || Scale == 1);
231 }
232
233 bool isLSRCostLess(const TTI::LSRCost &C1, const TTI::LSRCost &C2) const {
234 return std::tie(C1.NumRegs, C1.AddRecCost, C1.NumIVMuls, C1.NumBaseAdds,
235 C1.ScaleCost, C1.ImmCost, C1.SetupCost) <
236 std::tie(C2.NumRegs, C2.AddRecCost, C2.NumIVMuls, C2.NumBaseAdds,
237 C2.ScaleCost, C2.ImmCost, C2.SetupCost);
238 }
239
240 bool isNumRegsMajorCostOfLSR() const { return true; }
241
242 bool shouldFoldTerminatingConditionAfterLSR() const { return false; }
243
244 bool isProfitableLSRChainElement(Instruction *I) const { return false; }
245
246 bool canMacroFuseCmp() const { return false; }
247
250 TargetLibraryInfo *LibInfo) const {
251 return false;
252 }
253
256 return TTI::AMK_None;
257 }
258
259 bool isLegalMaskedStore(Type *DataType, Align Alignment) const {
260 return false;
261 }
262
263 bool isLegalMaskedLoad(Type *DataType, Align Alignment) const {
264 return false;
265 }
266
267 bool isLegalNTStore(Type *DataType, Align Alignment) const {
268 // By default, assume nontemporal memory stores are available for stores
269 // that are aligned and have a size that is a power of 2.
270 unsigned DataSize = DL.getTypeStoreSize(DataType);
271 return Alignment >= DataSize && isPowerOf2_32(DataSize);
272 }
273
274 bool isLegalNTLoad(Type *DataType, Align Alignment) const {
275 // By default, assume nontemporal memory loads are available for loads that
276 // are aligned and have a size that is a power of 2.
277 unsigned DataSize = DL.getTypeStoreSize(DataType);
278 return Alignment >= DataSize && isPowerOf2_32(DataSize);
279 }
280
281 bool isLegalBroadcastLoad(Type *ElementTy, ElementCount NumElements) const {
282 return false;
283 }
284
285 bool isLegalMaskedScatter(Type *DataType, Align Alignment) const {
286 return false;
287 }
288
289 bool isLegalMaskedGather(Type *DataType, Align Alignment) const {
290 return false;
291 }
292
293 bool forceScalarizeMaskedGather(VectorType *DataType, Align Alignment) const {
294 return false;
295 }
296
298 Align Alignment) const {
299 return false;
300 }
301
302 bool isLegalMaskedCompressStore(Type *DataType, Align Alignment) const {
303 return false;
304 }
305
306 bool isLegalAltInstr(VectorType *VecTy, unsigned Opcode0, unsigned Opcode1,
307 const SmallBitVector &OpcodeMask) const {
308 return false;
309 }
310
311 bool isLegalMaskedExpandLoad(Type *DataType, Align Alignment) const {
312 return false;
313 }
314
315 bool isLegalStridedLoadStore(Type *DataType, Align Alignment) const {
316 return false;
317 }
318
319 bool isLegalMaskedVectorHistogram(Type *AddrType, Type *DataType) const {
320 return false;
321 }
322
323 bool enableOrderedReductions() const { return false; }
324
325 bool hasDivRemOp(Type *DataType, bool IsSigned) const { return false; }
326
327 bool hasVolatileVariant(Instruction *I, unsigned AddrSpace) const {
328 return false;
329 }
330
331 bool prefersVectorizedAddressing() const { return true; }
332
334 StackOffset BaseOffset, bool HasBaseReg,
335 int64_t Scale,
336 unsigned AddrSpace) const {
337 // Guess that all legal addressing mode are free.
338 if (isLegalAddressingMode(Ty, BaseGV, BaseOffset.getFixed(), HasBaseReg,
339 Scale, AddrSpace, /*I=*/nullptr,
340 BaseOffset.getScalable()))
341 return 0;
342 return -1;
343 }
344
345 bool LSRWithInstrQueries() const { return false; }
346
347 bool isTruncateFree(Type *Ty1, Type *Ty2) const { return false; }
348
349 bool isProfitableToHoist(Instruction *I) const { return true; }
350
351 bool useAA() const { return false; }
352
353 bool isTypeLegal(Type *Ty) const { return false; }
354
355 unsigned getRegUsageForType(Type *Ty) const { return 1; }
356
357 bool shouldBuildLookupTables() const { return true; }
358
359 bool shouldBuildLookupTablesForConstant(Constant *C) const { return true; }
360
361 bool shouldBuildRelLookupTables() const { return false; }
362
363 bool useColdCCForColdCall(Function &F) const { return false; }
364
366 const APInt &DemandedElts,
367 bool Insert, bool Extract,
369 return 0;
370 }
371
376 return 0;
377 }
378
379 bool supportsEfficientVectorElementLoadStore() const { return false; }
380
381 bool supportsTailCalls() const { return true; }
382
383 bool enableAggressiveInterleaving(bool LoopHasReductions) const {
384 return false;
385 }
386
388 bool IsZeroCmp) const {
389 return {};
390 }
391
392 bool enableSelectOptimize() const { return true; }
393
395 // If the select is a logical-and/logical-or then it is better treated as a
396 // and/or by the backend.
397 using namespace llvm::PatternMatch;
398 return isa<SelectInst>(I) &&
401 }
402
403 bool enableInterleavedAccessVectorization() const { return false; }
404
405 bool enableMaskedInterleavedAccessVectorization() const { return false; }
406
407 bool isFPVectorizationPotentiallyUnsafe() const { return false; }
408
410 unsigned AddressSpace, Align Alignment,
411 unsigned *Fast) const {
412 return false;
413 }
414
415 TTI::PopcntSupportKind getPopcntSupport(unsigned IntTyWidthInBit) const {
416 return TTI::PSK_Software;
417 }
418
419 bool haveFastSqrt(Type *Ty) const { return false; }
420
421 bool isExpensiveToSpeculativelyExecute(const Instruction *I) { return true; }
422
423 bool isFCmpOrdCheaperThanFCmpZero(Type *Ty) const { return true; }
424
427 }
428
429 InstructionCost getIntImmCodeSizeCost(unsigned Opcode, unsigned Idx,
430 const APInt &Imm, Type *Ty) const {
431 return 0;
432 }
433
436 return TTI::TCC_Basic;
437 }
438
439 InstructionCost getIntImmCostInst(unsigned Opcode, unsigned Idx,
440 const APInt &Imm, Type *Ty,
442 Instruction *Inst = nullptr) const {
443 return TTI::TCC_Free;
444 }
445
447 const APInt &Imm, Type *Ty,
449 return TTI::TCC_Free;
450 }
451
453 const Function &Fn) const {
454 return false;
455 }
456
457 unsigned getNumberOfRegisters(unsigned ClassID) const { return 8; }
458
459 unsigned getRegisterClassForType(bool Vector, Type *Ty = nullptr) const {
460 return Vector ? 1 : 0;
461 };
462
463 const char *getRegisterClassName(unsigned ClassID) const {
464 switch (ClassID) {
465 default:
466 return "Generic::Unknown Register Class";
467 case 0:
468 return "Generic::ScalarRC";
469 case 1:
470 return "Generic::VectorRC";
471 }
472 }
473
475 return TypeSize::getFixed(32);
476 }
477
478 unsigned getMinVectorRegisterBitWidth() const { return 128; }
479
480 std::optional<unsigned> getMaxVScale() const { return std::nullopt; }
481 std::optional<unsigned> getVScaleForTuning() const { return std::nullopt; }
482 bool isVScaleKnownToBeAPowerOfTwo() const { return false; }
483
484 bool
486 return false;
487 }
488
489 ElementCount getMinimumVF(unsigned ElemWidth, bool IsScalable) const {
490 return ElementCount::get(0, IsScalable);
491 }
492
493 unsigned getMaximumVF(unsigned ElemWidth, unsigned Opcode) const { return 0; }
494 unsigned getStoreMinimumVF(unsigned VF, Type *, Type *) const { return VF; }
495
497 const Instruction &I, bool &AllowPromotionWithoutCommonHeader) const {
498 AllowPromotionWithoutCommonHeader = false;
499 return false;
500 }
501
502 unsigned getCacheLineSize() const { return 0; }
503 std::optional<unsigned>
505 switch (Level) {
507 [[fallthrough]];
509 return std::nullopt;
510 }
511 llvm_unreachable("Unknown TargetTransformInfo::CacheLevel");
512 }
513
514 std::optional<unsigned>
516 switch (Level) {
518 [[fallthrough]];
520 return std::nullopt;
521 }
522
523 llvm_unreachable("Unknown TargetTransformInfo::CacheLevel");
524 }
525
526 std::optional<unsigned> getMinPageSize() const { return {}; }
527
528 unsigned getPrefetchDistance() const { return 0; }
529 unsigned getMinPrefetchStride(unsigned NumMemAccesses,
530 unsigned NumStridedMemAccesses,
531 unsigned NumPrefetches, bool HasCall) const {
532 return 1;
533 }
534 unsigned getMaxPrefetchIterationsAhead() const { return UINT_MAX; }
535 bool enableWritePrefetching() const { return false; }
536 bool shouldPrefetchAddressSpace(unsigned AS) const { return !AS; }
537
538 unsigned getMaxInterleaveFactor(ElementCount VF) const { return 1; }
539
541 unsigned Opcode, Type *Ty, TTI::TargetCostKind CostKind,
544 const Instruction *CxtI = nullptr) const {
545 // Widenable conditions will eventually lower into constants, so some
546 // operations with them will be trivially optimized away.
547 auto IsWidenableCondition = [](const Value *V) {
548 if (auto *II = dyn_cast<IntrinsicInst>(V))
549 if (II->getIntrinsicID() == Intrinsic::experimental_widenable_condition)
550 return true;
551 return false;
552 };
553 // FIXME: A number of transformation tests seem to require these values
554 // which seems a little odd for how arbitary there are.
555 switch (Opcode) {
556 default:
557 break;
558 case Instruction::FDiv:
559 case Instruction::FRem:
560 case Instruction::SDiv:
561 case Instruction::SRem:
562 case Instruction::UDiv:
563 case Instruction::URem:
564 // FIXME: Unlikely to be true for CodeSize.
565 return TTI::TCC_Expensive;
566 case Instruction::And:
567 case Instruction::Or:
568 if (any_of(Args, IsWidenableCondition))
569 return TTI::TCC_Free;
570 break;
571 }
572
573 // Assume a 3cy latency for fp arithmetic ops.
575 if (Ty->getScalarType()->isFloatingPointTy())
576 return 3;
577
578 return 1;
579 }
580
582 unsigned Opcode1,
583 const SmallBitVector &OpcodeMask,
586 }
587
589 ArrayRef<int> Mask,
591 VectorType *SubTp,
592 ArrayRef<const Value *> Args = std::nullopt,
593 const Instruction *CxtI = nullptr) const {
594 return 1;
595 }
596
597 InstructionCost getCastInstrCost(unsigned Opcode, Type *Dst, Type *Src,
600 const Instruction *I) const {
601 switch (Opcode) {
602 default:
603 break;
604 case Instruction::IntToPtr: {
605 unsigned SrcSize = Src->getScalarSizeInBits();
606 if (DL.isLegalInteger(SrcSize) &&
607 SrcSize <= DL.getPointerTypeSizeInBits(Dst))
608 return 0;
609 break;
610 }
611 case Instruction::PtrToInt: {
612 unsigned DstSize = Dst->getScalarSizeInBits();
613 if (DL.isLegalInteger(DstSize) &&
614 DstSize >= DL.getPointerTypeSizeInBits(Src))
615 return 0;
616 break;
617 }
618 case Instruction::BitCast:
619 if (Dst == Src || (Dst->isPointerTy() && Src->isPointerTy()))
620 // Identity and pointer-to-pointer casts are free.
621 return 0;
622 break;
623 case Instruction::Trunc: {
624 // trunc to a native type is free (assuming the target has compare and
625 // shift-right of the same width).
626 TypeSize DstSize = DL.getTypeSizeInBits(Dst);
627 if (!DstSize.isScalable() && DL.isLegalInteger(DstSize.getFixedValue()))
628 return 0;
629 break;
630 }
631 }
632 return 1;
633 }
634
636 VectorType *VecTy,
637 unsigned Index) const {
638 return 1;
639 }
640
642 const Instruction *I = nullptr) const {
643 // A phi would be free, unless we're costing the throughput because it
644 // will require a register.
645 if (Opcode == Instruction::PHI && CostKind != TTI::TCK_RecipThroughput)
646 return 0;
647 return 1;
648 }
649
650 InstructionCost getCmpSelInstrCost(unsigned Opcode, Type *ValTy, Type *CondTy,
651 CmpInst::Predicate VecPred,
653 const Instruction *I) const {
654 return 1;
655 }
656
659 unsigned Index, Value *Op0,
660 Value *Op1) const {
661 return 1;
662 }
663
666 unsigned Index) const {
667 return 1;
668 }
669
670 unsigned getReplicationShuffleCost(Type *EltTy, int ReplicationFactor, int VF,
671 const APInt &DemandedDstElts,
673 return 1;
674 }
675
676 InstructionCost getMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment,
677 unsigned AddressSpace,
680 const Instruction *I) const {
681 return 1;
682 }
683
684 InstructionCost getVPMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment,
685 unsigned AddressSpace,
687 const Instruction *I) const {
688 return 1;
689 }
690
692 Align Alignment, unsigned AddressSpace,
694 return 1;
695 }
696
698 const Value *Ptr, bool VariableMask,
699 Align Alignment,
701 const Instruction *I = nullptr) const {
702 return 1;
703 }
704
706 const Value *Ptr, bool VariableMask,
707 Align Alignment,
709 const Instruction *I = nullptr) const {
711 }
712
714 unsigned Opcode, Type *VecTy, unsigned Factor, ArrayRef<unsigned> Indices,
715 Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind,
716 bool UseMaskForCond, bool UseMaskForGaps) const {
717 return 1;
718 }
719
722 switch (ICA.getID()) {
723 default:
724 break;
725 case Intrinsic::allow_runtime_check:
726 case Intrinsic::allow_ubsan_check:
727 case Intrinsic::annotation:
728 case Intrinsic::assume:
729 case Intrinsic::sideeffect:
730 case Intrinsic::pseudoprobe:
731 case Intrinsic::arithmetic_fence:
732 case Intrinsic::dbg_assign:
733 case Intrinsic::dbg_declare:
734 case Intrinsic::dbg_value:
735 case Intrinsic::dbg_label:
736 case Intrinsic::invariant_start:
737 case Intrinsic::invariant_end:
738 case Intrinsic::launder_invariant_group:
739 case Intrinsic::strip_invariant_group:
740 case Intrinsic::is_constant:
741 case Intrinsic::lifetime_start:
742 case Intrinsic::lifetime_end:
743 case Intrinsic::experimental_noalias_scope_decl:
744 case Intrinsic::objectsize:
745 case Intrinsic::ptr_annotation:
746 case Intrinsic::var_annotation:
747 case Intrinsic::experimental_gc_result:
748 case Intrinsic::experimental_gc_relocate:
749 case Intrinsic::coro_alloc:
750 case Intrinsic::coro_begin:
751 case Intrinsic::coro_free:
752 case Intrinsic::coro_end:
753 case Intrinsic::coro_frame:
754 case Intrinsic::coro_size:
755 case Intrinsic::coro_align:
756 case Intrinsic::coro_suspend:
757 case Intrinsic::coro_subfn_addr:
758 case Intrinsic::threadlocal_address:
759 case Intrinsic::experimental_widenable_condition:
760 case Intrinsic::ssa_copy:
761 // These intrinsics don't actually represent code after lowering.
762 return 0;
763 }
764 return 1;
765 }
766
770 return 1;
771 }
772
773 // Assume that we have a register of the right size for the type.
774 unsigned getNumberOfParts(Type *Tp) const { return 1; }
775
777 const SCEV *) const {
778 return 0;
779 }
780
782 std::optional<FastMathFlags> FMF,
783 TTI::TargetCostKind) const {
784 return 1;
785 }
786
789 TTI::TargetCostKind) const {
790 return 1;
791 }
792
793 InstructionCost getExtendedReductionCost(unsigned Opcode, bool IsUnsigned,
794 Type *ResTy, VectorType *Ty,
795 FastMathFlags FMF,
797 return 1;
798 }
799
801 VectorType *Ty,
803 return 1;
804 }
805
807 return 0;
808 }
809
811 return false;
812 }
813
815 // Note for overrides: You must ensure for all element unordered-atomic
816 // memory intrinsics that all power-of-2 element sizes up to, and
817 // including, the return value of this method have a corresponding
818 // runtime lib call. These runtime lib call definitions can be found
819 // in RuntimeLibcalls.h
820 return 0;
821 }
822
824 Type *ExpectedType) const {
825 return nullptr;
826 }
827
828 Type *
830 unsigned SrcAddrSpace, unsigned DestAddrSpace,
831 unsigned SrcAlign, unsigned DestAlign,
832 std::optional<uint32_t> AtomicElementSize) const {
833 return AtomicElementSize ? Type::getIntNTy(Context, *AtomicElementSize * 8)
835 }
836
838 SmallVectorImpl<Type *> &OpsOut, LLVMContext &Context,
839 unsigned RemainingBytes, unsigned SrcAddrSpace, unsigned DestAddrSpace,
840 unsigned SrcAlign, unsigned DestAlign,
841 std::optional<uint32_t> AtomicCpySize) const {
842 unsigned OpSizeInBytes = AtomicCpySize ? *AtomicCpySize : 1;
843 Type *OpType = Type::getIntNTy(Context, OpSizeInBytes * 8);
844 for (unsigned i = 0; i != RemainingBytes; i += OpSizeInBytes)
845 OpsOut.push_back(OpType);
846 }
847
848 bool areInlineCompatible(const Function *Caller,
849 const Function *Callee) const {
850 return (Caller->getFnAttribute("target-cpu") ==
851 Callee->getFnAttribute("target-cpu")) &&
852 (Caller->getFnAttribute("target-features") ==
853 Callee->getFnAttribute("target-features"));
854 }
855
856 unsigned getInlineCallPenalty(const Function *F, const CallBase &Call,
857 unsigned DefaultCallPenalty) const {
858 return DefaultCallPenalty;
859 }
860
861 bool areTypesABICompatible(const Function *Caller, const Function *Callee,
862 const ArrayRef<Type *> &Types) const {
863 return (Caller->getFnAttribute("target-cpu") ==
864 Callee->getFnAttribute("target-cpu")) &&
865 (Caller->getFnAttribute("target-features") ==
866 Callee->getFnAttribute("target-features"));
867 }
868
870 const DataLayout &DL) const {
871 return false;
872 }
873
875 const DataLayout &DL) const {
876 return false;
877 }
878
879 unsigned getLoadStoreVecRegBitWidth(unsigned AddrSpace) const { return 128; }
880
881 bool isLegalToVectorizeLoad(LoadInst *LI) const { return true; }
882
883 bool isLegalToVectorizeStore(StoreInst *SI) const { return true; }
884
885 bool isLegalToVectorizeLoadChain(unsigned ChainSizeInBytes, Align Alignment,
886 unsigned AddrSpace) const {
887 return true;
888 }
889
890 bool isLegalToVectorizeStoreChain(unsigned ChainSizeInBytes, Align Alignment,
891 unsigned AddrSpace) const {
892 return true;
893 }
894
896 ElementCount VF) const {
897 return true;
898 }
899
900 bool isElementTypeLegalForScalableVector(Type *Ty) const { return true; }
901
902 unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize,
903 unsigned ChainSizeInBytes,
904 VectorType *VecTy) const {
905 return VF;
906 }
907
908 unsigned getStoreVectorFactor(unsigned VF, unsigned StoreSize,
909 unsigned ChainSizeInBytes,
910 VectorType *VecTy) const {
911 return VF;
912 }
913
914 bool preferInLoopReduction(unsigned Opcode, Type *Ty,
915 TTI::ReductionFlags Flags) const {
916 return false;
917 }
918
919 bool preferPredicatedReductionSelect(unsigned Opcode, Type *Ty,
920 TTI::ReductionFlags Flags) const {
921 return false;
922 }
923
925 return true;
926 }
927
928 bool shouldExpandReduction(const IntrinsicInst *II) const { return true; }
929
930 unsigned getGISelRematGlobalCost() const { return 1; }
931
932 unsigned getMinTripCountTailFoldingThreshold() const { return 0; }
933
934 bool supportsScalableVectors() const { return false; }
935
936 bool enableScalableVectorization() const { return false; }
937
938 bool hasActiveVectorLength(unsigned Opcode, Type *DataType,
939 Align Alignment) const {
940 return false;
941 }
942
947 /* OperatorStrategy */ TargetTransformInfo::VPLegalization::Convert);
948 }
949
950 bool hasArmWideBranch(bool) const { return false; }
951
952 unsigned getMaxNumArgs() const { return UINT_MAX; }
953
954protected:
955 // Obtain the minimum required size to hold the value (without the sign)
956 // In case of a vector it returns the min required size for one element.
957 unsigned minRequiredElementSize(const Value *Val, bool &isSigned) const {
958 if (isa<ConstantDataVector>(Val) || isa<ConstantVector>(Val)) {
959 const auto *VectorValue = cast<Constant>(Val);
960
961 // In case of a vector need to pick the max between the min
962 // required size for each element
963 auto *VT = cast<FixedVectorType>(Val->getType());
964
965 // Assume unsigned elements
966 isSigned = false;
967
968 // The max required size is the size of the vector element type
969 unsigned MaxRequiredSize =
970 VT->getElementType()->getPrimitiveSizeInBits().getFixedValue();
971
972 unsigned MinRequiredSize = 0;
973 for (unsigned i = 0, e = VT->getNumElements(); i < e; ++i) {
974 if (auto *IntElement =
975 dyn_cast<ConstantInt>(VectorValue->getAggregateElement(i))) {
976 bool signedElement = IntElement->getValue().isNegative();
977 // Get the element min required size.
978 unsigned ElementMinRequiredSize =
979 IntElement->getValue().getSignificantBits() - 1;
980 // In case one element is signed then all the vector is signed.
981 isSigned |= signedElement;
982 // Save the max required bit size between all the elements.
983 MinRequiredSize = std::max(MinRequiredSize, ElementMinRequiredSize);
984 } else {
985 // not an int constant element
986 return MaxRequiredSize;
987 }
988 }
989 return MinRequiredSize;
990 }
991
992 if (const auto *CI = dyn_cast<ConstantInt>(Val)) {
993 isSigned = CI->getValue().isNegative();
994 return CI->getValue().getSignificantBits() - 1;
995 }
996
997 if (const auto *Cast = dyn_cast<SExtInst>(Val)) {
998 isSigned = true;
999 return Cast->getSrcTy()->getScalarSizeInBits() - 1;
1000 }
1001
1002 if (const auto *Cast = dyn_cast<ZExtInst>(Val)) {
1003 isSigned = false;
1004 return Cast->getSrcTy()->getScalarSizeInBits();
1005 }
1006
1007 isSigned = false;
1008 return Val->getType()->getScalarSizeInBits();
1009 }
1010
1011 bool isStridedAccess(const SCEV *Ptr) const {
1012 return Ptr && isa<SCEVAddRecExpr>(Ptr);
1013 }
1014
1016 const SCEV *Ptr) const {
1017 if (!isStridedAccess(Ptr))
1018 return nullptr;
1019 const SCEVAddRecExpr *AddRec = cast<SCEVAddRecExpr>(Ptr);
1020 return dyn_cast<SCEVConstant>(AddRec->getStepRecurrence(*SE));
1021 }
1022
1024 int64_t MergeDistance) const {
1025 const SCEVConstant *Step = getConstantStrideStep(SE, Ptr);
1026 if (!Step)
1027 return false;
1028 APInt StrideVal = Step->getAPInt();
1029 if (StrideVal.getBitWidth() > 64)
1030 return false;
1031 // FIXME: Need to take absolute value for negative stride case.
1032 return StrideVal.getSExtValue() < MergeDistance;
1033 }
1034};
1035
1036/// CRTP base class for use as a mix-in that aids implementing
1037/// a TargetTransformInfo-compatible class.
1038template <typename T>
1040private:
1042
1043protected:
1045
1046public:
1048
1052 assert(PointeeType && Ptr && "can't get GEPCost of nullptr");
1053 auto *BaseGV = dyn_cast<GlobalValue>(Ptr->stripPointerCasts());
1054 bool HasBaseReg = (BaseGV == nullptr);
1055
1056 auto PtrSizeBits = DL.getPointerTypeSizeInBits(Ptr->getType());
1057 APInt BaseOffset(PtrSizeBits, 0);
1058 int64_t Scale = 0;
1059
1060 auto GTI = gep_type_begin(PointeeType, Operands);
1061 Type *TargetType = nullptr;
1062
1063 // Handle the case where the GEP instruction has a single operand,
1064 // the basis, therefore TargetType is a nullptr.
1065 if (Operands.empty())
1066 return !BaseGV ? TTI::TCC_Free : TTI::TCC_Basic;
1067
1068 for (auto I = Operands.begin(); I != Operands.end(); ++I, ++GTI) {
1069 TargetType = GTI.getIndexedType();
1070 // We assume that the cost of Scalar GEP with constant index and the
1071 // cost of Vector GEP with splat constant index are the same.
1072 const ConstantInt *ConstIdx = dyn_cast<ConstantInt>(*I);
1073 if (!ConstIdx)
1074 if (auto Splat = getSplatValue(*I))
1075 ConstIdx = dyn_cast<ConstantInt>(Splat);
1076 if (StructType *STy = GTI.getStructTypeOrNull()) {
1077 // For structures the index is always splat or scalar constant
1078 assert(ConstIdx && "Unexpected GEP index");
1079 uint64_t Field = ConstIdx->getZExtValue();
1080 BaseOffset += DL.getStructLayout(STy)->getElementOffset(Field);
1081 } else {
1082 // If this operand is a scalable type, bail out early.
1083 // TODO: Make isLegalAddressingMode TypeSize aware.
1084 if (TargetType->isScalableTy())
1085 return TTI::TCC_Basic;
1086 int64_t ElementSize =
1087 GTI.getSequentialElementStride(DL).getFixedValue();
1088 if (ConstIdx) {
1089 BaseOffset +=
1090 ConstIdx->getValue().sextOrTrunc(PtrSizeBits) * ElementSize;
1091 } else {
1092 // Needs scale register.
1093 if (Scale != 0)
1094 // No addressing mode takes two scale registers.
1095 return TTI::TCC_Basic;
1096 Scale = ElementSize;
1097 }
1098 }
1099 }
1100
1101 // If we haven't been provided a hint, use the target type for now.
1102 //
1103 // TODO: Take a look at potentially removing this: This is *slightly* wrong
1104 // as it's possible to have a GEP with a foldable target type but a memory
1105 // access that isn't foldable. For example, this load isn't foldable on
1106 // RISC-V:
1107 //
1108 // %p = getelementptr i32, ptr %base, i32 42
1109 // %x = load <2 x i32>, ptr %p
1110 if (!AccessType)
1111 AccessType = TargetType;
1112
1113 // If the final address of the GEP is a legal addressing mode for the given
1114 // access type, then we can fold it into its users.
1115 if (static_cast<T *>(this)->isLegalAddressingMode(
1116 AccessType, const_cast<GlobalValue *>(BaseGV),
1117 BaseOffset.sextOrTrunc(64).getSExtValue(), HasBaseReg, Scale,
1118 Ptr->getType()->getPointerAddressSpace()))
1119 return TTI::TCC_Free;
1120
1121 // TODO: Instead of returning TCC_Basic here, we should use
1122 // getArithmeticInstrCost. Or better yet, provide a hook to let the target
1123 // model it.
1124 return TTI::TCC_Basic;
1125 }
1126
1128 const Value *Base,
1130 Type *AccessTy,
1133 // In the basic model we take into account GEP instructions only
1134 // (although here can come alloca instruction, a value, constants and/or
1135 // constant expressions, PHIs, bitcasts ... whatever allowed to be used as a
1136 // pointer). Typically, if Base is a not a GEP-instruction and all the
1137 // pointers are relative to the same base address, all the rest are
1138 // either GEP instructions, PHIs, bitcasts or constants. When we have same
1139 // base, we just calculate cost of each non-Base GEP as an ADD operation if
1140 // any their index is a non-const.
1141 // If no known dependecies between the pointers cost is calculated as a sum
1142 // of costs of GEP instructions.
1143 for (const Value *V : Ptrs) {
1144 const auto *GEP = dyn_cast<GetElementPtrInst>(V);
1145 if (!GEP)
1146 continue;
1147 if (Info.isSameBase() && V != Base) {
1148 if (GEP->hasAllConstantIndices())
1149 continue;
1150 Cost += static_cast<T *>(this)->getArithmeticInstrCost(
1151 Instruction::Add, GEP->getType(), CostKind,
1153 std::nullopt);
1154 } else {
1155 SmallVector<const Value *> Indices(GEP->indices());
1156 Cost += static_cast<T *>(this)->getGEPCost(GEP->getSourceElementType(),
1157 GEP->getPointerOperand(),
1158 Indices, AccessTy, CostKind);
1159 }
1160 }
1161 return Cost;
1162 }
1163
1167 using namespace llvm::PatternMatch;
1168
1169 auto *TargetTTI = static_cast<T *>(this);
1170 // Handle non-intrinsic calls, invokes, and callbr.
1171 // FIXME: Unlikely to be true for anything but CodeSize.
1172 auto *CB = dyn_cast<CallBase>(U);
1173 if (CB && !isa<IntrinsicInst>(U)) {
1174 if (const Function *F = CB->getCalledFunction()) {
1175 if (!TargetTTI->isLoweredToCall(F))
1176 return TTI::TCC_Basic; // Give a basic cost if it will be lowered
1177
1178 return TTI::TCC_Basic * (F->getFunctionType()->getNumParams() + 1);
1179 }
1180 // For indirect or other calls, scale cost by number of arguments.
1181 return TTI::TCC_Basic * (CB->arg_size() + 1);
1182 }
1183
1184 Type *Ty = U->getType();
1185 unsigned Opcode = Operator::getOpcode(U);
1186 auto *I = dyn_cast<Instruction>(U);
1187 switch (Opcode) {
1188 default:
1189 break;
1190 case Instruction::Call: {
1191 assert(isa<IntrinsicInst>(U) && "Unexpected non-intrinsic call");
1192 auto *Intrinsic = cast<IntrinsicInst>(U);
1193 IntrinsicCostAttributes CostAttrs(Intrinsic->getIntrinsicID(), *CB);
1194 return TargetTTI->getIntrinsicInstrCost(CostAttrs, CostKind);
1195 }
1196 case Instruction::Br:
1197 case Instruction::Ret:
1198 case Instruction::PHI:
1199 case Instruction::Switch:
1200 return TargetTTI->getCFInstrCost(Opcode, CostKind, I);
1201 case Instruction::ExtractValue:
1202 case Instruction::Freeze:
1203 return TTI::TCC_Free;
1204 case Instruction::Alloca:
1205 if (cast<AllocaInst>(U)->isStaticAlloca())
1206 return TTI::TCC_Free;
1207 break;
1208 case Instruction::GetElementPtr: {
1209 const auto *GEP = cast<GEPOperator>(U);
1210 Type *AccessType = nullptr;
1211 // For now, only provide the AccessType in the simple case where the GEP
1212 // only has one user.
1213 if (GEP->hasOneUser() && I)
1214 AccessType = I->user_back()->getAccessType();
1215
1216 return TargetTTI->getGEPCost(GEP->getSourceElementType(),
1217 Operands.front(), Operands.drop_front(),
1218 AccessType, CostKind);
1219 }
1220 case Instruction::Add:
1221 case Instruction::FAdd:
1222 case Instruction::Sub:
1223 case Instruction::FSub:
1224 case Instruction::Mul:
1225 case Instruction::FMul:
1226 case Instruction::UDiv:
1227 case Instruction::SDiv:
1228 case Instruction::FDiv:
1229 case Instruction::URem:
1230 case Instruction::SRem:
1231 case Instruction::FRem:
1232 case Instruction::Shl:
1233 case Instruction::LShr:
1234 case Instruction::AShr:
1235 case Instruction::And:
1236 case Instruction::Or:
1237 case Instruction::Xor:
1238 case Instruction::FNeg: {
1240 TTI::OperandValueInfo Op2Info;
1241 if (Opcode != Instruction::FNeg)
1242 Op2Info = TTI::getOperandInfo(Operands[1]);
1243 return TargetTTI->getArithmeticInstrCost(Opcode, Ty, CostKind, Op1Info,
1244 Op2Info, Operands, I);
1245 }
1246 case Instruction::IntToPtr:
1247 case Instruction::PtrToInt:
1248 case Instruction::SIToFP:
1249 case Instruction::UIToFP:
1250 case Instruction::FPToUI:
1251 case Instruction::FPToSI:
1252 case Instruction::Trunc:
1253 case Instruction::FPTrunc:
1254 case Instruction::BitCast:
1255 case Instruction::FPExt:
1256 case Instruction::SExt:
1257 case Instruction::ZExt:
1258 case Instruction::AddrSpaceCast: {
1259 Type *OpTy = Operands[0]->getType();
1260 return TargetTTI->getCastInstrCost(
1261 Opcode, Ty, OpTy, TTI::getCastContextHint(I), CostKind, I);
1262 }
1263 case Instruction::Store: {
1264 auto *SI = cast<StoreInst>(U);
1265 Type *ValTy = Operands[0]->getType();
1267 return TargetTTI->getMemoryOpCost(Opcode, ValTy, SI->getAlign(),
1268 SI->getPointerAddressSpace(), CostKind,
1269 OpInfo, I);
1270 }
1271 case Instruction::Load: {
1272 // FIXME: Arbitary cost which could come from the backend.
1274 return 4;
1275 auto *LI = cast<LoadInst>(U);
1276 Type *LoadType = U->getType();
1277 // If there is a non-register sized type, the cost estimation may expand
1278 // it to be several instructions to load into multiple registers on the
1279 // target. But, if the only use of the load is a trunc instruction to a
1280 // register sized type, the instruction selector can combine these
1281 // instructions to be a single load. So, in this case, we use the
1282 // destination type of the trunc instruction rather than the load to
1283 // accurately estimate the cost of this load instruction.
1284 if (CostKind == TTI::TCK_CodeSize && LI->hasOneUse() &&
1285 !LoadType->isVectorTy()) {
1286 if (const TruncInst *TI = dyn_cast<TruncInst>(*LI->user_begin()))
1287 LoadType = TI->getDestTy();
1288 }
1289 return TargetTTI->getMemoryOpCost(Opcode, LoadType, LI->getAlign(),
1291 {TTI::OK_AnyValue, TTI::OP_None}, I);
1292 }
1293 case Instruction::Select: {
1294 const Value *Op0, *Op1;
1295 if (match(U, m_LogicalAnd(m_Value(Op0), m_Value(Op1))) ||
1296 match(U, m_LogicalOr(m_Value(Op0), m_Value(Op1)))) {
1297 // select x, y, false --> x & y
1298 // select x, true, y --> x | y
1299 const auto Op1Info = TTI::getOperandInfo(Op0);
1300 const auto Op2Info = TTI::getOperandInfo(Op1);
1301 assert(Op0->getType()->getScalarSizeInBits() == 1 &&
1302 Op1->getType()->getScalarSizeInBits() == 1);
1303
1305 return TargetTTI->getArithmeticInstrCost(
1306 match(U, m_LogicalOr()) ? Instruction::Or : Instruction::And, Ty,
1307 CostKind, Op1Info, Op2Info, Operands, I);
1308 }
1309 Type *CondTy = Operands[0]->getType();
1310 return TargetTTI->getCmpSelInstrCost(Opcode, U->getType(), CondTy,
1312 CostKind, I);
1313 }
1314 case Instruction::ICmp:
1315 case Instruction::FCmp: {
1316 Type *ValTy = Operands[0]->getType();
1317 // TODO: Also handle ICmp/FCmp constant expressions.
1318 return TargetTTI->getCmpSelInstrCost(Opcode, ValTy, U->getType(),
1319 I ? cast<CmpInst>(I)->getPredicate()
1321 CostKind, I);
1322 }
1323 case Instruction::InsertElement: {
1324 auto *IE = dyn_cast<InsertElementInst>(U);
1325 if (!IE)
1326 return TTI::TCC_Basic; // FIXME
1327 unsigned Idx = -1;
1328 if (auto *CI = dyn_cast<ConstantInt>(Operands[2]))
1329 if (CI->getValue().getActiveBits() <= 32)
1330 Idx = CI->getZExtValue();
1331 return TargetTTI->getVectorInstrCost(*IE, Ty, CostKind, Idx);
1332 }
1333 case Instruction::ShuffleVector: {
1334 auto *Shuffle = dyn_cast<ShuffleVectorInst>(U);
1335 if (!Shuffle)
1336 return TTI::TCC_Basic; // FIXME
1337
1338 auto *VecTy = cast<VectorType>(U->getType());
1339 auto *VecSrcTy = cast<VectorType>(Operands[0]->getType());
1340 ArrayRef<int> Mask = Shuffle->getShuffleMask();
1341 int NumSubElts, SubIndex;
1342
1343 // TODO: move more of this inside improveShuffleKindFromMask.
1344 if (Shuffle->changesLength()) {
1345 // Treat a 'subvector widening' as a free shuffle.
1346 if (Shuffle->increasesLength() && Shuffle->isIdentityWithPadding())
1347 return 0;
1348
1349 if (Shuffle->isExtractSubvectorMask(SubIndex))
1350 return TargetTTI->getShuffleCost(TTI::SK_ExtractSubvector, VecSrcTy,
1351 Mask, CostKind, SubIndex, VecTy,
1352 Operands, Shuffle);
1353
1354 if (Shuffle->isInsertSubvectorMask(NumSubElts, SubIndex))
1355 return TargetTTI->getShuffleCost(
1356 TTI::SK_InsertSubvector, VecTy, Mask, CostKind, SubIndex,
1357 FixedVectorType::get(VecTy->getScalarType(), NumSubElts),
1358 Operands, Shuffle);
1359
1360 int ReplicationFactor, VF;
1361 if (Shuffle->isReplicationMask(ReplicationFactor, VF)) {
1362 APInt DemandedDstElts = APInt::getZero(Mask.size());
1363 for (auto I : enumerate(Mask)) {
1364 if (I.value() != PoisonMaskElem)
1365 DemandedDstElts.setBit(I.index());
1366 }
1367 return TargetTTI->getReplicationShuffleCost(
1368 VecSrcTy->getElementType(), ReplicationFactor, VF,
1369 DemandedDstElts, CostKind);
1370 }
1371
1372 bool IsUnary = isa<UndefValue>(Operands[1]);
1373 NumSubElts = VecSrcTy->getElementCount().getKnownMinValue();
1374 SmallVector<int, 16> AdjustMask(Mask.begin(), Mask.end());
1375
1376 // Widening shuffle - widening the source(s) to the new length
1377 // (treated as free - see above), and then perform the adjusted
1378 // shuffle at that width.
1379 if (Shuffle->increasesLength()) {
1380 for (int &M : AdjustMask)
1381 M = M >= NumSubElts ? (M + (Mask.size() - NumSubElts)) : M;
1382
1383 return TargetTTI->getShuffleCost(
1385 AdjustMask, CostKind, 0, nullptr, Operands, Shuffle);
1386 }
1387
1388 // Narrowing shuffle - perform shuffle at original wider width and
1389 // then extract the lower elements.
1390 AdjustMask.append(NumSubElts - Mask.size(), PoisonMaskElem);
1391
1392 InstructionCost ShuffleCost = TargetTTI->getShuffleCost(
1394 VecSrcTy, AdjustMask, CostKind, 0, nullptr, Operands, Shuffle);
1395
1396 SmallVector<int, 16> ExtractMask(Mask.size());
1397 std::iota(ExtractMask.begin(), ExtractMask.end(), 0);
1398 return ShuffleCost + TargetTTI->getShuffleCost(
1399 TTI::SK_ExtractSubvector, VecSrcTy,
1400 ExtractMask, CostKind, 0, VecTy, {}, Shuffle);
1401 }
1402
1403 if (Shuffle->isIdentity())
1404 return 0;
1405
1406 if (Shuffle->isReverse())
1407 return TargetTTI->getShuffleCost(TTI::SK_Reverse, VecTy, Mask, CostKind,
1408 0, nullptr, Operands, Shuffle);
1409
1410 if (Shuffle->isSelect())
1411 return TargetTTI->getShuffleCost(TTI::SK_Select, VecTy, Mask, CostKind,
1412 0, nullptr, Operands, Shuffle);
1413
1414 if (Shuffle->isTranspose())
1415 return TargetTTI->getShuffleCost(TTI::SK_Transpose, VecTy, Mask,
1416 CostKind, 0, nullptr, Operands,
1417 Shuffle);
1418
1419 if (Shuffle->isZeroEltSplat())
1420 return TargetTTI->getShuffleCost(TTI::SK_Broadcast, VecTy, Mask,
1421 CostKind, 0, nullptr, Operands,
1422 Shuffle);
1423
1424 if (Shuffle->isSingleSource())
1425 return TargetTTI->getShuffleCost(TTI::SK_PermuteSingleSrc, VecTy, Mask,
1426 CostKind, 0, nullptr, Operands,
1427 Shuffle);
1428
1429 if (Shuffle->isInsertSubvectorMask(NumSubElts, SubIndex))
1430 return TargetTTI->getShuffleCost(
1431 TTI::SK_InsertSubvector, VecTy, Mask, CostKind, SubIndex,
1432 FixedVectorType::get(VecTy->getScalarType(), NumSubElts), Operands,
1433 Shuffle);
1434
1435 if (Shuffle->isSplice(SubIndex))
1436 return TargetTTI->getShuffleCost(TTI::SK_Splice, VecTy, Mask, CostKind,
1437 SubIndex, nullptr, Operands, Shuffle);
1438
1439 return TargetTTI->getShuffleCost(TTI::SK_PermuteTwoSrc, VecTy, Mask,
1440 CostKind, 0, nullptr, Operands, Shuffle);
1441 }
1442 case Instruction::ExtractElement: {
1443 auto *EEI = dyn_cast<ExtractElementInst>(U);
1444 if (!EEI)
1445 return TTI::TCC_Basic; // FIXME
1446 unsigned Idx = -1;
1447 if (auto *CI = dyn_cast<ConstantInt>(Operands[1]))
1448 if (CI->getValue().getActiveBits() <= 32)
1449 Idx = CI->getZExtValue();
1450 Type *DstTy = Operands[0]->getType();
1451 return TargetTTI->getVectorInstrCost(*EEI, DstTy, CostKind, Idx);
1452 }
1453 }
1454
1455 // By default, just classify everything as 'basic' or -1 to represent that
1456 // don't know the throughput cost.
1458 }
1459
1461 auto *TargetTTI = static_cast<T *>(this);
1462 SmallVector<const Value *, 4> Ops(I->operand_values());
1463 InstructionCost Cost = TargetTTI->getInstructionCost(
1466 }
1467
1468 bool supportsTailCallFor(const CallBase *CB) const {
1469 return static_cast<const T *>(this)->supportsTailCalls();
1470 }
1471};
1472} // namespace llvm
1473
1474#endif
Analysis containing CSE Info
Definition: CSEInfo.cpp:27
static cl::opt< TargetTransformInfo::TargetCostKind > CostKind("cost-kind", cl::desc("Target cost kind"), cl::init(TargetTransformInfo::TCK_RecipThroughput), cl::values(clEnumValN(TargetTransformInfo::TCK_RecipThroughput, "throughput", "Reciprocal throughput"), clEnumValN(TargetTransformInfo::TCK_Latency, "latency", "Instruction latency"), clEnumValN(TargetTransformInfo::TCK_CodeSize, "code-size", "Code size"), clEnumValN(TargetTransformInfo::TCK_SizeAndLatency, "size-latency", "Code size and latency")))
return RetTy
Returns the sub type a function will return at a given Idx Should correspond to the result type of an ExtractValue instruction executed with just that one unsigned Idx
std::string Name
static bool isSigned(unsigned int Opcode)
Hexagon Common GEP
#define F(x, y, z)
Definition: MD5.cpp:55
#define I(x, y, z)
Definition: MD5.cpp:58
mir Rename Register Operands
LLVMContext & Context
static cl::opt< RegAllocEvictionAdvisorAnalysis::AdvisorMode > Mode("regalloc-enable-advisor", cl::Hidden, cl::init(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Default), cl::desc("Enable regalloc advisor mode"), cl::values(clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Default, "default", "Default"), clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Release, "release", "precompiled"), clEnumValN(RegAllocEvictionAdvisorAnalysis::AdvisorMode::Development, "development", "for training")))
assert(ImpDefSCC.getReg()==AMDGPU::SCC &&ImpDefSCC.isDef())
static SymbolRef::Type getType(const Symbol *Sym)
Definition: TapiFile.cpp:40
This pass exposes codegen information to IR-level passes.
Class for arbitrary precision integers.
Definition: APInt.h:76
void setBit(unsigned BitPosition)
Set the given bit to 1 whose position is given as "bitPosition".
Definition: APInt.h:1308
unsigned getBitWidth() const
Return the number of bits in the APInt.
Definition: APInt.h:1446
APInt sextOrTrunc(unsigned width) const
Sign extend or truncate to width.
Definition: APInt.cpp:1010
static APInt getZero(unsigned numBits)
Get the '0' value for the specified bit-width.
Definition: APInt.h:178
int64_t getSExtValue() const
Get sign extended value.
Definition: APInt.h:1520
an instruction to allocate memory on the stack
Definition: Instructions.h:59
ArrayRef - Represent a constant reference to an array (0 or more elements consecutively in memory),...
Definition: ArrayRef.h:41
A cache of @llvm.assume calls within a function.
BlockFrequencyInfo pass uses BlockFrequencyInfoImpl implementation to estimate IR basic block frequen...
Conditional or Unconditional Branch instruction.
Base class for all callable instructions (InvokeInst and CallInst) Holds everything related to callin...
Definition: InstrTypes.h:1494
Predicate
This enumeration lists the possible predicates for CmpInst subclasses.
Definition: InstrTypes.h:993
This is the shared class of boolean and integer constants.
Definition: Constants.h:80
uint64_t getZExtValue() const
Return the constant as a 64-bit unsigned integer value after it has been zero extended as appropriate...
Definition: Constants.h:154
const APInt & getValue() const
Return the constant as an APInt value reference.
Definition: Constants.h:145
This is an important base class in LLVM.
Definition: Constant.h:41
A parsed version of the target data layout string in and methods for querying it.
Definition: DataLayout.h:110
bool isLegalInteger(uint64_t Width) const
Returns true if the specified type is known to be a native integer type supported by the CPU.
Definition: DataLayout.h:260
const StructLayout * getStructLayout(StructType *Ty) const
Returns a StructLayout object, indicating the alignment of the struct, its size, and the offsets of i...
Definition: DataLayout.cpp:720
unsigned getPointerTypeSizeInBits(Type *) const
Layout pointer size, in bits, based on the type.
Definition: DataLayout.cpp:763
TypeSize getTypeSizeInBits(Type *Ty) const
Size examples:
Definition: DataLayout.h:672
TypeSize getTypeStoreSize(Type *Ty) const
Returns the maximum number of bytes that may be overwritten by storing the specified type.
Definition: DataLayout.h:472
Concrete subclass of DominatorTreeBase that is used to compute a normal dominator tree.
Definition: Dominators.h:162
static constexpr ElementCount get(ScalarTy MinVal, bool Scalable)
Definition: TypeSize.h:314
Convenience struct for specifying and reasoning about fast-math flags.
Definition: FMF.h:20
static FixedVectorType * get(Type *ElementType, unsigned NumElts)
Definition: Type.cpp:692
The core instruction combiner logic.
Definition: InstCombiner.h:47
static InstructionCost getInvalid(CostType Val=0)
A wrapper class for inspecting calls to intrinsic functions.
Definition: IntrinsicInst.h:47
Intrinsic::ID getIntrinsicID() const
Return the intrinsic ID of this intrinsic.
Definition: IntrinsicInst.h:54
This is an important class for using LLVM in a threaded context.
Definition: LLVMContext.h:67
An instruction for reading from memory.
Definition: Instructions.h:184
Represents a single loop in the control flow graph.
Definition: LoopInfo.h:44
unsigned getOpcode() const
Return the opcode for this Instruction or ConstantExpr.
Definition: Operator.h:41
The optimization diagnostic interface.
Analysis providing profile information.
The RecurrenceDescriptor is used to identify recurrences variables in a loop.
Definition: IVDescriptors.h:71
This node represents a polynomial recurrence on the trip count of the specified loop.
const SCEV * getStepRecurrence(ScalarEvolution &SE) const
Constructs and returns the recurrence indicating how much this expression steps by.
This class represents a constant integer value.
const APInt & getAPInt() const
This class represents an analyzed expression in the program.
The main scalar evolution driver.
This is a 'bitvector' (really, a variable-sized bit array), optimized for the case when the array is ...
This class consists of common code factored out of the SmallVector class to reduce code duplication b...
Definition: SmallVector.h:586
void append(ItTy in_start, ItTy in_end)
Add the specified range to the end of the SmallVector.
Definition: SmallVector.h:696
void push_back(const T &Elt)
Definition: SmallVector.h:426
This is a 'vector' (really, a variable-sized array), optimized for the case when the array is small.
Definition: SmallVector.h:1209
StackOffset holds a fixed and a scalable offset in bytes.
Definition: TypeSize.h:33
static StackOffset getScalable(int64_t Scalable)
Definition: TypeSize.h:43
static StackOffset getFixed(int64_t Fixed)
Definition: TypeSize.h:42
An instruction for storing to memory.
Definition: Instructions.h:317
StringRef - Represent a constant reference to a string, i.e.
Definition: StringRef.h:50
TypeSize getElementOffset(unsigned Idx) const
Definition: DataLayout.h:651
Class to represent struct types.
Definition: DerivedTypes.h:216
Multiway switch.
Provides information about what library functions are available for the current target.
Base class for use as a mix-in that aids implementing a TargetTransformInfo-compatible class.
const DataLayout & getDataLayout() const
void getMemcpyLoopResidualLoweringType(SmallVectorImpl< Type * > &OpsOut, LLVMContext &Context, unsigned RemainingBytes, unsigned SrcAddrSpace, unsigned DestAddrSpace, unsigned SrcAlign, unsigned DestAlign, std::optional< uint32_t > AtomicCpySize) const
bool isLegalToVectorizeStore(StoreInst *SI) const
bool addrspacesMayAlias(unsigned AS0, unsigned AS1) const
bool shouldTreatInstructionLikeSelect(const Instruction *I)
bool isLegalToVectorizeLoad(LoadInst *LI) const
bool isLegalBroadcastLoad(Type *ElementTy, ElementCount NumElements) const
std::optional< unsigned > getVScaleForTuning() const
bool shouldMaximizeVectorBandwidth(TargetTransformInfo::RegisterKind K) const
bool canSaveCmp(Loop *L, BranchInst **BI, ScalarEvolution *SE, LoopInfo *LI, DominatorTree *DT, AssumptionCache *AC, TargetLibraryInfo *LibInfo) const
bool isLegalStridedLoadStore(Type *DataType, Align Alignment) const
std::optional< Value * > simplifyDemandedVectorEltsIntrinsic(InstCombiner &IC, IntrinsicInst &II, APInt DemandedElts, APInt &UndefElts, APInt &UndefElts2, APInt &UndefElts3, std::function< void(Instruction *, unsigned, APInt, APInt &)> SimplifyAndSetOp) const
bool isLegalICmpImmediate(int64_t Imm) const
unsigned getRegUsageForType(Type *Ty) const
unsigned getInterleavedMemoryOpCost(unsigned Opcode, Type *VecTy, unsigned Factor, ArrayRef< unsigned > Indices, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind, bool UseMaskForCond, bool UseMaskForGaps) const
bool areTypesABICompatible(const Function *Caller, const Function *Callee, const ArrayRef< Type * > &Types) const
void getPeelingPreferences(Loop *, ScalarEvolution &, TTI::PeelingPreferences &) const
bool isAlwaysUniform(const Value *V) const
bool isProfitableToHoist(Instruction *I) const
unsigned getEstimatedNumberOfCaseClusters(const SwitchInst &SI, unsigned &JTSize, ProfileSummaryInfo *PSI, BlockFrequencyInfo *BFI) const
InstructionCost getIntImmCostInst(unsigned Opcode, unsigned Idx, const APInt &Imm, Type *Ty, TTI::TargetCostKind CostKind, Instruction *Inst=nullptr) const
bool isLSRCostLess(const TTI::LSRCost &C1, const TTI::LSRCost &C2) const
bool isExpensiveToSpeculativelyExecute(const Instruction *I)
bool isTruncateFree(Type *Ty1, Type *Ty2) const
bool isStridedAccess(const SCEV *Ptr) const
InstructionCost getCallInstrCost(Function *F, Type *RetTy, ArrayRef< Type * > Tys, TTI::TargetCostKind CostKind) const
InstructionCost getArithmeticReductionCost(unsigned, VectorType *, std::optional< FastMathFlags > FMF, TTI::TargetCostKind) const
InstructionCost getFPOpCost(Type *Ty) const
bool shouldConsiderAddressTypePromotion(const Instruction &I, bool &AllowPromotionWithoutCommonHeader) const
bool areInlineCompatible(const Function *Caller, const Function *Callee) const
std::pair< const Value *, unsigned > getPredicatedAddrSpace(const Value *V) const
InstructionCost getMemcpyCost(const Instruction *I) const
unsigned getInliningCostBenefitAnalysisProfitableMultiplier() const
unsigned getStoreVectorFactor(unsigned VF, unsigned StoreSize, unsigned ChainSizeInBytes, VectorType *VecTy) const
std::optional< unsigned > getMaxVScale() const
TTI::PopcntSupportKind getPopcntSupport(unsigned IntTyWidthInBit) const
unsigned getCallerAllocaCost(const CallBase *CB, const AllocaInst *AI) const
bool isProfitableLSRChainElement(Instruction *I) const
InstructionCost getShuffleCost(TTI::ShuffleKind Kind, VectorType *Ty, ArrayRef< int > Mask, TTI::TargetCostKind CostKind, int Index, VectorType *SubTp, ArrayRef< const Value * > Args=std::nullopt, const Instruction *CxtI=nullptr) const
InstructionCost getIntImmCodeSizeCost(unsigned Opcode, unsigned Idx, const APInt &Imm, Type *Ty) const
bool preferToKeepConstantsAttached(const Instruction &Inst, const Function &Fn) const
Value * getOrCreateResultFromMemIntrinsic(IntrinsicInst *Inst, Type *ExpectedType) const
InstructionCost getIntrinsicInstrCost(const IntrinsicCostAttributes &ICA, TTI::TargetCostKind CostKind) const
bool isLegalMaskedStore(Type *DataType, Align Alignment) const
InstructionCost getGatherScatterOpCost(unsigned Opcode, Type *DataTy, const Value *Ptr, bool VariableMask, Align Alignment, TTI::TargetCostKind CostKind, const Instruction *I=nullptr) const
unsigned getInlineCallPenalty(const Function *F, const CallBase &Call, unsigned DefaultCallPenalty) const
InstructionCost getVPMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind, const Instruction *I) const
InstructionCost getStridedMemoryOpCost(unsigned Opcode, Type *DataTy, const Value *Ptr, bool VariableMask, Align Alignment, TTI::TargetCostKind CostKind, const Instruction *I=nullptr) const
bool isNoopAddrSpaceCast(unsigned, unsigned) const
unsigned getStoreMinimumVF(unsigned VF, Type *, Type *) const
InstructionCost getVectorInstrCost(const Instruction &I, Type *Val, TTI::TargetCostKind CostKind, unsigned Index) const
bool preferPredicatedReductionSelect(unsigned Opcode, Type *Ty, TTI::ReductionFlags Flags) const
TargetTransformInfoImplBase(TargetTransformInfoImplBase &&Arg)
TargetTransformInfo::VPLegalization getVPLegalizationStrategy(const VPIntrinsic &PI) const
void getUnrollingPreferences(Loop *, ScalarEvolution &, TTI::UnrollingPreferences &, OptimizationRemarkEmitter *) const
bool isLegalToVectorizeLoadChain(unsigned ChainSizeInBytes, Align Alignment, unsigned AddrSpace) const
bool isLegalAddressingMode(Type *Ty, GlobalValue *BaseGV, int64_t BaseOffset, bool HasBaseReg, int64_t Scale, unsigned AddrSpace, Instruction *I=nullptr, int64_t ScalableOffset=0) const
bool isHardwareLoopProfitable(Loop *L, ScalarEvolution &SE, AssumptionCache &AC, TargetLibraryInfo *LibInfo, HardwareLoopInfo &HWLoopInfo) const
unsigned minRequiredElementSize(const Value *Val, bool &isSigned) const
Type * getMemcpyLoopLoweringType(LLVMContext &Context, Value *Length, unsigned SrcAddrSpace, unsigned DestAddrSpace, unsigned SrcAlign, unsigned DestAlign, std::optional< uint32_t > AtomicElementSize) const
std::optional< unsigned > getCacheSize(TargetTransformInfo::CacheLevel Level) const
unsigned getAssumedAddrSpace(const Value *V) const
bool isLegalMaskedExpandLoad(Type *DataType, Align Alignment) const
bool isLegalNTStore(Type *DataType, Align Alignment) const
unsigned getRegisterClassForType(bool Vector, Type *Ty=nullptr) const
bool isLegalMaskedGather(Type *DataType, Align Alignment) const
unsigned adjustInliningThreshold(const CallBase *CB) const
BranchProbability getPredictableBranchThreshold() const
std::optional< unsigned > getMinPageSize() const
bool collectFlatAddressOperands(SmallVectorImpl< int > &OpIndexes, Intrinsic::ID IID) const
InstructionCost getExtendedReductionCost(unsigned Opcode, bool IsUnsigned, Type *ResTy, VectorType *Ty, FastMathFlags FMF, TTI::TargetCostKind CostKind) const
bool allowsMisalignedMemoryAccesses(LLVMContext &Context, unsigned BitWidth, unsigned AddressSpace, Align Alignment, unsigned *Fast) const
const SCEVConstant * getConstantStrideStep(ScalarEvolution *SE, const SCEV *Ptr) const
unsigned getMinPrefetchStride(unsigned NumMemAccesses, unsigned NumStridedMemAccesses, unsigned NumPrefetches, bool HasCall) const
InstructionCost getIntImmCostIntrin(Intrinsic::ID IID, unsigned Idx, const APInt &Imm, Type *Ty, TTI::TargetCostKind CostKind) const
unsigned getLoadVectorFactor(unsigned VF, unsigned LoadSize, unsigned ChainSizeInBytes, VectorType *VecTy) const
bool isIndexedLoadLegal(TTI::MemIndexedMode Mode, Type *Ty, const DataLayout &DL) const
bool shouldPrefetchAddressSpace(unsigned AS) const
InstructionCost getIntImmCost(const APInt &Imm, Type *Ty, TTI::TargetCostKind CostKind) const
unsigned getReplicationShuffleCost(Type *EltTy, int ReplicationFactor, int VF, const APInt &DemandedDstElts, TTI::TargetCostKind CostKind)
bool isSourceOfDivergence(const Value *V) const
bool enableAggressiveInterleaving(bool LoopHasReductions) const
unsigned getMaxInterleaveFactor(ElementCount VF) const
InstructionCost getScalingFactorCost(Type *Ty, GlobalValue *BaseGV, StackOffset BaseOffset, bool HasBaseReg, int64_t Scale, unsigned AddrSpace) const
InstructionCost getCastInstrCost(unsigned Opcode, Type *Dst, Type *Src, TTI::CastContextHint CCH, TTI::TargetCostKind CostKind, const Instruction *I) const
std::optional< unsigned > getCacheAssociativity(TargetTransformInfo::CacheLevel Level) const
bool hasVolatileVariant(Instruction *I, unsigned AddrSpace) const
InstructionCost getArithmeticInstrCost(unsigned Opcode, Type *Ty, TTI::TargetCostKind CostKind, TTI::OperandValueInfo Opd1Info, TTI::OperandValueInfo Opd2Info, ArrayRef< const Value * > Args, const Instruction *CxtI=nullptr) const
TTI::AddressingModeKind getPreferredAddressingMode(const Loop *L, ScalarEvolution *SE) const
bool forceScalarizeMaskedGather(VectorType *DataType, Align Alignment) const
bool getTgtMemIntrinsic(IntrinsicInst *Inst, MemIntrinsicInfo &Info) const
unsigned getInliningCostBenefitAnalysisSavingsMultiplier() const
bool canHaveNonUndefGlobalInitializerInAddressSpace(unsigned AS) const
InstructionCost getMulAccReductionCost(bool IsUnsigned, Type *ResTy, VectorType *Ty, TTI::TargetCostKind CostKind) const
bool isIndexedStoreLegal(TTI::MemIndexedMode Mode, Type *Ty, const DataLayout &DL) const
bool hasDivRemOp(Type *DataType, bool IsSigned) const
InstructionCost getAddressComputationCost(Type *Tp, ScalarEvolution *, const SCEV *) const
InstructionCost getCmpSelInstrCost(unsigned Opcode, Type *ValTy, Type *CondTy, CmpInst::Predicate VecPred, TTI::TargetCostKind CostKind, const Instruction *I) const
TTI::MemCmpExpansionOptions enableMemCmpExpansion(bool OptSize, bool IsZeroCmp) const
bool preferInLoopReduction(unsigned Opcode, Type *Ty, TTI::ReductionFlags Flags) const
InstructionCost getOperandsScalarizationOverhead(ArrayRef< const Value * > Args, ArrayRef< Type * > Tys, TTI::TargetCostKind CostKind) const
TypeSize getRegisterBitWidth(TargetTransformInfo::RegisterKind K) const
bool isConstantStridedAccessLessThan(ScalarEvolution *SE, const SCEV *Ptr, int64_t MergeDistance) const
InstructionCost getMaskedMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind) const
bool isLoweredToCall(const Function *F) const
bool hasBranchDivergence(const Function *F=nullptr) const
TargetTransformInfoImplBase(const DataLayout &DL)
bool isLegalMaskedVectorHistogram(Type *AddrType, Type *DataType) const
const char * getRegisterClassName(unsigned ClassID) const
bool isElementTypeLegalForScalableVector(Type *Ty) const
bool preferPredicateOverEpilogue(TailFoldingInfo *TFI) const
bool isLegalToVectorizeReduction(const RecurrenceDescriptor &RdxDesc, ElementCount VF) const
InstructionCost getScalarizationOverhead(VectorType *Ty, const APInt &DemandedElts, bool Insert, bool Extract, TTI::TargetCostKind CostKind) const
unsigned getMaximumVF(unsigned ElemWidth, unsigned Opcode) const
TailFoldingStyle getPreferredTailFoldingStyle(bool IVUpdateMayOverflow=true) const
InstructionCost getCFInstrCost(unsigned Opcode, TTI::TargetCostKind CostKind, const Instruction *I=nullptr) const
InstructionCost getCostOfKeepingLiveOverCall(ArrayRef< Type * > Tys) const
InstructionCost getMinMaxReductionCost(Intrinsic::ID IID, VectorType *, FastMathFlags, TTI::TargetCostKind) const
InstructionCost getVectorInstrCost(unsigned Opcode, Type *Val, TTI::TargetCostKind CostKind, unsigned Index, Value *Op0, Value *Op1) const
unsigned getLoadStoreVecRegBitWidth(unsigned AddrSpace) const
InstructionCost getMemoryOpCost(unsigned Opcode, Type *Src, Align Alignment, unsigned AddressSpace, TTI::TargetCostKind CostKind, TTI::OperandValueInfo OpInfo, const Instruction *I) const
bool useColdCCForColdCall(Function &F) const
bool shouldExpandReduction(const IntrinsicInst *II) const
bool isLegalMaskedScatter(Type *DataType, Align Alignment) const
unsigned getNumberOfRegisters(unsigned ClassID) const
InstructionCost getExtractWithExtendCost(unsigned Opcode, Type *Dst, VectorType *VecTy, unsigned Index) const
InstructionCost getGEPCost(Type *PointeeType, const Value *Ptr, ArrayRef< const Value * > Operands, Type *AccessType, TTI::TargetCostKind CostKind) const
bool isLegalNTLoad(Type *DataType, Align Alignment) const
std::optional< Value * > simplifyDemandedUseBitsIntrinsic(InstCombiner &IC, IntrinsicInst &II, APInt DemandedMask, KnownBits &Known, bool &KnownBitsComputed) const
bool forceScalarizeMaskedScatter(VectorType *DataType, Align Alignment) const
bool hasActiveVectorLength(unsigned Opcode, Type *DataType, Align Alignment) const
bool isLegalAltInstr(VectorType *VecTy, unsigned Opcode0, unsigned Opcode1, const SmallBitVector &OpcodeMask) const
bool isLegalMaskedLoad(Type *DataType, Align Alignment) const
bool isValidAddrSpaceCast(unsigned FromAS, unsigned ToAS) const
std::optional< Instruction * > instCombineIntrinsic(InstCombiner &IC, IntrinsicInst &II) const
bool isLegalAddScalableImmediate(int64_t Imm) const
bool isLegalMaskedCompressStore(Type *DataType, Align Alignment) const
InstructionCost getAltInstrCost(VectorType *VecTy, unsigned Opcode0, unsigned Opcode1, const SmallBitVector &OpcodeMask, TTI::TargetCostKind CostKind) const
TargetTransformInfoImplBase(const TargetTransformInfoImplBase &Arg)=default
bool isLegalToVectorizeStoreChain(unsigned ChainSizeInBytes, Align Alignment, unsigned AddrSpace) const
ElementCount getMinimumVF(unsigned ElemWidth, bool IsScalable) const
bool shouldBuildLookupTablesForConstant(Constant *C) const
Value * rewriteIntrinsicWithAddressSpace(IntrinsicInst *II, Value *OldV, Value *NewV) const
CRTP base class for use as a mix-in that aids implementing a TargetTransformInfo-compatible class.
bool supportsTailCallFor(const CallBase *CB) const
InstructionCost getGEPCost(Type *PointeeType, const Value *Ptr, ArrayRef< const Value * > Operands, Type *AccessType, TTI::TargetCostKind CostKind)
InstructionCost getPointersChainCost(ArrayRef< const Value * > Ptrs, const Value *Base, const TTI::PointersChainInfo &Info, Type *AccessTy, TTI::TargetCostKind CostKind)
InstructionCost getInstructionCost(const User *U, ArrayRef< const Value * > Operands, TTI::TargetCostKind CostKind)
bool isExpensiveToSpeculativelyExecute(const Instruction *I)
This pass provides access to the codegen interfaces that are needed for IR-level transformations.
static CastContextHint getCastContextHint(const Instruction *I)
Calculates a CastContextHint from I.
static OperandValueInfo getOperandInfo(const Value *V)
Collect properties of V used in cost analysis, e.g. OP_PowerOf2.
TargetCostKind
The kind of cost model.
@ TCK_RecipThroughput
Reciprocal throughput.
@ TCK_CodeSize
Instruction code size.
@ TCK_SizeAndLatency
The weighted sum of size and latency.
@ TCK_Latency
The latency of instruction.
PopcntSupportKind
Flags indicating the kind of support for population count.
@ TCC_Expensive
The cost of a 'div' instruction on x86.
@ TCC_Free
Expected to fold away in lowering.
@ TCC_Basic
The cost of a typical 'add' instruction.
MemIndexedMode
The type of load/store indexing.
ShuffleKind
The various kinds of shuffle patterns for vector queries.
@ SK_InsertSubvector
InsertSubvector. Index indicates start offset.
@ SK_Select
Selects elements from the corresponding lane of either source operand.
@ SK_PermuteSingleSrc
Shuffle elements of single source vector with any shuffle mask.
@ SK_Transpose
Transpose two vectors.
@ SK_Splice
Concatenates elements from the first input vector with elements of the second input vector.
@ SK_Broadcast
Broadcast element 0 to all other elements.
@ SK_PermuteTwoSrc
Merge elements from two source vectors into one with any shuffle mask.
@ SK_Reverse
Reverse the order of the vector.
@ SK_ExtractSubvector
ExtractSubvector Index indicates start offset.
CastContextHint
Represents a hint about the context in which a cast is used.
CacheLevel
The possible cache levels.
This class represents a truncation of integer types.
static constexpr TypeSize getFixed(ScalarTy ExactSize)
Definition: TypeSize.h:342
The instances of the Type class are immutable: once they are created, they are never changed.
Definition: Type.h:45
bool isVectorTy() const
True if this is an instance of VectorType.
Definition: Type.h:265
unsigned getPointerAddressSpace() const
Get the address space of this pointer or pointer vector type.
static IntegerType * getIntNTy(LLVMContext &C, unsigned N)
unsigned getScalarSizeInBits() const LLVM_READONLY
If this is a vector type, return the getPrimitiveSizeInBits value for the element type.
static IntegerType * getInt8Ty(LLVMContext &C)
bool isFloatingPointTy() const
Return true if this is one of the floating-point types.
Definition: Type.h:185
bool isScalableTy() const
Return true if this is a type whose size is a known multiple of vscale.
Type * getScalarType() const
If this is a vector type, return the element type, otherwise return 'this'.
Definition: Type.h:348
This is the common base class for vector predication intrinsics.
LLVM Value Representation.
Definition: Value.h:74
Type * getType() const
All values are typed, get the type of this value.
Definition: Value.h:255
Base class of all SIMD vector types.
Definition: DerivedTypes.h:403
constexpr ScalarTy getFixedValue() const
Definition: TypeSize.h:199
constexpr bool isScalable() const
Returns whether the quantity is scaled by a runtime quantity (vscale).
Definition: TypeSize.h:171
#define llvm_unreachable(msg)
Marks that the current location is not supposed to be reachable.
@ Fast
Attempts to make calls as fast as possible (e.g.
Definition: CallingConv.h:41
@ C
The default llvm calling convention, compatible with C.
Definition: CallingConv.h:34
bool match(Val *V, const Pattern &P)
Definition: PatternMatch.h:49
auto m_LogicalOr()
Matches L || R where L and R are arbitrary values.
class_match< Value > m_Value()
Match an arbitrary value and ignore it.
Definition: PatternMatch.h:92
auto m_LogicalAnd()
Matches L && R where L and R are arbitrary values.
match_combine_or< LTy, RTy > m_CombineOr(const LTy &L, const RTy &R)
Combine two pattern matchers matching L || R.
Definition: PatternMatch.h:239
This is an optimization pass for GlobalISel generic memory operations.
Definition: AddressRanges.h:18
@ Length
Definition: DWP.cpp:456
auto enumerate(FirstRange &&First, RestRanges &&...Rest)
Given two or more input ranges, returns a new range whose values are are tuples (A,...
Definition: STLExtras.h:2406
AddressSpace
Definition: NVPTXBaseInfo.h:21
Value * getSplatValue(const Value *V)
Get splat value if the input is a splat vector or return nullptr.
bool any_of(R &&range, UnaryPredicate P)
Provide wrappers to std::any_of which take ranges instead of having to pass begin/end explicitly.
Definition: STLExtras.h:1729
constexpr bool isPowerOf2_32(uint32_t Value)
Return true if the argument is a power of two > 0.
Definition: MathExtras.h:275
constexpr int PoisonMaskElem
constexpr unsigned BitWidth
Definition: BitmaskEnum.h:191
gep_type_iterator gep_type_begin(const User *GEP)
InstructionCost Cost
@ DataWithoutLaneMask
Same as Data, but avoids using the get.active.lane.mask intrinsic to calculate the mask and instead i...
This struct is a compact representation of a valid (non-zero power of two) alignment.
Definition: Alignment.h:39
Attributes of a target dependent hardware loop.
Information about a load/store intrinsic defined by the target.
Returns options for expansion of memcmp. IsZeroCmp is.
Describe known properties for a set of pointers.
Flags describing the kind of vector reduction.
Parameters that control the generic loop unrolling transformation.