Line data Source code
1 : //===- TypeMetadataUtils.cpp - Utilities related to type metadata ---------===//
2 : //
3 : // The LLVM Compiler Infrastructure
4 : //
5 : // This file is distributed under the University of Illinois Open Source
6 : // License. See LICENSE.TXT for details.
7 : //
8 : //===----------------------------------------------------------------------===//
9 : //
10 : // This file contains functions that make it easier to manipulate type metadata
11 : // for devirtualization.
12 : //
13 : //===----------------------------------------------------------------------===//
14 :
15 : #include "llvm/Analysis/TypeMetadataUtils.h"
16 : #include "llvm/IR/Constants.h"
17 : #include "llvm/IR/Dominators.h"
18 : #include "llvm/IR/Intrinsics.h"
19 : #include "llvm/IR/Module.h"
20 :
21 : using namespace llvm;
22 :
23 : // Search for virtual calls that call FPtr and add them to DevirtCalls.
24 : static void
25 280 : findCallsAtConstantOffset(SmallVectorImpl<DevirtCallSite> &DevirtCalls,
26 : bool *HasNonCallUses, Value *FPtr, uint64_t Offset,
27 : const CallInst *CI, DominatorTree &DT) {
28 568 : for (const Use &U : FPtr->uses()) {
29 288 : Instruction *User = cast<Instruction>(U.getUser());
30 : // Ignore this instruction if it is not dominated by the type intrinsic
31 : // being analyzed. Otherwise we may transform a call sharing the same
32 : // vtable pointer incorrectly. Specifically, this situation can arise
33 : // after indirect call promotion and inlining, where we may have uses
34 : // of the vtable pointer guarded by a function pointer check, and a fallback
35 : // indirect call.
36 288 : if (!DT.dominates(CI, User))
37 : continue;
38 278 : if (isa<BitCastInst>(User)) {
39 131 : findCallsAtConstantOffset(DevirtCalls, HasNonCallUses, User, Offset, CI,
40 : DT);
41 : } else if (auto CI = dyn_cast<CallInst>(User)) {
42 284 : DevirtCalls.push_back({Offset, CI});
43 : } else if (auto II = dyn_cast<InvokeInst>(User)) {
44 2 : DevirtCalls.push_back({Offset, II});
45 4 : } else if (HasNonCallUses) {
46 0 : *HasNonCallUses = true;
47 : }
48 : }
49 280 : }
50 :
51 : // Search for virtual calls that load from VPtr and add them to DevirtCalls.
52 319 : static void findLoadCallsAtConstantOffset(
53 : const Module *M, SmallVectorImpl<DevirtCallSite> &DevirtCalls, Value *VPtr,
54 : int64_t Offset, const CallInst *CI, DominatorTree &DT) {
55 753 : for (const Use &U : VPtr->uses()) {
56 434 : Value *User = U.getUser();
57 : if (isa<BitCastInst>(User)) {
58 120 : findLoadCallsAtConstantOffset(M, DevirtCalls, User, Offset, CI, DT);
59 : } else if (isa<LoadInst>(User)) {
60 107 : findCallsAtConstantOffset(DevirtCalls, nullptr, User, Offset, CI, DT);
61 : } else if (auto GEP = dyn_cast<GetElementPtrInst>(User)) {
62 : // Take into account the GEP offset.
63 99 : if (VPtr == GEP->getPointerOperand() && GEP->hasAllConstantIndices()) {
64 99 : SmallVector<Value *, 8> Indices(GEP->op_begin() + 1, GEP->op_end());
65 99 : int64_t GEPOffset = M->getDataLayout().getIndexedOffsetInType(
66 : GEP->getSourceElementType(), Indices);
67 99 : findLoadCallsAtConstantOffset(M, DevirtCalls, User, Offset + GEPOffset,
68 : CI, DT);
69 : }
70 : }
71 : }
72 319 : }
73 :
74 137 : void llvm::findDevirtualizableCallsForTypeTest(
75 : SmallVectorImpl<DevirtCallSite> &DevirtCalls,
76 : SmallVectorImpl<CallInst *> &Assumes, const CallInst *CI,
77 : DominatorTree &DT) {
78 : assert(CI->getCalledFunction()->getIntrinsicID() == Intrinsic::type_test);
79 :
80 137 : const Module *M = CI->getParent()->getParent()->getParent();
81 :
82 : // Find llvm.assume intrinsics for this llvm.type.test call.
83 272 : for (const Use &CIU : CI->uses()) {
84 270 : if (auto *AssumeCI = dyn_cast<CallInst>(CIU.getUser())) {
85 : Function *F = AssumeCI->getCalledFunction();
86 100 : if (F && F->getIntrinsicID() == Intrinsic::assume)
87 100 : Assumes.push_back(AssumeCI);
88 : }
89 : }
90 :
91 : // If we found any, search for virtual calls based on %p and add them to
92 : // DevirtCalls.
93 137 : if (!Assumes.empty())
94 200 : findLoadCallsAtConstantOffset(
95 : M, DevirtCalls, CI->getArgOperand(0)->stripPointerCasts(), 0, CI, DT);
96 137 : }
97 :
98 46 : void llvm::findDevirtualizableCallsForTypeCheckedLoad(
99 : SmallVectorImpl<DevirtCallSite> &DevirtCalls,
100 : SmallVectorImpl<Instruction *> &LoadedPtrs,
101 : SmallVectorImpl<Instruction *> &Preds, bool &HasNonCallUses,
102 : const CallInst *CI, DominatorTree &DT) {
103 : assert(CI->getCalledFunction()->getIntrinsicID() ==
104 : Intrinsic::type_checked_load);
105 :
106 46 : auto *Offset = dyn_cast<ConstantInt>(CI->getArgOperand(1));
107 : if (!Offset) {
108 0 : HasNonCallUses = true;
109 0 : return;
110 : }
111 :
112 130 : for (const Use &U : CI->uses()) {
113 84 : auto CIU = U.getUser();
114 : if (auto EVI = dyn_cast<ExtractValueInst>(CIU)) {
115 81 : if (EVI->getNumIndices() == 1 && EVI->getIndices()[0] == 0) {
116 42 : LoadedPtrs.push_back(EVI);
117 42 : continue;
118 : }
119 39 : if (EVI->getNumIndices() == 1 && EVI->getIndices()[0] == 1) {
120 39 : Preds.push_back(EVI);
121 39 : continue;
122 : }
123 : }
124 3 : HasNonCallUses = true;
125 : }
126 :
127 88 : for (Value *LoadedPtr : LoadedPtrs)
128 42 : findCallsAtConstantOffset(DevirtCalls, &HasNonCallUses, LoadedPtr,
129 : Offset->getZExtValue(), CI, DT);
130 : }
|