| //===- CodeMetrics.cpp - Code cost measurements ---------------------------===// |
| // |
| // The LLVM Compiler Infrastructure |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| // This file implements code cost measurement utilities. |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #include "llvm/Analysis/CodeMetrics.h" |
| #include "llvm/Analysis/AssumptionCache.h" |
| #include "llvm/Analysis/LoopInfo.h" |
| #include "llvm/Analysis/TargetTransformInfo.h" |
| #include "llvm/Analysis/ValueTracking.h" |
| #include "llvm/IR/CallSite.h" |
| #include "llvm/IR/DataLayout.h" |
| #include "llvm/IR/Function.h" |
| #include "llvm/Support/Debug.h" |
| #include "llvm/Support/raw_ostream.h" |
| |
| #define DEBUG_TYPE "code-metrics" |
| |
| using namespace llvm; |
| |
| static void |
| appendSpeculatableOperands(const Value *V, |
| SmallPtrSetImpl<const Value *> &Visited, |
| SmallVectorImpl<const Value *> &Worklist) { |
| const User *U = dyn_cast<User>(V); |
| if (!U) |
| return; |
| |
| for (const Value *Operand : U->operands()) |
| if (Visited.insert(Operand).second) |
| if (isSafeToSpeculativelyExecute(Operand)) |
| Worklist.push_back(Operand); |
| } |
| |
| static void completeEphemeralValues(SmallPtrSetImpl<const Value *> &Visited, |
| SmallVectorImpl<const Value *> &Worklist, |
| SmallPtrSetImpl<const Value *> &EphValues) { |
| // Note: We don't speculate PHIs here, so we'll miss instruction chains kept |
| // alive only by ephemeral values. |
| |
| // Walk the worklist using an index but without caching the size so we can |
| // append more entries as we process the worklist. This forms a queue without |
| // quadratic behavior by just leaving processed nodes at the head of the |
| // worklist forever. |
| for (int i = 0; i < (int)Worklist.size(); ++i) { |
| const Value *V = Worklist[i]; |
| |
| assert(Visited.count(V) && |
| "Failed to add a worklist entry to our visited set!"); |
| |
| // If all uses of this value are ephemeral, then so is this value. |
| if (!all_of(V->users(), [&](const User *U) { return EphValues.count(U); })) |
| continue; |
| |
| EphValues.insert(V); |
| LLVM_DEBUG(dbgs() << "Ephemeral Value: " << *V << "\n"); |
| |
| // Append any more operands to consider. |
| appendSpeculatableOperands(V, Visited, Worklist); |
| } |
| } |
| |
| // Find all ephemeral values. |
| void CodeMetrics::collectEphemeralValues( |
| const Loop *L, AssumptionCache *AC, |
| SmallPtrSetImpl<const Value *> &EphValues) { |
| SmallPtrSet<const Value *, 32> Visited; |
| SmallVector<const Value *, 16> Worklist; |
| |
| for (auto &AssumeVH : AC->assumptions()) { |
| if (!AssumeVH) |
| continue; |
| Instruction *I = cast<Instruction>(AssumeVH); |
| |
| // Filter out call sites outside of the loop so we don't do a function's |
| // worth of work for each of its loops (and, in the common case, ephemeral |
| // values in the loop are likely due to @llvm.assume calls in the loop). |
| if (!L->contains(I->getParent())) |
| continue; |
| |
| if (EphValues.insert(I).second) |
| appendSpeculatableOperands(I, Visited, Worklist); |
| } |
| |
| completeEphemeralValues(Visited, Worklist, EphValues); |
| } |
| |
| void CodeMetrics::collectEphemeralValues( |
| const Function *F, AssumptionCache *AC, |
| SmallPtrSetImpl<const Value *> &EphValues) { |
| SmallPtrSet<const Value *, 32> Visited; |
| SmallVector<const Value *, 16> Worklist; |
| |
| for (auto &AssumeVH : AC->assumptions()) { |
| if (!AssumeVH) |
| continue; |
| Instruction *I = cast<Instruction>(AssumeVH); |
| assert(I->getParent()->getParent() == F && |
| "Found assumption for the wrong function!"); |
| |
| if (EphValues.insert(I).second) |
| appendSpeculatableOperands(I, Visited, Worklist); |
| } |
| |
| completeEphemeralValues(Visited, Worklist, EphValues); |
| } |
| |
| /// Fill in the current structure with information gleaned from the specified |
| /// block. |
| void CodeMetrics::analyzeBasicBlock(const BasicBlock *BB, |
| const TargetTransformInfo &TTI, |
| const SmallPtrSetImpl<const Value*> &EphValues) { |
| ++NumBlocks; |
| unsigned NumInstsBeforeThisBB = NumInsts; |
| for (const Instruction &I : *BB) { |
| // Skip ephemeral values. |
| if (EphValues.count(&I)) |
| continue; |
| |
| // Special handling for calls. |
| if (isa<CallInst>(I) || isa<InvokeInst>(I)) { |
| ImmutableCallSite CS(&I); |
| |
| if (const Function *F = CS.getCalledFunction()) { |
| // If a function is both internal and has a single use, then it is |
| // extremely likely to get inlined in the future (it was probably |
| // exposed by an interleaved devirtualization pass). |
| if (!CS.isNoInline() && F->hasInternalLinkage() && F->hasOneUse()) |
| ++NumInlineCandidates; |
| |
| // If this call is to function itself, then the function is recursive. |
| // Inlining it into other functions is a bad idea, because this is |
| // basically just a form of loop peeling, and our metrics aren't useful |
| // for that case. |
| if (F == BB->getParent()) |
| isRecursive = true; |
| |
| if (TTI.isLoweredToCall(F)) |
| ++NumCalls; |
| } else { |
| // We don't want inline asm to count as a call - that would prevent loop |
| // unrolling. The argument setup cost is still real, though. |
| if (!isa<InlineAsm>(CS.getCalledValue())) |
| ++NumCalls; |
| } |
| } |
| |
| if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) { |
| if (!AI->isStaticAlloca()) |
| this->usesDynamicAlloca = true; |
| } |
| |
| if (isa<ExtractElementInst>(I) || I.getType()->isVectorTy()) |
| ++NumVectorInsts; |
| |
| if (I.getType()->isTokenTy() && I.isUsedOutsideOfBlock(BB)) |
| notDuplicatable = true; |
| |
| if (const CallInst *CI = dyn_cast<CallInst>(&I)) { |
| if (CI->cannotDuplicate()) |
| notDuplicatable = true; |
| if (CI->isConvergent()) |
| convergent = true; |
| } |
| |
| if (const InvokeInst *InvI = dyn_cast<InvokeInst>(&I)) |
| if (InvI->cannotDuplicate()) |
| notDuplicatable = true; |
| |
| NumInsts += TTI.getUserCost(&I); |
| } |
| |
| if (isa<ReturnInst>(BB->getTerminator())) |
| ++NumRets; |
| |
| // We never want to inline functions that contain an indirectbr. This is |
| // incorrect because all the blockaddress's (in static global initializers |
| // for example) would be referring to the original function, and this indirect |
| // jump would jump from the inlined copy of the function into the original |
| // function which is extremely undefined behavior. |
| // FIXME: This logic isn't really right; we can safely inline functions |
| // with indirectbr's as long as no other function or global references the |
| // blockaddress of a block within the current function. And as a QOI issue, |
| // if someone is using a blockaddress without an indirectbr, and that |
| // reference somehow ends up in another function or global, we probably |
| // don't want to inline this function. |
| notDuplicatable |= isa<IndirectBrInst>(BB->getTerminator()); |
| |
| // Remember NumInsts for this BB. |
| NumBBInsts[BB] = NumInsts - NumInstsBeforeThisBB; |
| } |