| //===-- AutoUpgrade.cpp - Implement auto-upgrade helper functions ---------===// |
| // |
| // The LLVM Compiler Infrastructure |
| // |
| // This file is distributed under the University of Illinois Open Source |
| // License. See LICENSE.TXT for details. |
| // |
| //===----------------------------------------------------------------------===// |
| // |
| // This file implements the auto-upgrade helper functions |
| // |
| //===----------------------------------------------------------------------===// |
| |
| #include "llvm/AutoUpgrade.h" |
| #include "llvm/Constants.h" |
| #include "llvm/Function.h" |
| #include "llvm/Instruction.h" |
| #include "llvm/LLVMContext.h" |
| #include "llvm/Module.h" |
| #include "llvm/IntrinsicInst.h" |
| #include "llvm/ADT/DenseMap.h" |
| #include "llvm/ADT/SmallPtrSet.h" |
| #include "llvm/ADT/SmallVector.h" |
| #include "llvm/Support/CallSite.h" |
| #include "llvm/Support/CFG.h" |
| #include "llvm/Support/ErrorHandling.h" |
| #include "llvm/Support/IRBuilder.h" |
| #include <cstring> |
| using namespace llvm; |
| |
| |
| static bool UpgradeIntrinsicFunction1(Function *F, Function *&NewFn) { |
| assert(F && "Illegal to upgrade a non-existent Function."); |
| |
| // Quickly eliminate it, if it's not a candidate. |
| StringRef Name = F->getName(); |
| if (Name.size() <= 8 || !Name.startswith("llvm.")) |
| return false; |
| Name = Name.substr(5); // Strip off "llvm." |
| |
| FunctionType *FTy = F->getFunctionType(); |
| Module *M = F->getParent(); |
| |
| switch (Name[0]) { |
| default: break; |
| case 'a': |
| if (Name.startswith("atomic.cmp.swap") || |
| Name.startswith("atomic.swap") || |
| Name.startswith("atomic.load.add") || |
| Name.startswith("atomic.load.sub") || |
| Name.startswith("atomic.load.and") || |
| Name.startswith("atomic.load.nand") || |
| Name.startswith("atomic.load.or") || |
| Name.startswith("atomic.load.xor") || |
| Name.startswith("atomic.load.max") || |
| Name.startswith("atomic.load.min") || |
| Name.startswith("atomic.load.umax") || |
| Name.startswith("atomic.load.umin")) |
| return true; |
| case 'i': |
| // This upgrades the old llvm.init.trampoline to the new |
| // llvm.init.trampoline and llvm.adjust.trampoline pair. |
| if (Name == "init.trampoline") { |
| // The new llvm.init.trampoline returns nothing. |
| if (FTy->getReturnType()->isVoidTy()) |
| break; |
| |
| assert(FTy->getNumParams() == 3 && "old init.trampoline takes 3 args!"); |
| |
| // Change the name of the old intrinsic so that we can play with its type. |
| std::string NameTmp = F->getName(); |
| F->setName(""); |
| NewFn = cast<Function>(M->getOrInsertFunction( |
| NameTmp, |
| Type::getVoidTy(M->getContext()), |
| FTy->getParamType(0), FTy->getParamType(1), |
| FTy->getParamType(2), (Type *)0)); |
| return true; |
| } |
| case 'm': |
| if (Name == "memory.barrier") |
| return true; |
| case 'p': |
| // This upgrades the llvm.prefetch intrinsic to accept one more parameter, |
| // which is a instruction / data cache identifier. The old version only |
| // implicitly accepted the data version. |
| if (Name == "prefetch") { |
| // Don't do anything if it has the correct number of arguments already |
| if (FTy->getNumParams() == 4) |
| break; |
| |
| assert(FTy->getNumParams() == 3 && "old prefetch takes 3 args!"); |
| // We first need to change the name of the old (bad) intrinsic, because |
| // its type is incorrect, but we cannot overload that name. We |
| // arbitrarily unique it here allowing us to construct a correctly named |
| // and typed function below. |
| std::string NameTmp = F->getName(); |
| F->setName(""); |
| NewFn = cast<Function>(M->getOrInsertFunction(NameTmp, |
| FTy->getReturnType(), |
| FTy->getParamType(0), |
| FTy->getParamType(1), |
| FTy->getParamType(2), |
| FTy->getParamType(2), |
| (Type*)0)); |
| return true; |
| } |
| |
| break; |
| case 'x': { |
| const char *NewFnName = NULL; |
| // This fixes the poorly named crc32 intrinsics. |
| if (Name == "x86.sse42.crc32.8") |
| NewFnName = "llvm.x86.sse42.crc32.32.8"; |
| else if (Name == "x86.sse42.crc32.16") |
| NewFnName = "llvm.x86.sse42.crc32.32.16"; |
| else if (Name == "x86.sse42.crc32.32") |
| NewFnName = "llvm.x86.sse42.crc32.32.32"; |
| else if (Name == "x86.sse42.crc64.8") |
| NewFnName = "llvm.x86.sse42.crc32.64.8"; |
| else if (Name == "x86.sse42.crc64.64") |
| NewFnName = "llvm.x86.sse42.crc32.64.64"; |
| |
| if (NewFnName) { |
| F->setName(NewFnName); |
| NewFn = F; |
| return true; |
| } |
| |
| // Calls to these instructions are transformed into unaligned loads. |
| if (Name == "x86.sse.loadu.ps" || Name == "x86.sse2.loadu.dq" || |
| Name == "x86.sse2.loadu.pd") |
| return true; |
| |
| // Calls to these instructions are transformed into nontemporal stores. |
| if (Name == "x86.sse.movnt.ps" || Name == "x86.sse2.movnt.dq" || |
| Name == "x86.sse2.movnt.pd" || Name == "x86.sse2.movnt.i") |
| return true; |
| |
| break; |
| } |
| } |
| |
| // This may not belong here. This function is effectively being overloaded |
| // to both detect an intrinsic which needs upgrading, and to provide the |
| // upgraded form of the intrinsic. We should perhaps have two separate |
| // functions for this. |
| return false; |
| } |
| |
| bool llvm::UpgradeIntrinsicFunction(Function *F, Function *&NewFn) { |
| NewFn = 0; |
| bool Upgraded = UpgradeIntrinsicFunction1(F, NewFn); |
| |
| // Upgrade intrinsic attributes. This does not change the function. |
| if (NewFn) |
| F = NewFn; |
| if (unsigned id = F->getIntrinsicID()) |
| F->setAttributes(Intrinsic::getAttributes((Intrinsic::ID)id)); |
| return Upgraded; |
| } |
| |
| bool llvm::UpgradeGlobalVariable(GlobalVariable *GV) { |
| // Nothing to do yet. |
| return false; |
| } |
| |
| // UpgradeIntrinsicCall - Upgrade a call to an old intrinsic to be a call the |
| // upgraded intrinsic. All argument and return casting must be provided in |
| // order to seamlessly integrate with existing context. |
| void llvm::UpgradeIntrinsicCall(CallInst *CI, Function *NewFn) { |
| Function *F = CI->getCalledFunction(); |
| LLVMContext &C = CI->getContext(); |
| ImmutableCallSite CS(CI); |
| |
| assert(F && "CallInst has no function associated with it."); |
| |
| if (!NewFn) { |
| if (F->getName() == "llvm.x86.sse.loadu.ps" || |
| F->getName() == "llvm.x86.sse2.loadu.dq" || |
| F->getName() == "llvm.x86.sse2.loadu.pd") { |
| // Convert to a native, unaligned load. |
| Type *VecTy = CI->getType(); |
| Type *IntTy = IntegerType::get(C, 128); |
| IRBuilder<> Builder(C); |
| Builder.SetInsertPoint(CI->getParent(), CI); |
| |
| Value *BC = Builder.CreateBitCast(CI->getArgOperand(0), |
| PointerType::getUnqual(IntTy), |
| "cast"); |
| LoadInst *LI = Builder.CreateLoad(BC, CI->getName()); |
| LI->setAlignment(1); // Unaligned load. |
| BC = Builder.CreateBitCast(LI, VecTy, "new.cast"); |
| |
| // Fix up all the uses with our new load. |
| if (!CI->use_empty()) |
| CI->replaceAllUsesWith(BC); |
| |
| // Remove intrinsic. |
| CI->eraseFromParent(); |
| } else if (F->getName() == "llvm.x86.sse.movnt.ps" || |
| F->getName() == "llvm.x86.sse2.movnt.dq" || |
| F->getName() == "llvm.x86.sse2.movnt.pd" || |
| F->getName() == "llvm.x86.sse2.movnt.i") { |
| IRBuilder<> Builder(C); |
| Builder.SetInsertPoint(CI->getParent(), CI); |
| |
| Module *M = F->getParent(); |
| SmallVector<Value *, 1> Elts; |
| Elts.push_back(ConstantInt::get(Type::getInt32Ty(C), 1)); |
| MDNode *Node = MDNode::get(C, Elts); |
| |
| Value *Arg0 = CI->getArgOperand(0); |
| Value *Arg1 = CI->getArgOperand(1); |
| |
| // Convert the type of the pointer to a pointer to the stored type. |
| Value *BC = Builder.CreateBitCast(Arg0, |
| PointerType::getUnqual(Arg1->getType()), |
| "cast"); |
| StoreInst *SI = Builder.CreateStore(Arg1, BC); |
| SI->setMetadata(M->getMDKindID("nontemporal"), Node); |
| SI->setAlignment(16); |
| |
| // Remove intrinsic. |
| CI->eraseFromParent(); |
| } else if (F->getName().startswith("llvm.atomic.cmp.swap")) { |
| IRBuilder<> Builder(C); |
| Builder.SetInsertPoint(CI->getParent(), CI); |
| Value *Val = Builder.CreateAtomicCmpXchg(CI->getArgOperand(0), |
| CI->getArgOperand(1), |
| CI->getArgOperand(2), |
| Monotonic); |
| |
| // Replace intrinsic. |
| Val->takeName(CI); |
| if (!CI->use_empty()) |
| CI->replaceAllUsesWith(Val); |
| CI->eraseFromParent(); |
| } else if (F->getName().startswith("llvm.atomic")) { |
| IRBuilder<> Builder(C); |
| Builder.SetInsertPoint(CI->getParent(), CI); |
| |
| AtomicRMWInst::BinOp Op; |
| if (F->getName().startswith("llvm.atomic.swap")) |
| Op = AtomicRMWInst::Xchg; |
| else if (F->getName().startswith("llvm.atomic.load.add")) |
| Op = AtomicRMWInst::Add; |
| else if (F->getName().startswith("llvm.atomic.load.sub")) |
| Op = AtomicRMWInst::Sub; |
| else if (F->getName().startswith("llvm.atomic.load.and")) |
| Op = AtomicRMWInst::And; |
| else if (F->getName().startswith("llvm.atomic.load.nand")) |
| Op = AtomicRMWInst::Nand; |
| else if (F->getName().startswith("llvm.atomic.load.or")) |
| Op = AtomicRMWInst::Or; |
| else if (F->getName().startswith("llvm.atomic.load.xor")) |
| Op = AtomicRMWInst::Xor; |
| else if (F->getName().startswith("llvm.atomic.load.max")) |
| Op = AtomicRMWInst::Max; |
| else if (F->getName().startswith("llvm.atomic.load.min")) |
| Op = AtomicRMWInst::Min; |
| else if (F->getName().startswith("llvm.atomic.load.umax")) |
| Op = AtomicRMWInst::UMax; |
| else if (F->getName().startswith("llvm.atomic.load.umin")) |
| Op = AtomicRMWInst::UMin; |
| else |
| llvm_unreachable("Unknown atomic"); |
| |
| Value *Val = Builder.CreateAtomicRMW(Op, CI->getArgOperand(0), |
| CI->getArgOperand(1), |
| Monotonic); |
| |
| // Replace intrinsic. |
| Val->takeName(CI); |
| if (!CI->use_empty()) |
| CI->replaceAllUsesWith(Val); |
| CI->eraseFromParent(); |
| } else if (F->getName() == "llvm.memory.barrier") { |
| IRBuilder<> Builder(C); |
| Builder.SetInsertPoint(CI->getParent(), CI); |
| |
| // Note that this conversion ignores the "device" bit; it was not really |
| // well-defined, and got abused because nobody paid enough attention to |
| // get it right. In practice, this probably doesn't matter; application |
| // code generally doesn't need anything stronger than |
| // SequentiallyConsistent (and realistically, SequentiallyConsistent |
| // is lowered to a strong enough barrier for almost anything). |
| |
| if (cast<ConstantInt>(CI->getArgOperand(1))->getZExtValue()) |
| Builder.CreateFence(SequentiallyConsistent); |
| else if (!cast<ConstantInt>(CI->getArgOperand(0))->getZExtValue()) |
| Builder.CreateFence(Release); |
| else if (!cast<ConstantInt>(CI->getArgOperand(3))->getZExtValue()) |
| Builder.CreateFence(Acquire); |
| else |
| Builder.CreateFence(AcquireRelease); |
| |
| // Remove intrinsic. |
| CI->eraseFromParent(); |
| } else { |
| llvm_unreachable("Unknown function for CallInst upgrade."); |
| } |
| return; |
| } |
| |
| switch (NewFn->getIntrinsicID()) { |
| case Intrinsic::prefetch: { |
| IRBuilder<> Builder(C); |
| Builder.SetInsertPoint(CI->getParent(), CI); |
| llvm::Type *I32Ty = llvm::Type::getInt32Ty(CI->getContext()); |
| |
| // Add the extra "data cache" argument |
| Value *Operands[4] = { CI->getArgOperand(0), CI->getArgOperand(1), |
| CI->getArgOperand(2), |
| llvm::ConstantInt::get(I32Ty, 1) }; |
| CallInst *NewCI = CallInst::Create(NewFn, Operands, |
| CI->getName(), CI); |
| NewCI->setTailCall(CI->isTailCall()); |
| NewCI->setCallingConv(CI->getCallingConv()); |
| // Handle any uses of the old CallInst. |
| if (!CI->use_empty()) |
| // Replace all uses of the old call with the new cast which has the |
| // correct type. |
| CI->replaceAllUsesWith(NewCI); |
| |
| // Clean up the old call now that it has been completely upgraded. |
| CI->eraseFromParent(); |
| break; |
| } |
| case Intrinsic::init_trampoline: { |
| |
| // Transform |
| // %tramp = call i8* llvm.init.trampoline (i8* x, i8* y, i8* z) |
| // to |
| // call void llvm.init.trampoline (i8* %x, i8* %y, i8* %z) |
| // %tramp = call i8* llvm.adjust.trampoline (i8* %x) |
| |
| Function *AdjustTrampolineFn = |
| cast<Function>(Intrinsic::getDeclaration(F->getParent(), |
| Intrinsic::adjust_trampoline)); |
| |
| IRBuilder<> Builder(C); |
| Builder.SetInsertPoint(CI); |
| |
| Builder.CreateCall3(NewFn, CI->getArgOperand(0), CI->getArgOperand(1), |
| CI->getArgOperand(2)); |
| |
| CallInst *AdjustCall = Builder.CreateCall(AdjustTrampolineFn, |
| CI->getArgOperand(0), |
| CI->getName()); |
| if (!CI->use_empty()) |
| CI->replaceAllUsesWith(AdjustCall); |
| CI->eraseFromParent(); |
| break; |
| } |
| } |
| } |
| |
| // This tests each Function to determine if it needs upgrading. When we find |
| // one we are interested in, we then upgrade all calls to reflect the new |
| // function. |
| void llvm::UpgradeCallsToIntrinsic(Function* F) { |
| assert(F && "Illegal attempt to upgrade a non-existent intrinsic."); |
| |
| // Upgrade the function and check if it is a totaly new function. |
| Function *NewFn; |
| if (UpgradeIntrinsicFunction(F, NewFn)) { |
| if (NewFn != F) { |
| // Replace all uses to the old function with the new one if necessary. |
| for (Value::use_iterator UI = F->use_begin(), UE = F->use_end(); |
| UI != UE; ) { |
| if (CallInst *CI = dyn_cast<CallInst>(*UI++)) |
| UpgradeIntrinsicCall(CI, NewFn); |
| } |
| // Remove old function, no longer used, from the module. |
| F->eraseFromParent(); |
| } |
| } |
| } |
| |
| /// This function strips all debug info intrinsics, except for llvm.dbg.declare. |
| /// If an llvm.dbg.declare intrinsic is invalid, then this function simply |
| /// strips that use. |
| void llvm::CheckDebugInfoIntrinsics(Module *M) { |
| if (Function *FuncStart = M->getFunction("llvm.dbg.func.start")) { |
| while (!FuncStart->use_empty()) |
| cast<CallInst>(FuncStart->use_back())->eraseFromParent(); |
| FuncStart->eraseFromParent(); |
| } |
| |
| if (Function *StopPoint = M->getFunction("llvm.dbg.stoppoint")) { |
| while (!StopPoint->use_empty()) |
| cast<CallInst>(StopPoint->use_back())->eraseFromParent(); |
| StopPoint->eraseFromParent(); |
| } |
| |
| if (Function *RegionStart = M->getFunction("llvm.dbg.region.start")) { |
| while (!RegionStart->use_empty()) |
| cast<CallInst>(RegionStart->use_back())->eraseFromParent(); |
| RegionStart->eraseFromParent(); |
| } |
| |
| if (Function *RegionEnd = M->getFunction("llvm.dbg.region.end")) { |
| while (!RegionEnd->use_empty()) |
| cast<CallInst>(RegionEnd->use_back())->eraseFromParent(); |
| RegionEnd->eraseFromParent(); |
| } |
| |
| if (Function *Declare = M->getFunction("llvm.dbg.declare")) { |
| if (!Declare->use_empty()) { |
| DbgDeclareInst *DDI = cast<DbgDeclareInst>(Declare->use_back()); |
| if (!isa<MDNode>(DDI->getArgOperand(0)) || |
| !isa<MDNode>(DDI->getArgOperand(1))) { |
| while (!Declare->use_empty()) { |
| CallInst *CI = cast<CallInst>(Declare->use_back()); |
| CI->eraseFromParent(); |
| } |
| Declare->eraseFromParent(); |
| } |
| } |
| } |
| } |
| |
| /// FindExnAndSelIntrinsics - Find the eh_exception and eh_selector intrinsic |
| /// calls reachable from the unwind basic block. |
| static void FindExnAndSelIntrinsics(BasicBlock *BB, CallInst *&Exn, |
| CallInst *&Sel, |
| SmallPtrSet<BasicBlock*, 8> &Visited) { |
| if (!Visited.insert(BB)) return; |
| |
| for (BasicBlock::iterator |
| I = BB->begin(), E = BB->end(); I != E; ++I) { |
| if (CallInst *CI = dyn_cast<CallInst>(I)) { |
| switch (CI->getCalledFunction()->getIntrinsicID()) { |
| default: break; |
| case Intrinsic::eh_exception: |
| assert(!Exn && "Found more than one eh.exception call!"); |
| Exn = CI; |
| break; |
| case Intrinsic::eh_selector: |
| assert(!Sel && "Found more than one eh.selector call!"); |
| Sel = CI; |
| break; |
| } |
| |
| if (Exn && Sel) return; |
| } |
| } |
| |
| if (Exn && Sel) return; |
| |
| for (succ_iterator I = succ_begin(BB), E = succ_end(BB); I != E; ++I) { |
| FindExnAndSelIntrinsics(*I, Exn, Sel, Visited); |
| if (Exn && Sel) return; |
| } |
| } |
| |
| /// TransferClausesToLandingPadInst - Transfer the exception handling clauses |
| /// from the eh_selector call to the new landingpad instruction. |
| static void TransferClausesToLandingPadInst(LandingPadInst *LPI, |
| CallInst *EHSel) { |
| LLVMContext &Context = LPI->getContext(); |
| unsigned N = EHSel->getNumArgOperands(); |
| |
| for (unsigned i = N - 1; i > 1; --i) { |
| if (const ConstantInt *CI = dyn_cast<ConstantInt>(EHSel->getArgOperand(i))){ |
| unsigned FilterLength = CI->getZExtValue(); |
| unsigned FirstCatch = i + FilterLength + !FilterLength; |
| assert(FirstCatch <= N && "Invalid filter length"); |
| |
| if (FirstCatch < N) |
| for (unsigned j = FirstCatch; j < N; ++j) { |
| Value *Val = EHSel->getArgOperand(j); |
| if (!Val->hasName() || Val->getName() != "llvm.eh.catch.all.value") { |
| LPI->addClause(EHSel->getArgOperand(j)); |
| } else { |
| GlobalVariable *GV = cast<GlobalVariable>(Val); |
| LPI->addClause(GV->getInitializer()); |
| } |
| } |
| |
| if (!FilterLength) { |
| // Cleanup. |
| LPI->setCleanup(true); |
| } else { |
| // Filter. |
| SmallVector<Constant *, 4> TyInfo; |
| TyInfo.reserve(FilterLength - 1); |
| for (unsigned j = i + 1; j < FirstCatch; ++j) |
| TyInfo.push_back(cast<Constant>(EHSel->getArgOperand(j))); |
| ArrayType *AType = |
| ArrayType::get(!TyInfo.empty() ? TyInfo[0]->getType() : |
| PointerType::getUnqual(Type::getInt8Ty(Context)), |
| TyInfo.size()); |
| LPI->addClause(ConstantArray::get(AType, TyInfo)); |
| } |
| |
| N = i; |
| } |
| } |
| |
| if (N > 2) |
| for (unsigned j = 2; j < N; ++j) { |
| Value *Val = EHSel->getArgOperand(j); |
| if (!Val->hasName() || Val->getName() != "llvm.eh.catch.all.value") { |
| LPI->addClause(EHSel->getArgOperand(j)); |
| } else { |
| GlobalVariable *GV = cast<GlobalVariable>(Val); |
| LPI->addClause(GV->getInitializer()); |
| } |
| } |
| } |
| |
| /// This function upgrades the old pre-3.0 exception handling system to the new |
| /// one. N.B. This will be removed in 3.1. |
| void llvm::UpgradeExceptionHandling(Module *M) { |
| Function *EHException = M->getFunction("llvm.eh.exception"); |
| Function *EHSelector = M->getFunction("llvm.eh.selector"); |
| if (!EHException || !EHSelector) |
| return; |
| |
| LLVMContext &Context = M->getContext(); |
| Type *ExnTy = PointerType::getUnqual(Type::getInt8Ty(Context)); |
| Type *SelTy = Type::getInt32Ty(Context); |
| Type *LPadSlotTy = StructType::get(ExnTy, SelTy, NULL); |
| |
| // This map links the invoke instruction with the eh.exception and eh.selector |
| // calls associated with it. |
| DenseMap<InvokeInst*, std::pair<Value*, Value*> > InvokeToIntrinsicsMap; |
| for (Module::iterator |
| I = M->begin(), E = M->end(); I != E; ++I) { |
| Function &F = *I; |
| |
| for (Function::iterator |
| II = F.begin(), IE = F.end(); II != IE; ++II) { |
| BasicBlock *BB = &*II; |
| InvokeInst *Inst = dyn_cast<InvokeInst>(BB->getTerminator()); |
| if (!Inst) continue; |
| BasicBlock *UnwindDest = Inst->getUnwindDest(); |
| if (UnwindDest->isLandingPad()) continue; // Already converted. |
| |
| SmallPtrSet<BasicBlock*, 8> Visited; |
| CallInst *Exn = 0; |
| CallInst *Sel = 0; |
| FindExnAndSelIntrinsics(UnwindDest, Exn, Sel, Visited); |
| assert(Exn && Sel && "Cannot find eh.exception and eh.selector calls!"); |
| InvokeToIntrinsicsMap[Inst] = std::make_pair(Exn, Sel); |
| } |
| } |
| |
| // This map stores the slots where the exception object and selector value are |
| // stored within a function. |
| DenseMap<Function*, std::pair<Value*, Value*> > FnToLPadSlotMap; |
| SmallPtrSet<Instruction*, 32> DeadInsts; |
| for (DenseMap<InvokeInst*, std::pair<Value*, Value*> >::iterator |
| I = InvokeToIntrinsicsMap.begin(), E = InvokeToIntrinsicsMap.end(); |
| I != E; ++I) { |
| InvokeInst *Invoke = I->first; |
| BasicBlock *UnwindDest = Invoke->getUnwindDest(); |
| Function *F = UnwindDest->getParent(); |
| std::pair<Value*, Value*> EHIntrinsics = I->second; |
| CallInst *Exn = cast<CallInst>(EHIntrinsics.first); |
| CallInst *Sel = cast<CallInst>(EHIntrinsics.second); |
| |
| // Store the exception object and selector value in the entry block. |
| Value *ExnSlot = 0; |
| Value *SelSlot = 0; |
| if (!FnToLPadSlotMap[F].first) { |
| BasicBlock *Entry = &F->front(); |
| ExnSlot = new AllocaInst(ExnTy, "exn", Entry->getTerminator()); |
| SelSlot = new AllocaInst(SelTy, "sel", Entry->getTerminator()); |
| FnToLPadSlotMap[F] = std::make_pair(ExnSlot, SelSlot); |
| } else { |
| ExnSlot = FnToLPadSlotMap[F].first; |
| SelSlot = FnToLPadSlotMap[F].second; |
| } |
| |
| if (!UnwindDest->getSinglePredecessor()) { |
| // The unwind destination doesn't have a single predecessor. Create an |
| // unwind destination which has only one predecessor. |
| BasicBlock *NewBB = BasicBlock::Create(Context, "new.lpad", |
| UnwindDest->getParent()); |
| BranchInst::Create(UnwindDest, NewBB); |
| Invoke->setUnwindDest(NewBB); |
| |
| // Fix up any PHIs in the original unwind destination block. |
| for (BasicBlock::iterator |
| II = UnwindDest->begin(); isa<PHINode>(II); ++II) { |
| PHINode *PN = cast<PHINode>(II); |
| int Idx = PN->getBasicBlockIndex(Invoke->getParent()); |
| if (Idx == -1) continue; |
| PN->setIncomingBlock(Idx, NewBB); |
| } |
| |
| UnwindDest = NewBB; |
| } |
| |
| IRBuilder<> Builder(Context); |
| Builder.SetInsertPoint(UnwindDest, UnwindDest->getFirstInsertionPt()); |
| |
| Value *PersFn = Sel->getArgOperand(1); |
| LandingPadInst *LPI = Builder.CreateLandingPad(LPadSlotTy, PersFn, 0); |
| Value *LPExn = Builder.CreateExtractValue(LPI, 0); |
| Value *LPSel = Builder.CreateExtractValue(LPI, 1); |
| Builder.CreateStore(LPExn, ExnSlot); |
| Builder.CreateStore(LPSel, SelSlot); |
| |
| TransferClausesToLandingPadInst(LPI, Sel); |
| |
| DeadInsts.insert(Exn); |
| DeadInsts.insert(Sel); |
| } |
| |
| // Replace the old intrinsic calls with the values from the landingpad |
| // instruction(s). These values were stored in allocas for us to use here. |
| for (DenseMap<InvokeInst*, std::pair<Value*, Value*> >::iterator |
| I = InvokeToIntrinsicsMap.begin(), E = InvokeToIntrinsicsMap.end(); |
| I != E; ++I) { |
| std::pair<Value*, Value*> EHIntrinsics = I->second; |
| CallInst *Exn = cast<CallInst>(EHIntrinsics.first); |
| CallInst *Sel = cast<CallInst>(EHIntrinsics.second); |
| BasicBlock *Parent = Exn->getParent(); |
| |
| std::pair<Value*,Value*> ExnSelSlots = FnToLPadSlotMap[Parent->getParent()]; |
| |
| IRBuilder<> Builder(Context); |
| Builder.SetInsertPoint(Parent, Exn); |
| LoadInst *LPExn = Builder.CreateLoad(ExnSelSlots.first, "exn.load"); |
| LoadInst *LPSel = Builder.CreateLoad(ExnSelSlots.second, "sel.load"); |
| |
| Exn->replaceAllUsesWith(LPExn); |
| Sel->replaceAllUsesWith(LPSel); |
| } |
| |
| // Remove the dead instructions. |
| for (SmallPtrSet<Instruction*, 32>::iterator |
| I = DeadInsts.begin(), E = DeadInsts.end(); I != E; ++I) { |
| Instruction *Inst = *I; |
| Inst->eraseFromParent(); |
| } |
| |
| // Replace calls to "llvm.eh.resume" with the 'resume' instruction. Load the |
| // exception and selector values from the stored place. |
| Function *EHResume = M->getFunction("llvm.eh.resume"); |
| if (!EHResume) return; |
| |
| while (!EHResume->use_empty()) { |
| CallInst *Resume = cast<CallInst>(EHResume->use_back()); |
| BasicBlock *BB = Resume->getParent(); |
| |
| IRBuilder<> Builder(Context); |
| Builder.SetInsertPoint(BB, Resume); |
| |
| Value *LPadVal = |
| Builder.CreateInsertValue(UndefValue::get(LPadSlotTy), |
| Resume->getArgOperand(0), 0, "lpad.val"); |
| LPadVal = Builder.CreateInsertValue(LPadVal, Resume->getArgOperand(1), |
| 1, "lpad.val"); |
| Builder.CreateResume(LPadVal); |
| |
| // Remove all instructions after the 'resume.' |
| BasicBlock::iterator I = Resume; |
| while (I != BB->end()) { |
| Instruction *Inst = &*I++; |
| Inst->eraseFromParent(); |
| } |
| } |
| } |