/external/llvm/include/llvm/Transforms/Utils/ |
H A D | SimplifyLibCalls.h | 25 class CallInst; 51 Value *optimizeCall(CallInst *CI); 54 Value *optimizeMemCpyChk(CallInst *CI, IRBuilder<> &B); 55 Value *optimizeMemMoveChk(CallInst *CI, IRBuilder<> &B); 56 Value *optimizeMemSetChk(CallInst *CI, IRBuilder<> &B); 59 Value *optimizeStrpCpyChk(CallInst *CI, IRBuilder<> &B, LibFunc::Func Func); 60 Value *optimizeStrpNCpyChk(CallInst *CI, IRBuilder<> &B, LibFunc::Func Func); 64 bool isFortifiedCallFoldable(CallInst *CI, unsigned ObjSizeOp, 100 Value *optimizeCall(CallInst *CI); 104 Value *optimizeStrCat(CallInst *C [all...] |
/external/llvm/include/llvm/CodeGen/ |
H A D | IntrinsicLowering.h | 22 class CallInst; 49 void LowerIntrinsicCall(CallInst *CI); 54 static bool LowerToByteSwap(CallInst *CI);
|
/external/llvm/lib/IR/ |
H A D | Statepoint.cpp | 32 if (isa<InvokeInst>(inst) || isa<CallInst>(inst)) { 51 if (const CallInst *call = dyn_cast<CallInst>(inst)) { 68 if (const CallInst *call = dyn_cast<CallInst>(inst)) {
|
H A D | IRBuilder.cpp | 59 static CallInst *createCallHelper(Value *Callee, ArrayRef<Value *> Ops, 62 CallInst *CI = CallInst::Create(Callee, Ops, Name); 81 CallInst *IRBuilderBase:: 91 CallInst *CI = createCallHelper(TheFn, Ops, this); 106 CallInst *IRBuilderBase:: 118 CallInst *CI = createCallHelper(TheFn, Ops, this); 137 CallInst *IRBuilderBase:: 149 CallInst *CI = createCallHelper(TheFn, Ops, this); 164 CallInst *IRBuilderBas [all...] |
H A D | Instruction.cpp | 296 if (const CallInst *CI = dyn_cast<CallInst>(I1)) 297 return CI->isTailCall() == cast<CallInst>(I2)->isTailCall() && 298 CI->getCallingConv() == cast<CallInst>(I2)->getCallingConv() && 299 CI->getAttributes() == cast<CallInst>(I2)->getAttributes() && 300 CI->hasIdenticalOperandBundleSchema(*cast<CallInst>(I2)); 426 return !cast<CallInst>(this)->doesNotAccessMemory(); 448 return !cast<CallInst>(this)->onlyReadsMemory(); 472 if (const CallInst *CI = dyn_cast<CallInst>(thi [all...] |
/external/llvm/lib/CodeGen/SelectionDAG/ |
H A D | StatepointLowering.h | 62 void scheduleRelocCall(const CallInst &RelocCall) { 68 void relocCallVisited(const CallInst &RelocCall) { 69 SmallVectorImpl<const CallInst *>::iterator itr = 112 SmallVector<const CallInst *, 10> PendingGCRelocateCalls;
|
H A D | SelectionDAGBuilder.h | 38 class CallInst; 484 const CallInst &StackProtCheckCall) { 824 void visitMaskedLoad(const CallInst &I); 825 void visitMaskedStore(const CallInst &I); 826 void visitMaskedGather(const CallInst &I); 827 void visitMaskedScatter(const CallInst &I); 832 void visitCall(const CallInst &I); 833 bool visitMemCmpCall(const CallInst &I); 834 bool visitMemChrCall(const CallInst &I); 835 bool visitStrCpyCall(const CallInst [all...] |
/external/llvm/lib/Analysis/ |
H A D | AssumptionCache.cpp | 42 void AssumptionCache::registerAssumption(CallInst *CI) { 69 assert(match(cast<CallInst>(VH), m_Intrinsic<Intrinsic::assume>()) && 86 OS << " " << *cast<CallInst>(VH)->getArgOperand(0) << "\n"; 117 SmallPtrSet<const CallInst *, 4> AssumptionSet; 121 AssumptionSet.insert(cast<CallInst>(VH)); 126 assert(AssumptionSet.count(cast<CallInst>(&II)) &&
|
/external/llvm/include/llvm/Analysis/ |
H A D | MemoryBuiltins.h | 28 class CallInst; 76 /// extractMallocCall - Returns the corresponding CallInst if the instruction 77 /// is a malloc call. Since CallInst::CreateMalloc() only creates calls, we 79 const CallInst *extractMallocCall(const Value *I, const TargetLibraryInfo *TLI); 80 static inline CallInst *extractMallocCall(Value *I, 82 return const_cast<CallInst*>(extractMallocCall((const Value*)I, TLI)); 90 PointerType *getMallocType(const CallInst *CI, const TargetLibraryInfo *TLI); 97 Type *getMallocAllocatedType(const CallInst *CI, const TargetLibraryInfo *TLI); 104 Value *getMallocArraySize(CallInst *CI, const DataLayout &DL, 112 /// extractCallocCall - Returns the corresponding CallInst i [all...] |
H A D | VectorUtils.h | 48 Intrinsic::ID checkUnaryFloatSignature(const CallInst &I, 56 Intrinsic::ID checkBinaryFloatSignature(const CallInst &I, 62 Intrinsic::ID getIntrinsicIDForCall(CallInst *CI, const TargetLibraryInfo *TLI);
|
H A D | ObjCARCInstKind.h | 102 if (const CallInst *CI = dyn_cast<CallInst>(V)) {
|
/external/llvm/include/llvm/IR/ |
H A D | AutoUpgrade.h | 20 class CallInst; 37 void UpgradeIntrinsicCall(CallInst *CI, Function *NewFn);
|
H A D | CallSite.h | 37 class CallInst; 46 typename CallTy = const CallInst, 75 /// isCall - true if a CallInst is enclosed. 194 /// CallInst can be tail call optimized. 196 return isCall() && cast<CallInst>(getInstruction())->isMustTailCall(); 201 return isCall() && cast<CallInst>(getInstruction())->isTailCall(); 207 ? cast<CallInst>(II)->METHOD \ 213 cast<CallInst>(II)->METHOD; \ 227 return cast<CallInst>(getInstruction())->isInlineAsm(); 397 cast<CallInst>(I [all...] |
/external/llvm/lib/Transforms/ObjCARC/ |
H A D | ObjCARC.h | 54 Value *OldArg = cast<CallInst>(CI)->getArgOperand(0);
|
H A D | ObjCARCOpts.cpp | 64 cast<CallInst>(Arg)->getArgOperand(0)); 633 cast<CallInst>(RetainRV)->setCalledFunction(NewDecl); 667 CallInst *AutoreleaseRVCI = cast<CallInst>(AutoreleaseRV); 716 CallInst *CI = cast<CallInst>(Inst); 734 CallInst *CI = cast<CallInst>(Inst); 764 CallInst *Call = cast<CallInst>(Ins [all...] |
H A D | ObjCARCContract.cpp | 74 SmallPtrSet<CallInst *, 8> StoreStrongCalls; 139 cast<CallInst>(Retain)->setCalledFunction(Decl); 154 CallInst *Retain = nullptr; 170 Retain = dyn_cast_or_null<CallInst>(*DependingInstructions.begin()); 385 CallInst *StoreStrong = CallInst::Create(Decl, Args, "", Store); 458 CallInst::Create(IA, "", Inst); 465 CallInst *CI = cast<CallInst>(Inst); 549 Value *Arg = cast<CallInst>(Ins [all...] |
/external/llvm/tools/bugpoint-passes/ |
H A D | TestPasses.cpp | 38 if (isa<CallInst>(*I)) 61 if (CallInst *CI = dyn_cast<CallInst>(I)) {
|
/external/llvm/unittests/IR/ |
H A D | WaymarkTest.cpp | 34 const CallInst *A = CallInst::Create(F.get(), makeArrayRef(values)); 35 ASSERT_NE(A, (const CallInst*)nullptr);
|
/external/llvm/lib/Transforms/Scalar/ |
H A D | LowerExpectIntrinsic.cpp | 45 CallInst *CI = dyn_cast<CallInst>(SI.getCondition()); 87 CallInst *CI; 91 CI = dyn_cast<CallInst>(BI.getCondition()); 95 CI = dyn_cast<CallInst>(CmpI->getOperand(0)); 144 CallInst *CI = dyn_cast<CallInst>(BI++);
|
H A D | PartiallyInlineLibCalls.cpp | 44 bool optimizeSQRT(CallInst *Call, Function *CalledFunc, 72 CallInst *Call = dyn_cast<CallInst>(&*II); 104 bool PartiallyInlineLibCalls::optimizeSQRT(CallInst *Call,
|
H A D | TailRecursionElimination.cpp | 105 CallInst *FindTRECandidate(Instruction *I, 107 bool EliminateRecursiveTailCall(CallInst *CI, ReturnInst *Ret, 121 bool CanMoveAboveCall(Instruction *I, CallInst *CI); 122 Value *CanTransformAccumulatorRecursion(Instruction *I, CallInst *CI); 294 SmallVector<CallInst *, 32> DeferredTails; 303 CallInst *CI = dyn_cast<CallInst>(&I); 371 for (CallInst *CI : DeferredTails) { 446 bool TailCallElim::CanMoveAboveCall(Instruction *I, CallInst *CI) { 482 static bool isDynamicConstant(Value *V, CallInst *C [all...] |
/external/llvm/examples/Fibonacci/ |
H A D | fibonacci.cpp | 72 CallInst *CallFibX1 = CallInst::Create(FibF, Sub, "fibx1", RecurseBB); 77 CallInst *CallFibX2 = CallInst::Create(FibF, Sub, "fibx2", RecurseBB);
|
/external/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUAnnotateKernelFeatures.cpp | 63 // CallInst is the only valid user for an intrinsic. 64 CallInst *CI = cast<CallInst>(U);
|
/external/llvm/lib/Transforms/Utils/ |
H A D | LowerInvoke.cpp | 60 CallInst *NewCall = CallInst::Create(II->getCalledValue(),
|
/external/llvm/lib/Transforms/IPO/ |
H A D | PruneEH.cpp | 129 if (const auto *CI = dyn_cast<CallInst>(&I)) { 194 CallInst *Call = CallInst::Create(II->getCalledValue(), Args, OpBundles, 225 if (CallInst *CI = dyn_cast<CallInst>(I++)) 257 if (CallInst *CI = dyn_cast<CallInst>(I)) {
|