/external/llvm/lib/Target/X86/ |
H A D | X86AtomicExpandPass.cpp | 59 bool shouldExpandAtomicRMW(AtomicRMWInst *AI); 61 bool expandAtomicRMW(AtomicRMWInst *AI); 89 if (AtomicRMWInst *AI = dyn_cast<AtomicRMWInst>(Inst)) 90 MadeChange |= expandAtomicRMW(AI); 120 bool X86AtomicExpandPass::shouldExpandAtomicRMW(AtomicRMWInst *AI) { argument 124 if (needsCmpXchgNb(AI->getType())) 127 if (AI->getType()->getPrimitiveSizeInBits() > NativeWidth) 130 AtomicRMWInst::BinOp Op = AI->getOperation(); 144 return !AI->use_empty(); 164 if (AtomicRMWInst *AI 209 expandAtomicRMW(AtomicRMWInst *AI) argument 275 AtomicRMWInst *AI = local [all...] |
/external/chromium_org/third_party/harfbuzz-ng/src/ |
H A D | hb-atomic-private.hh | 63 #define hb_atomic_int_add(AI, V) InterlockedExchangeAdd (&(AI), (V)) 79 #define hb_atomic_int_add(AI, V) (OSAtomicAdd32Barrier ((V), &(AI)) - (V)) 96 #define hb_atomic_int_add(AI, V) __sync_fetch_and_add (&(AI), (V)) 108 #define hb_atomic_int_add(AI, V) ( ({__machine_rw_barrier ();}), atomic_add_int_nv (&(AI), (V)) - (V)) 118 #define hb_atomic_int_add(AI, V) (((AI) [all...] |
/external/harfbuzz_ng/src/ |
H A D | hb-atomic-private.hh | 62 #define hb_atomic_int_add(AI, V) InterlockedExchangeAdd (&(AI), (V)) 78 #define hb_atomic_int_add(AI, V) (OSAtomicAdd32Barrier ((V), &(AI)) - (V)) 95 #define hb_atomic_int_add(AI, V) __sync_fetch_and_add (&(AI), (V)) 107 #define hb_atomic_int_add(AI, V) ( ({__machine_rw_barrier ();}), atomic_add_int_nv (&(AI), (V)) - (V)) 117 #define hb_atomic_int_add(AI, V) (((AI) [all...] |
/external/clang/include/clang/CodeGen/ |
H A D | CGFunctionInfo.h | 101 auto AI = ABIArgInfo(Direct); local 102 AI.setCoerceToType(T); 103 AI.setDirectOffset(Offset); 104 AI.setPaddingType(Padding); 105 return AI; 108 auto AI = getDirect(T); local 109 AI.setInReg(true); 110 return AI; 113 auto AI = ABIArgInfo(Extend); local 114 AI 119 auto AI = getExtend(T); local 129 auto AI = ABIArgInfo(Indirect); local 139 auto AI = getIndirect(Alignment, ByVal, Realign); local 144 auto AI = ABIArgInfo(InAlloca); local 153 auto AI = getExpand(); local [all...] |
/external/llvm/lib/Transforms/Utils/ |
H A D | InstructionNamer.cpp | 35 for (Function::arg_iterator AI = F.arg_begin(), AE = F.arg_end(); 36 AI != AE; ++AI) 37 if (!AI->hasName() && !AI->getType()->isVoidTy()) 38 AI->setName("arg");
|
H A D | MetaRenamer.cpp | 74 for (Module::alias_iterator AI = M.alias_begin(), AE = M.alias_end(); 75 AI != AE; ++AI) { 76 StringRef Name = AI->getName(); 80 AI->setName("alias"); 119 for (Function::arg_iterator AI = F.arg_begin(), AE = F.arg_end(); 120 AI != AE; ++AI) 121 if (!AI->getType()->isVoidTy()) 122 AI [all...] |
H A D | PromoteMemoryToRegister.cpp | 60 bool llvm::isAllocaPromotable(const AllocaInst *AI) { argument 63 unsigned AS = AI->getType()->getAddressSpace(); 66 for (const User *U : AI->users()) { 73 if (SI->getOperand(0) == AI) 74 return false; // Don't allow a store OF the AI, only INTO the AI. 128 void AnalyzeAlloca(AllocaInst *AI) { argument 134 for (auto UI = AI->user_begin(), E = AI->user_end(); UI != E;) { 158 DbgDeclare = FindAllocaDbgDeclare(AI); 315 removeLifetimeIntrinsicUsers(AllocaInst *AI) argument 347 rewriteSingleStoreAlloca(AllocaInst *AI, AllocaInfo &Info, LargeBlockInfo &LBI, DominatorTree &DT, AliasSetTracker *AST) argument 444 promoteSingleBlockAlloca(AllocaInst *AI, const AllocaInfo &Info, LargeBlockInfo &LBI, AliasSetTracker *AST) argument 530 AllocaInst *AI = Allocas[AllocaNum]; local 767 ComputeLiveInBlocks( AllocaInst *AI, AllocaInfo &Info, const SmallPtrSet<BasicBlock *, 32> &DefBlocks, SmallPtrSet<BasicBlock *, 32> &LiveInBlocks) argument 842 DetermineInsertionPoint(AllocaInst *AI, unsigned AllocaNum, AllocaInfo &Info) argument 1017 DenseMap<AllocaInst *, unsigned>::iterator AI = AllocaLookup.find(Src); local [all...] |
H A D | Mem2Reg.cpp | 73 if (AllocaInst *AI = dyn_cast<AllocaInst>(I)) // Is it an alloca? 74 if (isAllocaPromotable(AI)) 75 Allocas.push_back(AI);
|
/external/llvm/lib/CodeGen/ |
H A D | MachineCopyPropagation.cpp | 69 for (MCRegAliasIterator AI(Reg, TRI, true); AI.isValid(); ++AI) { 70 SourceMap::iterator SI = SrcMap.find(*AI); 194 for (MCRegAliasIterator AI(Src, TRI, true); AI.isValid(); ++AI) { 195 CI = CopyMap.find(*AI); 218 for (MCRegAliasIterator AI(Def, TRI, false); AI [all...] |
H A D | AtomicExpandLoadLinkedPass.cpp | 45 bool expandAtomicRMW(AtomicRMWInst *AI); 84 if (AtomicRMWInst *AI = dyn_cast<AtomicRMWInst>(Inst)) 85 MadeChange |= expandAtomicRMW(AI); 125 AtomicRMWInst *AI = local 131 return expandAtomicRMW(AI); 134 bool AtomicExpandLoadLinked::expandAtomicRMW(AtomicRMWInst *AI) { argument 135 AtomicOrdering Order = AI->getOrdering(); 136 Value *Addr = AI->getPointerOperand(); 137 BasicBlock *BB = AI->getParent(); 155 BasicBlock *ExitBB = BB->splitBasicBlock(AI, "atomicrm [all...] |
H A D | StackProtector.cpp | 57 StackProtector::getSSPLayout(const AllocaInst *AI) const { 58 return AI ? Layout.lookup(AI) : SSPLK_None; 153 bool StackProtector::HasAddressTaken(const Instruction *AI) { argument 154 for (const User *U : AI->users()) { 156 if (AI == SI->getValueOperand()) 159 if (AI == SI->getOperand(0)) 217 if (AllocaInst *AI = dyn_cast<AllocaInst>(II)) { 218 if (AI->isArrayAllocation()) { 225 dyn_cast<ConstantInt>(AI 336 CreatePrologue(Function *F, Module *M, ReturnInst *RI, const TargetLoweringBase *TLI, const Triple &Trip, AllocaInst *&AI, Value *&StackGuardVar) argument 361 AI); local 376 AllocaInst *AI = nullptr; // Place on stack that stores the stack guard. local [all...] |
/external/llvm/include/llvm/Transforms/Utils/ |
H A D | PromoteMemToReg.h | 32 bool isAllocaPromotable(const AllocaInst *AI);
|
H A D | ASanStackFrameLayout.h | 34 AllocaInst *AI; // The actual AllocaInst. member in struct:llvm::ASanStackVariableDescription
|
/external/llvm/lib/Target/Hexagon/ |
H A D | HexagonRemoveSZExtArgs.cpp | 58 for (Function::arg_iterator AI = F.arg_begin(), AE = F.arg_end(); AI != AE; 59 ++AI, ++Idx) { 61 Argument* Arg = AI;
|
/external/llvm/tools/opt/ |
H A D | AnalysisWrappers.cpp | 47 for (CallSite::arg_iterator AI = CS.arg_begin(), 48 E = CS.arg_end(); AI != E; ++AI) { 49 if (!isa<Constant>(*AI)) continue;
|
/external/llvm/unittests/Transforms/Utils/ |
H A D | IntegerDivision.cpp | 37 Function::arg_iterator AI = F->arg_begin(); local 38 Value *A = AI++; 39 Value *B = AI++; 67 Function::arg_iterator AI = F->arg_begin(); local 68 Value *A = AI++; 69 Value *B = AI++; 97 Function::arg_iterator AI = F->arg_begin(); local 98 Value *A = AI++; 99 Value *B = AI++; 127 Function::arg_iterator AI local 158 Function::arg_iterator AI = F->arg_begin(); local 188 Function::arg_iterator AI = F->arg_begin(); local 218 Function::arg_iterator AI = F->arg_begin(); local 248 Function::arg_iterator AI = F->arg_begin(); local [all...] |
/external/llvm/lib/IR/ |
H A D | Mangler.cpp | 66 for (Function::const_arg_iterator AI = F->arg_begin(), AE = F->arg_end(); 67 AI != AE; ++AI) { 68 Type *Ty = AI->getType(); 70 if (AI->hasByValOrInAllocaAttr())
|
/external/llvm/lib/Transforms/Scalar/ |
H A D | ScalarReplAggregates.cpp | 102 AllocaInst *AI; member in struct:__anon25495::SROA::AllocaInfo 128 : AI(ai), isUnsafe(false), isMemCpySrc(false), isMemCpyDst(false), 152 bool isSafeAllocaToScalarRepl(AllocaInst *AI); 165 void DoScalarReplacement(AllocaInst *AI, 169 void RewriteForScalarRepl(Instruction *I, AllocaInst *AI, uint64_t Offset, 171 void RewriteBitCast(BitCastInst *BC, AllocaInst *AI, uint64_t Offset, 173 void RewriteGEP(GetElementPtrInst *GEPI, AllocaInst *AI, uint64_t Offset, 175 void RewriteLifetimeIntrinsic(IntrinsicInst *II, AllocaInst *AI, 179 AllocaInst *AI, 181 void RewriteStoreUserOfWholeAlloca(StoreInst *SI, AllocaInst *AI, 333 TryConvert(AllocaInst *AI) argument 1054 AllocaInst *AI; member in class:__anon25498::AllocaPromoter 1063 run(AllocaInst *AI, const SmallVectorImpl<Instruction*> &Insts) argument 1244 tryToMakeAllocaBePromotable(AllocaInst *AI, const DataLayout *DL) argument 1432 AllocaInst *AI = Allocas[i]; local 1451 ShouldAttemptScalarRepl(AllocaInst *AI) argument 1478 AllocaInst *AI = WorkList.back(); local 1537 DoScalarReplacement(AllocaInst *AI, std::vector<AllocaInst*> &WorkList) argument 1863 RewriteForScalarRepl(Instruction *I, AllocaInst *AI, uint64_t Offset, SmallVectorImpl<AllocaInst *> &NewElts) argument 1977 RewriteBitCast(BitCastInst *BC, AllocaInst *AI, uint64_t Offset, SmallVectorImpl<AllocaInst *> &NewElts) argument 2035 RewriteGEP(GetElementPtrInst *GEPI, AllocaInst *AI, uint64_t Offset, SmallVectorImpl<AllocaInst *> &NewElts) argument 2096 RewriteLifetimeIntrinsic(IntrinsicInst *II, AllocaInst *AI, uint64_t Offset, SmallVectorImpl<AllocaInst *> &NewElts) argument 2154 RewriteMemIntrinUserOfAlloca(MemIntrinsic *MI, Instruction *Inst, AllocaInst *AI, SmallVectorImpl<AllocaInst *> &NewElts) argument 2328 RewriteStoreUserOfWholeAlloca(StoreInst *SI, AllocaInst *AI, SmallVectorImpl<AllocaInst *> &NewElts) argument 2443 RewriteLoadUserOfWholeAlloca(LoadInst *LI, AllocaInst *AI, SmallVectorImpl<AllocaInst *> &NewElts) argument 2567 isSafeAllocaToScalarRepl(AllocaInst *AI) argument [all...] |
/external/llvm/lib/Transforms/IPO/ |
H A D | IPConstantPropagation.cpp | 106 CallSite::arg_iterator AI = CS.arg_begin(); 109 ++i, ++AI, ++Arg) { 115 Constant *C = dyn_cast<Constant>(*AI); 120 } else if (*AI == &*Arg) { 135 Function::arg_iterator AI = F.arg_begin(); local 136 for (unsigned i = 0, e = ArgumentConstants.size(); i != e; ++i, ++AI) { 138 if (ArgumentConstants[i].second || AI->use_empty() || 139 AI->hasInAllocaAttr() || (AI->hasByValAttr() && !F.onlyReadsMemory())) 143 if (!V) V = UndefValue::get(AI [all...] |
/external/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineLoadStoreAlloca.cpp | 157 isOnlyCopiedFromConstantGlobal(AllocaInst *AI, argument 160 if (isOnlyCopiedFromConstantGlobal(AI, TheCopy, ToDelete)) 165 Instruction *InstCombiner::visitAllocaInst(AllocaInst &AI) { argument 169 Type *IntPtrTy = DL->getIntPtrType(AI.getType()); 170 if (AI.getArraySize()->getType() != IntPtrTy) { 171 Value *V = Builder->CreateIntCast(AI.getArraySize(), 173 AI.setOperand(0, V); 174 return &AI; 179 if (AI.isArrayAllocation()) { // Check C != 1 180 if (const ConstantInt *C = dyn_cast<ConstantInt>(AI [all...] |
/external/llvm/lib/DebugInfo/ |
H A D | DWARFDebugLoc.cpp | 48 RelocAddrMap::const_iterator AI = RelocMap.find(Offset); local 51 if (AI != RelocMap.end()) 52 E.Begin += AI->second.second; 54 AI = RelocMap.find(Offset); 57 if (AI != RelocMap.end()) 58 E.End += AI->second.second;
|
H A D | DWARFFormValue.cpp | 146 RelocAddrMap::const_iterator AI = cu->getRelocMap()->find(*offset_ptr); local 147 if (AI != cu->getRelocMap()->end()) { 148 const std::pair<uint8_t, int64_t> &R = AI->second; 182 RelocAddrMap::const_iterator AI = cu->getRelocMap()->find(*offset_ptr); local 184 if (AI != cu->getRelocMap()->end()) 185 Value.uval += AI->second.second; 196 RelocAddrMap::const_iterator AI local 198 if (AI != cu->getRelocMap()->end()) { 199 const std::pair<uint8_t, int64_t> &R = AI->second; 218 RelocAddrMap::const_iterator AI local [all...] |
/external/llvm/include/llvm/ |
H A D | PassSupport.h | 166 PassInfo *AI = new PassInfo(name, & agName :: ID); \ 167 Registry.registerAnalysisGroup(& agName ::ID, 0, *AI, false, true); \ 168 return AI; \ 182 PassInfo *AI = new PassInfo(name, & agName :: ID); \ 184 *AI, def, true); \ 185 return AI; \ 201 PassInfo *AI = new PassInfo(n, & agName :: ID); \ 203 *AI, def, true); \ 204 return AI; \
|
/external/llvm/include/llvm/CodeGen/ |
H A D | StackProtector.h | 100 bool HasAddressTaken(const Instruction *AI); 122 SSPLayoutKind getSSPLayout(const AllocaInst *AI) const;
|
/external/llvm/lib/Analysis/ |
H A D | CodeMetrics.cpp | 59 if (const AllocaInst *AI = dyn_cast<AllocaInst>(II)) { 60 if (!AI->isStaticAlloca())
|