/external/clang/test/Modules/Inputs/merge-template-members/ |
H A D | a1.h | 8 typedef A<int> AI; typedef in namespace:N
|
H A D | a2.h | 8 typedef A<int> AI; typedef in namespace:N
|
/external/skia/src/opts/ |
H A D | SkNx_neon.h | 19 AI static float32x4_t armv7_vrndmq_f32(float32x4_t v) { 28 AI SkNx(float32x2_t vec) : fVec(vec) {} 30 AI SkNx() {} 31 AI SkNx(float val) : fVec(vdup_n_f32(val)) {} 32 AI SkNx(float a, float b) { fVec = (float32x2_t) { a, b }; } 34 AI static SkNx Load(const void* ptr) { return vld1_f32((const float*)ptr); } 35 AI void store(void* ptr) const { vst1_f32((float*)ptr, fVec); } 37 AI SkNx invert() const { 43 AI SkNx operator + (const SkNx& o) const { return vadd_f32(fVec, o.fVec); } 44 AI SkN [all...] |
H A D | SkNx_sse.h | 21 AI SkNx(const __m128& vec) : fVec(vec) {} 23 AI SkNx() {} 24 AI SkNx(float val) : fVec(_mm_set1_ps(val)) {} 25 AI static SkNx Load(const void* ptr) { 28 AI SkNx(float a, float b) : fVec(_mm_setr_ps(a,b,0,0)) {} 30 AI void store(void* ptr) const { _mm_storel_pi((__m64*)ptr, fVec); } 32 AI SkNx operator + (const SkNx& o) const { return _mm_add_ps(fVec, o.fVec); } 33 AI SkNx operator - (const SkNx& o) const { return _mm_sub_ps(fVec, o.fVec); } 34 AI SkNx operator * (const SkNx& o) const { return _mm_mul_ps(fVec, o.fVec); } 35 AI SkN [all...] |
/external/llvm/lib/Transforms/Utils/ |
H A D | InstructionNamer.cpp | 35 for (Function::arg_iterator AI = F.arg_begin(), AE = F.arg_end(); 36 AI != AE; ++AI) 37 if (!AI->hasName() && !AI->getType()->isVoidTy()) 38 AI->setName("arg");
|
H A D | MetaRenamer.cpp | 84 for (auto AI = M.alias_begin(), AE = M.alias_end(); AI != AE; ++AI) { 85 StringRef Name = AI->getName(); 89 AI->setName("alias"); 125 for (auto AI = F.arg_begin(), AE = F.arg_end(); AI != AE; ++AI) 126 if (!AI->getType()->isVoidTy()) 127 AI [all...] |
H A D | PromoteMemoryToRegister.cpp | 51 bool llvm::isAllocaPromotable(const AllocaInst *AI) { argument 54 unsigned AS = AI->getType()->getAddressSpace(); 57 for (const User *U : AI->users()) { 64 if (SI->getOperand(0) == AI) 65 return false; // Don't allow a store OF the AI, only INTO the AI. 119 void AnalyzeAlloca(AllocaInst *AI) { argument 125 for (auto UI = AI->user_begin(), E = AI->user_end(); UI != E;) { 149 DbgDeclare = FindAllocaDbgDeclare(AI); 304 removeLifetimeIntrinsicUsers(AllocaInst *AI) argument 336 rewriteSingleStoreAlloca(AllocaInst *AI, AllocaInfo &Info, LargeBlockInfo &LBI, DominatorTree &DT, AliasSetTracker *AST) argument 436 promoteSingleBlockAlloca(AllocaInst *AI, const AllocaInfo &Info, LargeBlockInfo &LBI, AliasSetTracker *AST) argument 529 AllocaInst *AI = Allocas[AllocaNum]; local 775 ComputeLiveInBlocks( AllocaInst *AI, AllocaInfo &Info, const SmallPtrSetImpl<BasicBlock *> &DefBlocks, SmallPtrSetImpl<BasicBlock *> &LiveInBlocks) argument 935 DenseMap<AllocaInst *, unsigned>::iterator AI = AllocaLookup.find(Src); local [all...] |
/external/swiftshader/third_party/LLVM/lib/Transforms/Utils/ |
H A D | InstructionNamer.cpp | 35 for (Function::arg_iterator AI = F.arg_begin(), AE = F.arg_end(); 36 AI != AE; ++AI) 37 if (!AI->hasName() && !AI->getType()->isVoidTy()) 38 AI->setName("arg");
|
H A D | PromoteMemoryToRegister.cpp | 80 bool llvm::isAllocaPromotable(const AllocaInst *AI) { argument 85 for (Value::const_use_iterator UI = AI->use_begin(), UE = AI->use_end(); 94 if (SI->getOperand(0) == AI) 95 return false; // Don't allow a store OF the AI, only INTO the AI. 275 void DetermineInsertionPoint(AllocaInst *AI, unsigned AllocaNum, 277 void ComputeLiveInBlocks(AllocaInst *AI, AllocaInfo &Info, 281 void RewriteSingleStoreAlloca(AllocaInst *AI, AllocaInfo &Info, 283 void PromoteSingleBlockAlloca(AllocaInst *AI, AllocaInf 315 AnalyzeAlloca(AllocaInst *AI) argument 359 removeLifetimeIntrinsicUsers(AllocaInst *AI) argument 395 AllocaInst *AI = Allocas[AllocaNum]; local 672 ComputeLiveInBlocks(AllocaInst *AI, AllocaInfo &Info, const SmallPtrSet<BasicBlock*, 32> &DefBlocks, SmallPtrSet<BasicBlock*, 32> &LiveInBlocks) argument 743 DetermineInsertionPoint(AllocaInst *AI, unsigned AllocaNum, AllocaInfo &Info) argument 831 RewriteSingleStoreAlloca(AllocaInst *AI, AllocaInfo &Info, LargeBlockInfo &LBI) argument 919 PromoteSingleBlockAlloca(AllocaInst *AI, AllocaInfo &Info, LargeBlockInfo &LBI) argument 1072 DenseMap<AllocaInst*, unsigned>::iterator AI = AllocaLookup.find(Src); local [all...] |
H A D | Mem2Reg.cpp | 72 if (AllocaInst *AI = dyn_cast<AllocaInst>(I)) // Is it an alloca? 73 if (isAllocaPromotable(AI)) 74 Allocas.push_back(AI);
|
/external/skia/src/core/ |
H A D | SkNx.h | 18 #define AI SK_ALWAYS_INLINE macro 29 AI SkNx() = default; 30 AI SkNx(const Half& lo, const Half& hi) : fLo(lo), fHi(hi) {} 32 AI SkNx(T v) : fLo(v), fHi(v) {} 34 AI SkNx(T a, T b) : fLo(a) , fHi(b) { static_assert(N==2, ""); } 35 AI SkNx(T a, T b, T c, T d) : fLo(a,b), fHi(c,d) { static_assert(N==4, ""); } 36 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h) : fLo(a,b,c,d), fHi(e,f,g,h) { 39 AI SkNx(T a, T b, T c, T d, T e, T f, T g, T h, 45 AI T operator[](int k) const { 50 AI stati 380 #undef AI macro [all...] |
/external/harfbuzz_ng/src/ |
H A D | hb-atomic-private.hh | 67 #define hb_atomic_int_impl_add(AI, V) InterlockedExchangeAdd (&(AI), (V)) 85 #define hb_atomic_int_impl_add(AI, V) (OSAtomicAdd32Barrier ((V), &(AI)) - (V)) 103 #define hb_atomic_int_impl_add(AI, V) __sync_fetch_and_add (&(AI), (V)) 116 #define hb_atomic_int_impl_add(AI, V) ( ({__machine_rw_barrier ();}), atomic_add_int_nv (&(AI), (V)) - (V)) 127 static inline int hb_fetch_and_add(volatile int* AI, unsigned int V) { argument 129 int result = __fetch_and_add(AI, [all...] |
/external/llvm/include/llvm/CodeGen/ |
H A D | AtomicExpandUtils.h | 56 expandAtomicRMWToCmpXchg(AtomicRMWInst *AI, CreateCmpXchgInstFun Factory);
|
/external/llvm/tools/opt/ |
H A D | AnalysisWrappers.cpp | 47 for (CallSite::arg_iterator AI = CS.arg_begin(), 48 E = CS.arg_end(); AI != E; ++AI) { 49 if (!isa<Constant>(*AI)) continue;
|
/external/llvm/lib/CodeGen/ |
H A D | StackProtector.cpp | 61 StackProtector::getSSPLayout(const AllocaInst *AI) const { 62 return AI ? Layout.lookup(AI) : SSPLK_None; 166 bool StackProtector::HasAddressTaken(const Instruction *AI) { argument 167 for (const User *U : AI->users()) { 169 if (AI == SI->getValueOperand()) 172 if (AI == SI->getOperand(0)) 237 if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) { 238 if (AI->isArrayAllocation()) { 244 if (const auto *CI = dyn_cast<ConstantInt>(AI 318 CreatePrologue(Function *F, Module *M, ReturnInst *RI, const TargetLoweringBase *TLI, AllocaInst *&AI) argument 340 AllocaInst *AI = nullptr; // Place on stack that stores the stack guard. local [all...] |
H A D | RegUsageInfoCollector.cpp | 82 for (MCRegAliasIterator AI(PReg, TRI, true); AI.isValid(); ++AI) 83 RegMask[*AI / 32] &= ~(1u << *AI % 32);
|
/external/swiftshader/third_party/LLVM/lib/Transforms/Instrumentation/ |
H A D | ProfilingUtils.cpp | 67 Function::arg_iterator AI; local 71 AI = MainFn->arg_begin(); ++AI; 72 if (AI->getType() != ArgVTy) { 73 Instruction::CastOps opcode = CastInst::getCastOpcode(AI, false, ArgVTy, 76 CastInst::Create(opcode, AI, ArgVTy, "argv.cast", InitCall)); 78 InitCall->setArgOperand(1, AI); 83 AI = MainFn->arg_begin(); 86 if (!AI->getType()->isIntegerTy(32)) { 88 if (!AI [all...] |
/external/clang/include/clang/CodeGen/ |
H A D | CGFunctionInfo.h | 124 auto AI = ABIArgInfo(Direct); local 125 AI.setCoerceToType(T); 126 AI.setPaddingType(Padding); 127 AI.setDirectOffset(Offset); 128 AI.setCanBeFlattened(CanBeFlattened); 129 return AI; 132 auto AI = getDirect(T); local 133 AI.setInReg(true); 134 return AI; 137 auto AI local 144 auto AI = getExtend(T); local 154 auto AI = ABIArgInfo(Indirect); local 164 auto AI = getIndirect(Alignment, ByVal, Realign); local 169 auto AI = ABIArgInfo(InAlloca); local 174 auto AI = ABIArgInfo(Expand); local 180 auto AI = getExpand(); local 219 auto AI = ABIArgInfo(CoerceAndExpand); local [all...] |
/external/llvm/include/llvm/Transforms/Scalar/ |
H A D | SROA.h | 115 bool presplitLoadsAndStores(AllocaInst &AI, sroa::AllocaSlices &AS); 116 AllocaInst *rewritePartition(AllocaInst &AI, sroa::AllocaSlices &AS, 118 bool splitAlloca(AllocaInst &AI, sroa::AllocaSlices &AS); 119 bool runOnAlloca(AllocaInst &AI);
|
/external/llvm/include/llvm/Transforms/Utils/ |
H A D | PromoteMemToReg.h | 32 bool isAllocaPromotable(const AllocaInst *AI);
|
/external/swiftshader/third_party/LLVM/include/llvm/Transforms/Utils/ |
H A D | PromoteMemToReg.h | 30 bool isAllocaPromotable(const AllocaInst *AI);
|
/external/swiftshader/third_party/LLVM/tools/opt/ |
H A D | AnalysisWrappers.cpp | 48 for (CallSite::arg_iterator AI = CS.arg_begin(),
49 E = CS.arg_end(); AI != E; ++AI) {
50 if (!isa<Constant>(*AI)) continue;
|
/external/llvm/unittests/Transforms/Utils/ |
H A D | IntegerDivision.cpp | 37 Function::arg_iterator AI = F->arg_begin(); local 38 Value *A = &*AI++; 39 Value *B = &*AI++; 67 Function::arg_iterator AI = F->arg_begin(); local 68 Value *A = &*AI++; 69 Value *B = &*AI++; 97 Function::arg_iterator AI = F->arg_begin(); local 98 Value *A = &*AI++; 99 Value *B = &*AI++; 127 Function::arg_iterator AI local 158 Function::arg_iterator AI = F->arg_begin(); local 188 Function::arg_iterator AI = F->arg_begin(); local 218 Function::arg_iterator AI = F->arg_begin(); local 248 Function::arg_iterator AI = F->arg_begin(); local [all...] |
/external/swiftshader/third_party/LLVM/lib/Transforms/Scalar/ |
H A D | ScalarReplAggregates.cpp | 88 AllocaInst *AI; member in struct:__anon18977::SROA::AllocaInfo 114 : AI(ai), isUnsafe(false), isMemCpySrc(false), isMemCpyDst(false), 125 bool isSafeAllocaToScalarRepl(AllocaInst *AI); 138 void DoScalarReplacement(AllocaInst *AI, 142 void RewriteForScalarRepl(Instruction *I, AllocaInst *AI, uint64_t Offset, 144 void RewriteBitCast(BitCastInst *BC, AllocaInst *AI, uint64_t Offset, 146 void RewriteGEP(GetElementPtrInst *GEPI, AllocaInst *AI, uint64_t Offset, 148 void RewriteLifetimeIntrinsic(IntrinsicInst *II, AllocaInst *AI, 152 AllocaInst *AI, 154 void RewriteStoreUserOfWholeAlloca(StoreInst *SI, AllocaInst *AI, 290 TryConvert(AllocaInst *AI) argument 925 AllocaInst *AI; member in class:__anon18980::AllocaPromoter 934 run(AllocaInst *AI, const SmallVectorImpl<Instruction*> &Insts) argument 1123 tryToMakeAllocaBePromotable(AllocaInst *AI, const TargetData *TD) argument 1317 AllocaInst *AI = Allocas[i]; local 1337 ShouldAttemptScalarRepl(AllocaInst *AI) argument 1366 AllocaInst *AI = WorkList.back(); local 1445 DoScalarReplacement(AllocaInst *AI, std::vector<AllocaInst*> &WorkList) argument 1756 RewriteForScalarRepl(Instruction *I, AllocaInst *AI, uint64_t Offset, SmallVector<AllocaInst*, 32> &NewElts) argument 1870 RewriteBitCast(BitCastInst *BC, AllocaInst *AI, uint64_t Offset, SmallVector<AllocaInst*, 32> &NewElts) argument 1915 RewriteGEP(GetElementPtrInst *GEPI, AllocaInst *AI, uint64_t Offset, SmallVector<AllocaInst*, 32> &NewElts) argument 1958 RewriteLifetimeIntrinsic(IntrinsicInst *II, AllocaInst *AI, uint64_t Offset, SmallVector<AllocaInst*, 32> &NewElts) argument 2014 RewriteMemIntrinUserOfAlloca(MemIntrinsic *MI, Instruction *Inst, AllocaInst *AI, SmallVector<AllocaInst*, 32> &NewElts) argument 2186 RewriteStoreUserOfWholeAlloca(StoreInst *SI, AllocaInst *AI, SmallVector<AllocaInst*, 32> &NewElts) argument 2300 RewriteLoadUserOfWholeAlloca(LoadInst *LI, AllocaInst *AI, SmallVector<AllocaInst*, 32> &NewElts) argument 2424 isSafeAllocaToScalarRepl(AllocaInst *AI) argument 2585 isOnlyCopiedFromConstantGlobal(AllocaInst *AI, SmallVector<Instruction*, 4> &ToDelete) argument [all...] |
/external/llvm/lib/Transforms/IPO/ |
H A D | IPConstantPropagation.cpp | 76 CallSite::arg_iterator AI = CS.arg_begin(); 79 ++i, ++AI, ++Arg) { 85 Constant *C = dyn_cast<Constant>(*AI); 90 } else if (*AI == &*Arg) { 105 Function::arg_iterator AI = F.arg_begin(); local 106 for (unsigned i = 0, e = ArgumentConstants.size(); i != e; ++i, ++AI) { 108 if (ArgumentConstants[i].second || AI->use_empty() || 109 AI->hasInAllocaAttr() || (AI->hasByValAttr() && !F.onlyReadsMemory())) 113 if (!V) V = UndefValue::get(AI [all...] |