/external/clang/test/CXX/dcl.dcl/dcl.attr/dcl.attr.unused/ |
H A D | p2.cpp | 5 static int SI [[maybe_unused]]; // expected-warning {{'maybe_unused' attribute only applies to variables, functions, methods, types, enumerations, enumerators, labels, and non-static data members}} member in struct:S
|
/external/swiftshader/third_party/LLVM/include/llvm-c/ |
H A D | Object.h | 40 void LLVMDisposeSectionIterator(LLVMSectionIteratorRef SI); 42 LLVMSectionIteratorRef SI); 43 void LLVMMoveToNextSection(LLVMSectionIteratorRef SI); 44 const char *LLVMGetSectionName(LLVMSectionIteratorRef SI); 45 uint64_t LLVMGetSectionSize(LLVMSectionIteratorRef SI); 46 const char *LLVMGetSectionContents(LLVMSectionIteratorRef SI); 62 inline section_iterator *unwrap(LLVMSectionIteratorRef SI) { argument 63 return reinterpret_cast<section_iterator*>(SI); 67 wrap(const section_iterator *SI) { argument 69 (const_cast<section_iterator*>(SI)); [all...] |
/external/swiftshader/third_party/LLVM/include/llvm/Transforms/Utils/ |
H A D | BasicBlockUtils.h | 115 inline BasicBlock *SplitCriticalEdge(BasicBlock *BB, succ_iterator SI, argument 117 return SplitCriticalEdge(BB->getTerminator(), SI.getSuccessorIndex(), P);
|
/external/swiftshader/third_party/LLVM/lib/Object/ |
H A D | Object.cpp | 30 section_iterator SI = unwrap(ObjectFile)->begin_sections(); local 31 return wrap(new section_iterator(SI)); 34 void LLVMDisposeSectionIterator(LLVMSectionIteratorRef SI) { argument 35 delete unwrap(SI); 39 LLVMSectionIteratorRef SI) { 40 return (*unwrap(SI) == unwrap(ObjectFile)->end_sections()) ? 1 : 0; 43 void LLVMMoveToNextSection(LLVMSectionIteratorRef SI) { argument 45 unwrap(SI)->increment(ec); 49 const char *LLVMGetSectionName(LLVMSectionIteratorRef SI) { argument 51 if (error_code ec = (*unwrap(SI)) 38 LLVMIsSectionIteratorAtEnd(LLVMObjectFileRef ObjectFile, LLVMSectionIteratorRef SI) argument 56 LLVMGetSectionSize(LLVMSectionIteratorRef SI) argument 63 LLVMGetSectionContents(LLVMSectionIteratorRef SI) argument [all...] |
/external/llvm/lib/Analysis/ |
H A D | MemoryLocation.cpp | 30 MemoryLocation MemoryLocation::get(const StoreInst *SI) { argument 32 SI->getAAMetadata(AATags); 33 const auto &DL = SI->getModule()->getDataLayout(); 35 return MemoryLocation(SI->getPointerOperand(), 36 DL.getTypeStoreSize(SI->getValueOperand()->getType()),
|
H A D | SparsePropagation.cpp | 174 SwitchInst &SI = cast<SwitchInst>(TI); local 177 SCValue = getOrInitValueState(SI.getCondition()); 179 SCValue = getLatticeState(SI.getCondition()); 192 Constant *C = LatticeFunc->GetConstant(SCValue, SI.getCondition(), *this); 198 SwitchInst::CaseIt Case = SI.findCaseValue(cast<ConstantInt>(C));
|
/external/llvm/lib/Target/Hexagon/ |
H A D | HexagonOptimizeSZextends.cpp | 85 SExtInst* SI = new SExtInst(&Arg, Use->getType()); local 86 assert (EVT::getEVT(SI->getType()) == 89 Use->replaceAllUsesWith(SI); 91 SI->insertBefore(First);
|
H A D | HexagonCFGOptimizer.cpp | 148 MachineBasicBlock::succ_iterator SI = MBB->succ_begin(); local 149 MachineBasicBlock* FirstSucc = *SI; 150 MachineBasicBlock* SecondSucc = *(++SI);
|
/external/llvm/lib/Target/NVPTX/ |
H A D | NVPTXLowerAlloca.cpp | 95 auto SI = dyn_cast<StoreInst>(AllocaUse.getUser()); local 96 if (SI && SI->getPointerOperand() == allocaInst && !SI->isVolatile()) { 97 SI->setOperand(SI->getPointerOperandIndex(), NewASCToGeneric);
|
/external/llvm/lib/Transforms/Scalar/ |
H A D | LowerExpectIntrinsic.cpp | 56 static bool handleSwitchExpect(SwitchInst &SI) { argument 57 CallInst *CI = dyn_cast<CallInst>(SI.getCondition()); 70 SwitchInst::CaseIt Case = SI.findCaseValue(ExpectedValue); 71 unsigned n = SI.getNumCases(); // +1 for default case. 74 if (Case == SI.case_default()) 79 SI.setMetadata(LLVMContext::MD_prof, 82 SI.setCondition(ArgValue); 149 } else if (SwitchInst *SI = dyn_cast<SwitchInst>(BB.getTerminator())) { 150 if (handleSwitchExpect(*SI))
|
/external/skia/src/jumper/ |
H A D | SkJumper_misc.h | 15 // Every function in this file should be marked static and inline using SI. 17 #define SI __attribute__((always_inline)) static inline macro 19 #define SI static inline macro 24 SI T unaligned_load(const P* p) { // const void* would work too, but const P* helps ARMv7 codegen. 31 SI void unaligned_store(P* p, T v) { 36 SI Dst bit_cast(const Src& src) { 42 SI Dst widen_cast(const Src& src) { 54 SI void* load_and_inc(void**& program) {
|
/external/skqp/src/jumper/ |
H A D | SkJumper_misc.h | 15 // Every function in this file should be marked static and inline using SI. 17 #define SI __attribute__((always_inline)) static inline macro 19 #define SI static inline macro 24 SI T unaligned_load(const P* p) { // const void* would work too, but const P* helps ARMv7 codegen. 31 SI void unaligned_store(P* p, T v) { 36 SI Dst bit_cast(const Src& src) { 42 SI Dst widen_cast(const Src& src) { 54 SI void* load_and_inc(void**& program) {
|
/external/strace/ |
H A D | sysent_shorthand_undefs.h | 17 #undef SI macro
|
H A D | sysent_shorthand_defs.h | 20 # define SI 0 macro 44 # define SI STACKTRACE_INVALIDATE_CACHE macro
|
/external/swiftshader/third_party/LLVM/lib/Transforms/Scalar/ |
H A D | BasicBlockPlacement.cpp | 127 succ_iterator SI = succ_begin(BB), E = succ_end(BB); local 130 for (; SI != E && PlacedBlocks.count(*SI); ++SI) 132 if (SI == E) return; // No more successors to place. 134 double MaxExecutionCount = PI->getExecutionCount(*SI); 135 BasicBlock *MaxSuccessor = *SI; 138 for (; SI != E; ++SI) 139 if (!PlacedBlocks.count(*SI)) { [all...] |
H A D | LowerAtomic.cpp | 103 static bool LowerStoreInst(StoreInst *SI) { argument 104 SI->setAtomic(NotAtomic); 127 } else if (StoreInst *SI = dyn_cast<StoreInst>(Inst)) { 128 if (SI->isAtomic()) 129 LowerStoreInst(SI);
|
/external/llvm/include/llvm/Transforms/Utils/ |
H A D | BasicBlockUtils.h | 126 SplitCriticalEdge(BasicBlock *BB, succ_iterator SI, argument 129 return SplitCriticalEdge(BB->getTerminator(), SI.getSuccessorIndex(),
|
/external/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUInstrInfo.cpp | 73 SI = 0, enumerator in enum:SIEncodingFamily 92 return SIEncodingFamily::SI; 101 return SIEncodingFamily::SI;
|
/external/swiftshader/third_party/LLVM/lib/Transforms/IPO/ |
H A D | FunctionAttrs.cpp | 172 } else if (StoreInst *SI = dyn_cast<StoreInst>(I)) { 174 if (!SI->isVolatile()) { 175 AliasAnalysis::Location Loc = AA->getLocation(SI); 288 SelectInst *SI = cast<SelectInst>(RVI); local 289 FlowsToReturn.insert(SI->getTrueValue()); 290 FlowsToReturn.insert(SI->getFalseValue());
|
/external/swiftshader/third_party/LLVM/lib/Transforms/Utils/ |
H A D | LowerExpectIntrinsic.cpp | 31 bool HandleSwitchExpect(SwitchInst *SI); 46 bool LowerExpectIntrinsic::HandleSwitchExpect(SwitchInst *SI) { argument 47 CallInst *CI = dyn_cast<CallInst>(SI->getCondition()); 63 unsigned caseNo = SI->findCaseValue(ExpectedValue); 65 unsigned n = SI->getNumCases(); 74 SI->setMetadata(LLVMContext::MD_prof, WeightsNode); 76 SI->setCondition(ArgValue); 135 } else if (SwitchInst *SI = dyn_cast<SwitchInst>(BB->getTerminator())) { 136 if (HandleSwitchExpect(SI))
|
/external/syslinux/com32/cmenu/libmenu/ |
H A D | tui.h | 27 #define SI '\017' macro
|
/external/clang/lib/StaticAnalyzer/Core/ |
H A D | Environment.cpp | 178 SymExpr::symbol_iterator SI = X.symbol_begin(), SE = X.symbol_end(); local 179 for (; SI != SE; ++SI) 180 SymReaper.maybeDead(*SI);
|
/external/clang/unittests/Lex/ |
H A D | HeaderMapTest.cpp | 64 unsigned SI = 1; member in struct:__anon3709::FileMaker 69 assert(SI + S.size() + 1 <= sizeof(File.Bytes)); 70 std::copy(S.begin(), S.end(), File.Bytes + SI); 71 auto OldSI = SI; 72 SI += S.size() + 1;
|
/external/llvm/lib/CodeGen/ |
H A D | InterleavedAccessPass.cpp | 92 bool lowerInterleavedStore(StoreInst *SI, 335 StoreInst *SI, SmallVector<Instruction *, 32> &DeadInsts) { 336 if (!SI->isSimple()) 339 ShuffleVectorInst *SVI = dyn_cast<ShuffleVectorInst>(SI->getValueOperand()); 348 DEBUG(dbgs() << "IA: Found an interleaved store: " << *SI << "\n"); 351 if (!TLI->lowerInterleavedStore(SI, SVI, Factor)) 355 DeadInsts.push_back(SI); 378 if (StoreInst *SI = dyn_cast<StoreInst>(&I)) 379 Changed |= lowerInterleavedStore(SI, DeadInsts); 334 lowerInterleavedStore( StoreInst *SI, SmallVector<Instruction *, 32> &DeadInsts) argument
|
/external/llvm/lib/Target/AArch64/ |
H A D | AArch64PBQPRegAlloc.cpp | 323 SlotIndex SI = LIs.getInstructionIndex(MI); local 324 return LI.expiredAt(SI);
|