/external/tensorflow/tensorflow/core/lib/core/ |
H A D | arena.cc | 40 // Arena::Arena() 41 // Arena::~Arena() 45 Arena::Arena(const size_t block_size) function in class:tensorflow::core::Arena 61 Arena::~Arena() { 72 bool Arena::SatisfyAlignment(size_t alignment) { 87 // Arena [all...] |
H A D | arena.h | 16 // TODO(vrv): Switch this to an open-sourced version of Arena. 35 class Arena { class in namespace:tensorflow::core 38 explicit Arena(const size_t block_size); 39 ~Arena(); 104 TF_DISALLOW_COPY_AND_ASSIGN(Arena);
|
/external/python/cpython2/Lib/multiprocessing/ |
H A D | heap.py | 57 class Arena(object): class in inherits:object 63 self.name = 'pym-%d-%d' % (os.getpid(), Arena._counter.next()) 79 class Arena(object): class in inherits:object 120 arena = Arena(length)
|
/external/python/cpython3/Lib/multiprocessing/ |
H A D | heap.py | 30 class Arena(object): class in inherits:object 62 class Arena(object): class in inherits:object 85 raise ValueError('Arena is unpicklable because ' 90 return Arena(size, dupfd.detach()) 92 reduction.register(Arena, reduce_arena) 128 arena = Arena(length)
|
/external/protobuf/src/google/protobuf/ |
H A D | arena.cc | 42 google::protobuf::internal::SequenceNumber Arena::lifecycle_id_generator_; 44 Arena::ThreadCache& Arena::thread_cache() { 50 Arena::ThreadCache& Arena::thread_cache() { 55 GOOGLE_THREAD_LOCAL Arena::ThreadCache Arena::thread_cache_ = { -1, NULL }; 58 void Arena::Init() { 91 Arena::~Arena() { [all...] |
H A D | arena.h | 63 class Arena; // defined below 101 // creator of the Arena retains ownership of the block after the Arena is 128 void* (*on_arena_init)(Arena* arena); 129 void (*on_arena_reset)(Arena* arena, void* cookie, uint64 space_used); 130 void (*on_arena_destruction)(Arena* arena, void* cookie, uint64 space_used); 167 // Arena allocator. Arena allocation replaces ordinary (heap-based) allocation 170 // allocated on an arena by using Arena::CreateMessage<T>(Arena*), belo 223 explicit Arena(const ArenaOptions& options) : options_(options) { function in class:google::protobuf::final 229 Arena() { function in class:google::protobuf::final [all...] |
/external/clang/include/clang/Analysis/Analyses/ |
H A D | ThreadSafetyCommon.h | 341 : Arena(A), SelfVar(nullptr), Scfg(nullptr), CurrentBB(nullptr), 344 SelfVar = new (Arena) til::Variable(nullptr); 478 til::MemRegionRef Arena; member in class:clang::threadSafety::SExprBuilder
|
H A D | ThreadSafetyTraverse.h | 140 Container(CopyReducerBase &S, unsigned N) : Elems(S.Arena, N) {} 148 CopyReducerBase(MemRegionRef A) : Arena(A) {} 151 MemRegionRef Arena; member in class:clang::threadSafety::til::CopyReducerBase
|
H A D | ThreadSafetyTIL.h | 1550 : SExpr(COP_BasicBlock), Arena(A), CFGPtr(nullptr), BlockID(0), 1554 : SExpr(COP_BasicBlock), Arena(A), CFGPtr(nullptr), BlockID(0),Visited(0), 1600 Args.reserveCheck(1, Arena); 1605 Instrs.reserveCheck(1, Arena); 1613 void reserveArguments(unsigned Nargs) { Args.reserve(Nargs, Arena); } 1616 void reserveInstructions(unsigned Nins) { Instrs.reserve(Nins, Arena); } 1667 MemRegionRef Arena; // The arena used to allocate this block. member in class:clang::threadSafety::BasicBlock 1695 : SExpr(COP_SCFG), Arena(A), Blocks(A, Nblocks), 1706 : SExpr(COP_SCFG), Arena(Cfg.Arena), Block 1776 MemRegionRef Arena; member in class:clang::threadSafety::SCFG [all...] |
/external/clang/lib/Analysis/ |
H A D | ThreadSafetyCommon.cpp | 160 return CapabilityExpr(new (Arena) til::Wildcard(), false); 253 return new (Arena) til::Literal(cast<Expr>(S)); 263 return new (Arena) til::Undefined(S); 287 return new (Arena) til::LiteralPtr(VD); 337 til::SExpr *E = new (Arena) til::SApply(BE); 344 til::Project *P = new (Arena) til::Project(E, D); 370 E = new (Arena) til::Apply(E, A); 372 return new (Arena) til::Call(E, CE); 382 return new (Arena) til::Cast(til::CAST_objToPtr, E); 397 return new (Arena) ti 856 static_cast<unsigned>(CurrentArguments.size()), Arena); local 877 static_cast<unsigned>(CurrentInstructions.size()), Arena); local [all...] |
/external/swiftshader/third_party/subzero/src/ |
H A D | IceDefs.h | 166 VariableDeclarationList() : Arena(new ArenaAllocator()) {} 174 return Arena->Allocate<T>(Count); 181 T *Ret = Arena->Allocate<T>(); 194 addArena(std::move(Other->Arena)); 210 if (Arena == nullptr) { 211 // Arena is only null if this was merged, so we ensure there's no state 225 Arena->Reset(); 269 ArenaPtr Arena; member in class:Ice::VariableDeclarationList
|
/external/valgrind/coregrind/ |
H A D | m_mallocfree.c | 79 red zone bytes (depends on Arena.rz_szB, but >= sizeof(void*)) 81 red zone bytes (depends on Arena.rz_szB, but >= sizeof(void*)) 89 excess red zone bytes (if Arena.rz_szB > sizeof(void*)) 91 excess red zone bytes (if Arena.rz_szB > sizeof(void*)) 249 Arena; typedef in typeref:struct:__anon29026 343 SizeT overhead_szB_lo ( Arena* a ) 348 SizeT overhead_szB_hi ( Arena* a ) 353 SizeT overhead_szB ( Arena* a ) 363 SizeT min_useful_bszB ( Arena* a ) 372 SizeT pszB_to_bszB ( Arena* [all...] |
/external/protobuf/ruby/ext/google/protobuf_c/ |
H A D | upb.h | 86 class Arena; 539 /* upb::Arena *****************************************************************/ 541 /* upb::Arena is a specific allocator implementation that uses arena allocation. 544 * to be freed. However the Arena does allow users to register cleanup 547 * A upb::Arena is *not* thread-safe. 552 UPB_DECLARE_TYPE(upb::Arena, upb_arena) 573 class upb::Arena { class in class:upb 576 Arena() { upb_arena_init(this); } function in class:upb::Arena 579 * with the given allocator. The given allocator must outlive the Arena. 584 Arena(voi function in class:upb::Arena [all...] |