Searched refs:arena (Results 1 - 25 of 47) sorted by relevance

12

/external/valgrind/VEX/unused/
H A Darena.h2 /* This is a modified version of the file "arena.h" from
28 /* $Id: H:/drh/idioms/book/RCS/arena.doc,v 1.10 1997/02/21 19:45:19 drh Exp $ */
40 extern void *Arena_alloc (T arena, long nbytes,
42 extern void *Arena_calloc(T arena, long count,
44 extern void Arena_free (T arena);
/external/jemalloc/src/
H A Darena.c40 static void arena_purge_to_limit(arena_t *arena, size_t ndirty_limit);
41 static void arena_run_dalloc(arena_t *arena, arena_run_t *run, bool dirty,
43 static void arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk,
45 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk,
203 arena_runs_avail_get(arena_t *arena, szind_t ind) argument
209 return (&arena->runs_avail[ind - runs_avail_bias]);
213 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, argument
220 arena_run_tree_insert(arena_runs_avail_get(arena, ind),
225 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, argument
232 arena_run_tree_remove(arena_runs_avail_get(arena, in
237 arena_run_dirty_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument
254 arena_run_dirty_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages) argument
278 arena_chunk_cache_maybe_insert(arena_t *arena, extent_node_t *node, bool cache) argument
290 arena_chunk_cache_maybe_remove(arena_t *arena, extent_node_t *node, bool dirty) argument
376 arena_nactive_add(arena_t *arena, size_t add_pages) argument
390 arena_nactive_sub(arena_t *arena, size_t sub_pages) argument
403 arena_run_split_remove(arena_t *arena, arena_chunk_t *chunk, size_t run_ind, size_t flag_dirty, size_t flag_decommitted, size_t need_pages) argument
445 arena_run_split_large_helper(arena_t *arena, arena_run_t *run, size_t size, bool remove, bool zero) argument
518 arena_run_split_large(arena_t *arena, arena_run_t *run, size_t size, bool zero) argument
525 arena_run_init_large(arena_t *arena, arena_run_t *run, size_t size, bool zero) argument
532 arena_run_split_small(arena_t *arena, arena_run_t *run, size_t size, szind_t binind) argument
570 arena_chunk_init_spare(arena_t *arena) argument
592 arena_chunk_register(arena_t *arena, arena_chunk_t *chunk, bool zero) argument
607 arena_chunk_alloc_internal_hard(arena_t *arena, chunk_hooks_t *chunk_hooks, bool *zero, bool *commit) argument
641 arena_chunk_alloc_internal(arena_t *arena, bool *zero, bool *commit) argument
670 arena_chunk_init_hard(arena_t *arena) argument
724 arena_chunk_alloc(arena_t *arena) argument
742 arena_chunk_dalloc(arena_t *arena, arena_chunk_t *chunk) argument
799 arena_huge_malloc_stats_update(arena_t *arena, size_t usize) argument
812 arena_huge_malloc_stats_update_undo(arena_t *arena, size_t usize) argument
825 arena_huge_dalloc_stats_update(arena_t *arena, size_t usize) argument
838 arena_huge_dalloc_stats_update_undo(arena_t *arena, size_t usize) argument
851 arena_huge_ralloc_stats_update(arena_t *arena, size_t oldsize, size_t usize) argument
859 arena_huge_ralloc_stats_update_undo(arena_t *arena, size_t oldsize, size_t usize) argument
868 arena_node_alloc(arena_t *arena) argument
884 arena_node_dalloc(arena_t *arena, extent_node_t *node) argument
894 arena_chunk_alloc_huge_hard(arena_t *arena, chunk_hooks_t *chunk_hooks, size_t usize, size_t alignment, bool *zero, size_t csize) argument
917 arena_chunk_alloc_huge(arena_t *arena, size_t usize, size_t alignment, bool *zero) argument
945 arena_chunk_dalloc_huge(arena_t *arena, void *chunk, size_t usize) argument
963 arena_chunk_ralloc_huge_similar(arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument
981 arena_chunk_ralloc_huge_shrink(arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument
1006 arena_chunk_ralloc_huge_expand_hard(arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t oldsize, size_t usize, bool *zero, void *nchunk, size_t udiff, size_t cdiff) argument
1035 arena_chunk_ralloc_huge_expand(arena_t *arena, void *chunk, size_t oldsize, size_t usize, bool *zero) argument
1076 arena_run_first_best_fit(arena_t *arena, size_t size) argument
1092 arena_run_alloc_large_helper(arena_t *arena, size_t size, bool zero) argument
1103 arena_run_alloc_large(arena_t *arena, size_t size, bool zero) argument
1136 arena_run_alloc_small_helper(arena_t *arena, size_t size, szind_t binind) argument
1147 arena_run_alloc_small(arena_t *arena, size_t size, szind_t binind) argument
1189 arena_lg_dirty_mult_get(arena_t *arena) argument
1201 arena_lg_dirty_mult_set(arena_t *arena, ssize_t lg_dirty_mult) argument
1216 arena_decay_deadline_init(arena_t *arena) argument
1237 arena_decay_deadline_reached(const arena_t *arena, const nstime_t *time) argument
1246 arena_decay_backlog_npages_limit(const arena_t *arena) argument
1274 arena_decay_epoch_advance(arena_t *arena, const nstime_t *time) argument
1317 arena_decay_npages_limit(arena_t *arena) argument
1333 arena_decay_init(arena_t *arena, ssize_t decay_time) argument
1359 arena_decay_time_get(arena_t *arena) argument
1371 arena_decay_time_set(arena_t *arena, ssize_t decay_time) argument
1394 arena_maybe_purge_ratio(arena_t *arena) argument
1422 arena_maybe_purge_decay(arena_t *arena) argument
1457 arena_maybe_purge(arena_t *arena) argument
1471 arena_dirty_count(arena_t *arena) argument
1505 arena_stash_dirty(arena_t *arena, chunk_hooks_t *chunk_hooks, size_t ndirty_limit, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument
1595 arena_purge_stashed(arena_t *arena, chunk_hooks_t *chunk_hooks, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument
1688 arena_unstash_purged(arena_t *arena, chunk_hooks_t *chunk_hooks, arena_runs_dirty_link_t *purge_runs_sentinel, extent_node_t *purge_chunks_sentinel) argument
1737 arena_purge_to_limit(arena_t *arena, size_t ndirty_limit) argument
1778 arena_purge(arena_t *arena, bool all) argument
1790 arena_run_coalesce(arena_t *arena, arena_chunk_t *chunk, size_t *p_size, size_t *p_run_ind, size_t *p_run_pages, size_t flag_dirty, size_t flag_decommitted) argument
1882 arena_run_size_get(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t run_ind) argument
1903 arena_run_dalloc(arena_t *arena, arena_run_t *run, bool dirty, bool cleaned, bool decommitted) argument
1979 arena_run_trim_head(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize) argument
2019 arena_run_trim_tail(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize, bool dirty) argument
2106 arena_bin_nonfull_run_get(arena_t *arena, arena_bin_t *bin) argument
2157 arena_bin_malloc_hard(arena_t *arena, arena_bin_t *bin) argument
2206 arena_tcache_fill_small(tsd_t *tsd, arena_t *arena, tcache_bin_t *tbin, szind_t binind, uint64_t prof_accumbytes) argument
2365 arena_malloc_small(tsd_t *tsd, arena_t *arena, szind_t binind, bool zero) argument
2419 arena_malloc_large(tsd_t *tsd, arena_t *arena, szind_t binind, bool zero) argument
2481 arena_malloc_hard(tsd_t *tsd, arena_t *arena, size_t size, szind_t ind, bool zero, tcache_t *tcache) argument
2498 arena_palloc_large(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero) argument
2584 arena_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument
2668 arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument
2688 arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument
2709 arena_dalloc_bin_locked_impl(arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm, bool junked) argument
2742 arena_dalloc_bin_junked_locked(arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_bits_t *bitselm) argument
2750 arena_dalloc_bin(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind, arena_chunk_map_bits_t *bitselm) argument
2766 arena_dalloc_small(tsd_t *tsd, arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind) argument
2800 arena_dalloc_large_locked_impl(arena_t *arena, arena_chunk_t *chunk, void *ptr, bool junked) argument
2827 arena_dalloc_large_junked_locked(arena_t *arena, arena_chunk_t *chunk, void *ptr) argument
2835 arena_dalloc_large(tsd_t *tsd, arena_t *arena, arena_chunk_t *chunk, void *ptr) argument
2845 arena_ralloc_large_shrink(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size) argument
2881 arena_ralloc_large_grow(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t usize_min, size_t usize_max, bool zero) argument
3009 arena_t *arena; local
3087 arena_ralloc_move_helper(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument
3101 arena_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize, size_t size, size_t alignment, bool zero, tcache_t *tcache) argument
3145 arena_dss_prec_get(arena_t *arena) argument
3156 arena_dss_prec_set(arena_t *arena, dss_prec_t dss_prec) argument
3206 arena_basic_stats_merge_locked(arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty) argument
3220 arena_basic_stats_merge(arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty) argument
3232 arena_stats_merge(arena_t *arena, unsigned *nthreads, const char **dss, ssize_t *lg_dirty_mult, ssize_t *decay_time, size_t *nactive, size_t *ndirty, arena_stats_t *astats, malloc_bin_stats_t *bstats, malloc_large_stats_t *lstats, malloc_huge_stats_t *hstats) argument
3293 arena_nthreads_get(arena_t *arena) argument
3300 arena_nthreads_inc(arena_t *arena) argument
3307 arena_nthreads_dec(arena_t *arena) argument
3316 arena_t *arena; local
3649 arena_prefork(arena_t *arena) argument
3662 arena_postfork_parent(arena_t *arena) argument
3675 arena_postfork_child(arena_t *arena) argument
[all...]
H A Dhuge.c34 huge_malloc(tsd_t *tsd, arena_t *arena, size_t usize, bool zero, argument
40 return (huge_palloc(tsd, arena, usize, chunksize, zero, tcache));
44 huge_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, argument
61 CACHELINE, false, tcache, true, arena);
72 /* On 32 bit systems, using a per arena cache can exhaust
74 * always take place in the first arena.
77 arena = a0get();
79 arena = arena_choose(tsd, arena);
82 if (unlikely(arena
141 arena_t *arena; local
198 arena_t *arena; local
247 arena_t *arena; local
341 huge_ralloc_move_helper(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument
351 huge_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument
384 arena_t *arena; local
414 arena_t *arena; local
430 arena_t *arena; local
445 arena_t *arena; local
[all...]
H A Dchunk.c52 static void chunk_record(arena_t *arena, chunk_hooks_t *chunk_hooks,
59 chunk_hooks_get_locked(arena_t *arena) argument
62 return (arena->chunk_hooks);
66 chunk_hooks_get(arena_t *arena) argument
70 malloc_mutex_lock(&arena->chunks_mtx);
71 chunk_hooks = chunk_hooks_get_locked(arena);
72 malloc_mutex_unlock(&arena->chunks_mtx);
78 chunk_hooks_set(arena_t *arena, const chunk_hooks_t *chunk_hooks) argument
82 malloc_mutex_lock(&arena->chunks_mtx);
83 old_chunk_hooks = arena
113 chunk_hooks_assure_initialized_impl(arena_t *arena, chunk_hooks_t *chunk_hooks, bool locked) argument
127 chunk_hooks_assure_initialized_locked(arena_t *arena, chunk_hooks_t *chunk_hooks) argument
135 chunk_hooks_assure_initialized(arena_t *arena, chunk_hooks_t *chunk_hooks) argument
188 chunk_first_best_fit(arena_t *arena, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, size_t size) argument
200 chunk_recycle(arena_t *arena, chunk_hooks_t *chunk_hooks, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, bool cache, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, bool dalloc_node) argument
331 chunk_alloc_core(arena_t *arena, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit, dss_prec_t dss_prec) argument
383 chunk_alloc_cache(arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool dalloc_node) argument
409 arena_t *arena; local
425 arena_t *arena; local
439 chunk_alloc_retained(arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument
454 chunk_alloc_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks, void *new_addr, size_t size, size_t alignment, bool *zero, bool *commit) argument
476 chunk_record(arena_t *arena, chunk_hooks_t *chunk_hooks, extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, bool cache, void *chunk, size_t size, bool zeroed, bool committed) argument
568 chunk_dalloc_cache(arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool committed) argument
593 chunk_dalloc_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, bool zeroed, bool committed) argument
651 chunk_purge_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t size, size_t offset, size_t length) argument
[all...]
H A Dtcache.c73 tcache_alloc_small_hard(tsd_t *tsd, arena_t *arena, tcache_t *tcache, argument
78 arena_tcache_fill_small(tsd, arena, tbin, binind, config_prof ?
91 arena_t *arena; local
99 arena = arena_choose(tsd, NULL);
100 assert(arena != NULL);
102 /* Lock the arena bin associated with the first object. */
108 if (config_prof && bin_arena == arena) {
109 if (arena_prof_accum(arena, tcache->prof_accumbytes))
115 if (config_stats && bin_arena == arena) {
137 * arena bi
172 arena_t *arena; local
253 tcache_arena_associate(tcache_t *tcache, arena_t *arena) argument
274 tcache_arena_dissociate(tcache_t *tcache, arena_t *arena) argument
300 arena_t *arena; local
314 tcache_create(tsd_t *tsd, arena_t *arena) argument
356 arena_t *arena; local
417 tcache_stats_merge(tcache_t *tcache, arena_t *arena) argument
[all...]
H A Djemalloc.c387 arena_set(unsigned ind, arena_t *arena) argument
390 atomic_write_p((void **)&arenas[ind], arena);
414 /* Create a new arena and insert it into the arenas array at index ind. */
418 arena_t *arena; local
428 * auto arena.
430 arena = arena_get(ind, false);
431 if (arena != NULL) {
433 return (arena);
436 /* Actually initialize the arena. */
437 arena
445 arena_t *arena; local
456 arena_t *arena; local
480 arena_t *arena; local
643 arena_t *arena; local
694 arena_t *arena = arena_get(i, false); local
2012 imallocx_flags_decode_hard(tsd_t *tsd, size_t size, int flags, size_t *usize, size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena) argument
2044 imallocx_flags_decode(tsd_t *tsd, size_t size, int flags, size_t *usize, size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena) argument
2064 imallocx_flags(tsd_t *tsd, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena) argument
2079 imallocx_prof_sample(tsd_t *tsd, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena) argument
2105 arena_t *arena; local
2136 arena_t *arena; local
2197 irallocx_prof_sample(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, prof_tctx_t *tctx) argument
2220 irallocx_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t size, size_t alignment, size_t *usize, bool zero, tcache_t *tcache, arena_t *arena) argument
2272 arena_t *arena; local
2672 arena_t *arena; local
2701 arena_t *arena; local
2722 arena_t *arena; local
[all...]
H A Dchunk_dss.c69 chunk_alloc_dss(arena_t *arena, void *new_addr, size_t size, size_t alignment, argument
137 chunk_dalloc_wrapper(arena,
/external/jemalloc/include/jemalloc/internal/
H A Dchunk.h55 chunk_hooks_t chunk_hooks_get(arena_t *arena);
56 chunk_hooks_t chunk_hooks_set(arena_t *arena,
62 void *chunk_alloc_cache(arena_t *arena, chunk_hooks_t *chunk_hooks,
65 void *chunk_alloc_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks,
67 void chunk_dalloc_cache(arena_t *arena, chunk_hooks_t *chunk_hooks,
69 void chunk_dalloc_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks,
71 bool chunk_purge_wrapper(arena_t *arena, chunk_hooks_t *chunk_hooks,
H A Dhuge.h12 void *huge_malloc(tsd_t *tsd, arena_t *arena, size_t usize, bool zero,
14 void *huge_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment,
18 void *huge_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize,
H A Darena.h17 * The minimum ratio of active:dirty pages per arena is computed as:
191 * A pointer to the arena that owns the chunk is stored within the node.
209 * than one per arena), but mainly to avoid false cacheline sharing.
268 * locked. Run allocation/deallocation are protected by the arena lock,
294 /* This arena's index within the arenas array. */
298 * Number of threads currently assigned to this arena. This field is
304 * There are three classes of arena operations from a locking
314 * List of tcaches for extant threads associated with this arena.
331 * In order to avoid rapid chunk allocation/deallocation when an arena
333 * recently freed chunk. The spare is left in the arena'
[all...]
H A Dchunk_dss.h26 void *chunk_alloc_dss(arena_t *arena, void *new_addr, size_t size,
H A Dtcache.h147 void *tcache_alloc_small_hard(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
153 void tcache_arena_associate(tcache_t *tcache, arena_t *arena);
156 void tcache_arena_dissociate(tcache_t *tcache, arena_t *arena);
158 tcache_t *tcache_create(tsd_t *tsd, arena_t *arena);
161 void tcache_stats_merge(tcache_t *tcache, arena_t *arena);
178 void *tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
180 void *tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache,
294 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, argument
308 arena = arena_choose(tsd, arena);
[all...]
H A Dextent.h36 * don't return arena chunks.
43 /* Linkage for arena's runs_dirty and chunks_cache rings. */
51 /* Linkage for arena's huge and node_cache lists. */
80 void extent_node_arena_set(extent_node_t *node, arena_t *arena);
87 void extent_node_init(extent_node_t *node, arena_t *arena, void *addr,
147 extent_node_arena_set(extent_node_t *node, arena_t *arena) argument
150 node->en_arena = arena;
196 extent_node_init(extent_node_t *node, arena_t *arena, void *addr, size_t size, argument
200 extent_node_arena_set(node, arena);
H A Djemalloc_internal.h170 * jemalloc can conceptually be broken into components (arena, tcache, etc.),
193 * a: arena
374 #include "jemalloc/internal/arena.h"
405 #include "jemalloc/internal/arena.h"
409 #include "jemalloc/internal/arena.h"
497 #include "jemalloc/internal/arena.h"
545 arena_t *arena_choose(tsd_t *tsd, arena_t *arena);
781 /* Choose an arena based on a per-thread value. */
783 arena_choose(tsd_t *tsd, arena_t *arena)
787 if (arena !
[all...]
/external/jemalloc/test/unit/
H A Dstats.c77 unsigned arena; local
85 arena = 0;
86 assert_d_eq(mallctl("thread.arena", NULL, NULL, &arena, sizeof(arena)),
100 assert_d_eq(mallctl("arena.0.purge", NULL, NULL, NULL, 0), 0,
144 unsigned arena; local
152 arena = 0;
153 assert_d_eq(mallctl("thread.arena", NULL, NULL, &arena, sizeo
193 unsigned arena; local
237 unsigned arena; local
277 unsigned arena; local
351 unsigned arena; local
395 unsigned arena; local
[all...]
/external/valgrind/coregrind/
H A Dpub_core_mallocfree.h51 When adding a new arena, remember also to add it to ensure_mm_init().
94 int arena; /* total space allocated from system */ member in struct:vg_mallinfo
106 extern void* VG_(arena_malloc) ( ArenaId arena, const HChar* cc, SizeT nbytes );
107 extern void VG_(arena_free) ( ArenaId arena, void* ptr );
108 extern void* VG_(arena_calloc) ( ArenaId arena, const HChar* cc,
110 extern void* VG_(arena_realloc) ( ArenaId arena, const HChar* cc,
150 const HChar* name; // arena name, !NULL if Addr a points in an arena.
153 Bool free; // True if this is in the arena free zone.
/external/libchrome/crypto/third_party/nss/
H A Dsecsign.cc54 SECStatus DerSignData(PLArenaPool *arena, argument
60 return SEC_DerSignData(arena, result, input->data, input->len, key,
115 rv = SECOID_SetAlgorithmID(arena, &sd.signatureAlgorithm, algo_id, 0);
127 arena, result, &sd, SEC_ASN1_GET(CERT_SignedDataTemplate));
H A Dchromium-nss.h73 CRYPTO_EXPORT SECStatus DerSignData(PLArenaPool *arena,
/external/valgrind/memcheck/tests/
H A Dsbfragment.c22 printf("%10d int arena; /* non-mmapped space allocated from system */\n", mallinfo_result.arena);
79 // We consider that an arena of up to 3 times more than bigsize is ok.
84 // Under valgrind, hblkhd is 0 : all the space is in arena.
88 if (mallinfo_result.arena + mallinfo_result.hblkhd > 3 * bigsize)
90 (unsigned long) mallinfo_result.arena
H A Dmallinfo.c21 printf("arena = %d\n", mi.arena); /* non-mmapped space allocated from system */
64 // arena should be reasonably close to fordblks + uordblks
65 if (mi.arena < mi.fordblks + mi.uordblks)
68 if (mi.arena/5*4 > mi.fordblks + mi.uordblks)
/external/regex-re2/util/
H A Darena.h16 // #include "lib/arena-inl.h"
18 // UnsafeArena arena(1000);
19 // Foo* foo = new (AllocateInArena, &arena) Foo;
89 // Operators for allocation on the arena
90 // Syntax: new (AllocateInArena, arena) MyClass;
98 re2::UnsafeArena *arena) {
99 return reinterpret_cast<char*>(arena->GetMemory(size, 1));
96 operator new(size_t size, re2::AllocateInArenaType , re2::UnsafeArena *arena) argument
/external/libchrome/base/trace_event/
H A Dmalloc_dump_provider.cc67 DCHECK_GE(info.arena + info.hblkhd, info.uordblks);
69 // In case of Android's jemalloc |arena| is 0 and the outer pages size is
71 // |arena| + |hblkhd|. For more details see link: http://goo.gl/fMR8lF.
72 total_virtual_size = info.arena + info.hblkhd;
/external/libchrome/crypto/
H A Dsignature_verifier_nss.cc93 PLArenaPool* arena = PORT_NewArena(DER_DEFAULT_CHUNKSIZE);
94 if (!arena) {
105 rv = SEC_QuickDERDecodeItem(arena, &sig_alg_id,
110 PORT_FreeArena(arena, PR_TRUE);
123 PORT_FreeArena(arena, PR_TRUE); // Done with sig_alg_id.
H A Dnss_key_util.cc91 ScopedPLArenaPool arena(PORT_NewArena(DER_DEFAULT_CHUNKSIZE));
92 DCHECK(arena);
101 SEC_QuickDERDecodeItem(arena.get(), &der_private_key_info,
/external/valgrind/gdbserver_tests/
H A Dhginfo.stderrB.exp19 Address 0x........ is 0 bytes inside an unallocated block of size 1,008 in arena "client"

Completed in 1015 milliseconds

12