/external/jemalloc/src/ |
H A D | valgrind.c | 9 valgrind_make_mem_noaccess(void *ptr, size_t usize) argument 12 VALGRIND_MAKE_MEM_NOACCESS(ptr, usize); 16 valgrind_make_mem_undefined(void *ptr, size_t usize) argument 19 VALGRIND_MAKE_MEM_UNDEFINED(ptr, usize); 23 valgrind_make_mem_defined(void *ptr, size_t usize) argument 26 VALGRIND_MAKE_MEM_DEFINED(ptr, usize); 30 valgrind_freelike_block(void *ptr, size_t usize) argument 33 VALGRIND_FREELIKE_BLOCK(ptr, usize);
|
H A D | base.c | 82 size_t csize, usize; local 92 usize = s2u(csize); 93 extent_node_init(&key, NULL, NULL, usize, false, false);
|
H A D | quarantine.c | 102 assert(obj->usize == isalloc(obj->ptr, config_prof)); 104 quarantine->curbytes -= obj->usize; 122 size_t usize = isalloc(ptr, config_prof); local 135 if (quarantine->curbytes + usize > opt_quarantine) { 136 size_t upper_bound = (opt_quarantine >= usize) ? opt_quarantine 137 - usize : 0; 146 if (quarantine->curbytes + usize <= opt_quarantine) { 151 obj->usize = usize; 152 quarantine->curbytes += usize; [all...] |
H A D | ckh.c | 265 size_t usize; local 268 usize = sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); 269 if (unlikely(usize == 0 || usize > HUGE_MAXCLASS)) { 273 tab = (ckhc_t *)ipallocztm(tsd, usize, CACHELINE, true, NULL, 305 size_t usize; local 314 usize = sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); 315 if (unlikely(usize == 0 || usize > HUGE_MAXCLASS)) 317 tab = (ckhc_t *)ipallocztm(tsd, usize, CACHELIN 354 size_t mincells, usize; local [all...] |
H A D | huge.c | 34 huge_malloc(tsd_t *tsd, arena_t *arena, size_t usize, bool zero, argument 38 assert(usize == s2u(usize)); 40 return (huge_palloc(tsd, arena, usize, chunksize, zero, tcache)); 44 huge_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, argument 54 ausize = sa2u(usize, alignment); 83 usize, alignment, &is_zeroed)) == NULL) { 88 extent_node_init(node, arena, ret, usize, is_zeroed, true); 91 arena_chunk_dalloc_huge(arena, ret, usize); 104 memset(ret, 0, usize); 117 huge_dalloc_junk(void *ptr, size_t usize) argument 139 size_t usize, usize_next; local 195 huge_ralloc_no_move_shrink(void *ptr, size_t oldsize, size_t usize) argument 245 huge_ralloc_no_move_expand(void *ptr, size_t oldsize, size_t usize, bool zero) argument 341 huge_ralloc_move_helper(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 351 huge_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument [all...] |
H A D | jemalloc.c | 1419 imalloc_prof_sample(tsd_t *tsd, size_t usize, szind_t ind, argument 1426 if (usize <= SMALL_MAXCLASS) { 1431 arena_prof_promoted(p, usize); 1433 p = imalloc(tsd, usize, ind, slow_path); 1439 imalloc_prof(tsd_t *tsd, size_t usize, szind_t ind, bool slow_path) argument 1444 tctx = prof_alloc_prep(tsd, usize, prof_active_get_unlocked(), true); 1446 p = imalloc_prof_sample(tsd, usize, ind, tctx, slow_path); 1448 p = imalloc(tsd, usize, ind, slow_path); 1453 prof_malloc(p, usize, tctx); 1459 imalloc_body(size_t size, tsd_t **tsd, size_t *usize, boo argument 1483 imalloc_post_check(void *ret, tsd_t *tsd, size_t usize, bool slow_path) argument 1529 imemalign_prof_sample(tsd_t *tsd, size_t alignment, size_t usize, prof_tctx_t *tctx) argument 1549 imemalign_prof(tsd_t *tsd, size_t alignment, size_t usize) argument 1574 size_t usize; local 1662 icalloc_prof_sample(tsd_t *tsd, size_t usize, szind_t ind, prof_tctx_t *tctx) argument 1681 icalloc_prof(tsd_t *tsd, size_t usize, szind_t ind) argument 1771 irealloc_prof_sample(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, prof_tctx_t *tctx) argument 1790 irealloc_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize) argument 1816 size_t usize; local 1841 isfree(tsd_t *tsd, void *ptr, size_t usize, tcache_t *tcache) argument 2012 imallocx_flags_decode_hard(tsd_t *tsd, size_t size, int flags, size_t *usize, size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena) argument 2044 imallocx_flags_decode(tsd_t *tsd, size_t size, int flags, size_t *usize, size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena) argument 2064 imallocx_flags(tsd_t *tsd, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena) argument 2079 imallocx_prof_sample(tsd_t *tsd, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena) argument 2099 imallocx_prof(tsd_t *tsd, size_t size, int flags, size_t *usize) argument 2130 imallocx_no_prof(tsd_t *tsd, size_t size, int flags, size_t *usize) argument 2165 size_t usize; local 2197 irallocx_prof_sample(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, prof_tctx_t *tctx) argument 2220 irallocx_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t size, size_t alignment, size_t *usize, bool zero, tcache_t *tcache, arena_t *arena) argument 2267 size_t usize; local 2340 size_t usize; local 2353 size_t usize; local 2367 size_t usize_max, usize; local 2417 size_t usize, old_usize; local 2475 size_t usize; local 2513 size_t usize; local 2527 size_t usize; local 2551 size_t usize; local [all...] |
H A D | prof.c | 226 prof_malloc_sample_object(const void *ptr, size_t usize, prof_tctx_t *tctx) argument 229 prof_tctx_set(ptr, usize, tctx); 233 tctx->cnts.curbytes += usize; 236 tctx->cnts.accumbytes += usize; 243 prof_free_sampled_object(tsd_t *tsd, size_t usize, prof_tctx_t *tctx) argument 248 assert(tctx->cnts.curbytes >= usize); 250 tctx->cnts.curbytes -= usize;
|
H A D | arena.c | 799 arena_huge_malloc_stats_update(arena_t *arena, size_t usize) argument 801 szind_t index = size2index(usize) - nlclasses - NBINS; 806 arena->stats.allocated_huge += usize; 812 arena_huge_malloc_stats_update_undo(arena_t *arena, size_t usize) argument 814 szind_t index = size2index(usize) - nlclasses - NBINS; 819 arena->stats.allocated_huge -= usize; 825 arena_huge_dalloc_stats_update(arena_t *arena, size_t usize) argument 827 szind_t index = size2index(usize) - nlclasses - NBINS; 832 arena->stats.allocated_huge -= usize; 838 arena_huge_dalloc_stats_update_undo(arena_t *arena, size_t usize) argument 851 arena_huge_ralloc_stats_update(arena_t *arena, size_t oldsize, size_t usize) argument 859 arena_huge_ralloc_stats_update_undo(arena_t *arena, size_t oldsize, size_t usize) argument 894 arena_chunk_alloc_huge_hard(arena_t *arena, chunk_hooks_t *chunk_hooks, size_t usize, size_t alignment, bool *zero, size_t csize) argument 917 arena_chunk_alloc_huge(arena_t *arena, size_t usize, size_t alignment, bool *zero) argument 945 arena_chunk_dalloc_huge(arena_t *arena, void *chunk, size_t usize) argument 963 arena_chunk_ralloc_huge_similar(arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument 981 arena_chunk_ralloc_huge_shrink(arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument 1006 arena_chunk_ralloc_huge_expand_hard(arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t oldsize, size_t usize, bool *zero, void *nchunk, size_t udiff, size_t cdiff) argument 1035 arena_chunk_ralloc_huge_expand(arena_t *arena, void *chunk, size_t oldsize, size_t usize, bool *zero) argument 2278 arena_redzone_corruption(void *ptr, size_t usize, bool after, size_t offset, uint8_t byte) argument 2350 arena_quarantine_junk_small(void *ptr, size_t usize) argument 2369 size_t usize; local 2422 size_t usize; local 2498 arena_palloc_large(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 2584 arena_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 2786 arena_dalloc_junk_large(void *ptr, size_t usize) argument 2808 size_t usize = arena_mapbits_large_size_get(chunk, pageind) - local 2904 size_t usize, splitsize, size, flag_dirty, flag_unzeroed_mask; local 2985 arena_ralloc_junk_large(void *ptr, size_t old_usize, size_t usize) argument 3087 arena_ralloc_move_helper(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 3105 size_t usize; local [all...] |
/external/jemalloc/include/jemalloc/internal/ |
H A D | quarantine.h | 16 size_t usize; member in struct:quarantine_obj_s
|
H A D | jemalloc_internal.h | 628 size_t usize = grp_size + mod_size; local 629 return (usize); 669 size_t usize = (size + delta_mask) & ~delta_mask; 670 return (usize); 704 size_t usize; 724 usize = s2u(ALIGNMENT_CEILING(size, alignment)); 725 if (usize < LARGE_MINCLASS) 726 return (usize); 738 usize = (size <= LARGE_MINCLASS) ? LARGE_MINCLASS : s2u(size); 744 if (usize [all...] |
H A D | prof.h | 284 void prof_malloc_sample_object(const void *ptr, size_t usize, 286 void prof_free_sampled_object(tsd_t *tsd, size_t usize, prof_tctx_t *tctx); 332 bool prof_sample_accum_update(tsd_t *tsd, size_t usize, bool commit, 334 prof_tctx_t *prof_alloc_prep(tsd_t *tsd, size_t usize, bool prof_active, 337 void prof_tctx_set(const void *ptr, size_t usize, prof_tctx_t *tctx); 338 void prof_tctx_reset(const void *ptr, size_t usize, const void *old_ptr, 340 void prof_malloc_sample_object(const void *ptr, size_t usize, 342 void prof_malloc(const void *ptr, size_t usize, prof_tctx_t *tctx); 343 void prof_realloc(tsd_t *tsd, const void *ptr, size_t usize, 346 void prof_free(tsd_t *tsd, const void *ptr, size_t usize); 410 prof_tctx_set(const void *ptr, size_t usize, prof_tctx_t *tctx) argument 420 prof_tctx_reset(const void *ptr, size_t usize, const void *old_ptr, prof_tctx_t *old_tctx) argument 431 prof_sample_accum_update(tsd_t *tsd, size_t usize, bool update, prof_tdata_t **tdata_out) argument 461 prof_alloc_prep(tsd_t *tsd, size_t usize, bool prof_active, bool update) argument 482 prof_malloc(const void *ptr, size_t usize, prof_tctx_t *tctx) argument 496 prof_realloc(tsd_t *tsd, const void *ptr, size_t usize, prof_tctx_t *tctx, bool prof_active, bool updated, const void *old_ptr, size_t old_usize, prof_tctx_t *old_tctx) argument 532 prof_free(tsd_t *tsd, const void *ptr, size_t usize) argument [all...] |
/external/jemalloc/test/integration/ |
H A D | allocated.c | 18 size_t sz, usize; local 69 usize = malloc_usable_size(p); 70 assert_u64_le(a0 + usize, a1, 86 assert_u64_le(d0 + usize, d1,
|
/external/jemalloc/test/unit/ |
H A D | quarantine.c | 62 arena_redzone_corruption_replacement(void *ptr, size_t usize, bool after, argument
|
H A D | junk.c | 41 arena_dalloc_junk_large_intercept(void *ptr, size_t usize) argument 45 arena_dalloc_junk_large_orig(ptr, usize); 46 for (i = 0; i < usize; i++) { 49 i, usize); 56 huge_dalloc_junk_intercept(void *ptr, size_t usize) argument 59 huge_dalloc_junk_orig(ptr, usize); 171 arena_ralloc_junk_large_intercept(void *ptr, size_t old_usize, size_t usize) argument 174 arena_ralloc_junk_large_orig(ptr, old_usize, usize); 176 assert_zu_eq(usize, shrink_size(large_maxclass), "Unexpected usize"); 203 arena_redzone_corruption_replacement(void *ptr, size_t usize, bool after, size_t offset, uint8_t byte) argument [all...] |
/external/tpm2/ |
H A D | TpmFail.c | 116 UINT32 usize; local 119 || UINT32_Unmarshal(&usize, buffer, size) != TPM_RC_SUCCESS 123 header->size = usize;
|
/external/v8/src/runtime/ |
H A D | runtime-atomics.cc | 580 uint32_t usize = NumberToUint32(*size); local 581 return isolate->heap()->ToBoolean(AtomicIsLockFree(usize));
|
/external/opencv3/modules/core/src/ |
H A D | arithm.cpp | 5566 Size ssize = _src.size(), lsize = _lowerb.size(), usize = _upperb.size(); local 5582 ssize != usize || stype != utype )
|
/external/elfutils/src/ |
H A D | readelf.c | 4354 uint8_t usize = *(uint8_t *) data++; local 4355 NEED (usize); 4358 print_block (usize, data); 4359 data += usize; 4385 usize = *(uint8_t *) data++; 4392 op_name, usize, uleb);
|