/external/jemalloc/src/ |
H A D | valgrind.c | 9 valgrind_make_mem_noaccess(void *ptr, size_t usize) argument 12 VALGRIND_MAKE_MEM_NOACCESS(ptr, usize); 16 valgrind_make_mem_undefined(void *ptr, size_t usize) argument 19 VALGRIND_MAKE_MEM_UNDEFINED(ptr, usize); 23 valgrind_make_mem_defined(void *ptr, size_t usize) argument 26 VALGRIND_MAKE_MEM_DEFINED(ptr, usize); 30 valgrind_freelike_block(void *ptr, size_t usize) argument 33 VALGRIND_FREELIKE_BLOCK(ptr, usize);
|
H A D | quarantine.c | 100 assert(obj->usize == isalloc(tsdn, obj->ptr, config_prof)); 102 quarantine->curbytes -= obj->usize; 120 size_t usize = isalloc(tsd_tsdn(tsd), ptr, config_prof); local 133 if (quarantine->curbytes + usize > opt_quarantine) { 134 size_t upper_bound = (opt_quarantine >= usize) ? opt_quarantine 135 - usize : 0; 144 if (quarantine->curbytes + usize <= opt_quarantine) { 149 obj->usize = usize; 150 quarantine->curbytes += usize; [all...] |
H A D | base.c | 93 size_t csize, usize; local 103 usize = s2u(csize); 104 extent_node_init(&key, NULL, NULL, usize, 0, false, false);
|
H A D | ckh.c | 266 size_t usize; local 269 usize = sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); 270 if (unlikely(usize == 0 || usize > HUGE_MAXCLASS)) { 274 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELINE, 306 size_t usize; local 315 usize = sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); 316 if (unlikely(usize == 0 || usize > HUGE_MAXCLASS)) 318 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELIN 355 size_t mincells, usize; local [all...] |
H A D | huge.c | 43 huge_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) argument 46 assert(usize == s2u(usize)); 48 return (huge_palloc(tsdn, arena, usize, chunksize, zero)); 52 huge_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, argument 66 ausize = sa2u(usize, alignment); 99 arena, usize, alignment, &sn, &is_zeroed)) == NULL) { 104 extent_node_init(node, arena, ret, usize, sn, is_zeroed, true); 107 arena_chunk_dalloc_huge(tsdn, arena, ret, usize, sn); 120 memset(ret, 0, usize); 133 huge_dalloc_junk(void *ptr, size_t usize) argument 155 size_t usize, usize_next; local 215 huge_ralloc_no_move_shrink(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t usize) argument 269 huge_ralloc_no_move_expand(tsdn_t *tsdn, void *ptr, size_t oldsize, size_t usize, bool zero) argument 369 huge_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 379 huge_ralloc(tsd_t *tsd, arena_t *arena, void *ptr, size_t oldsize, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument [all...] |
H A D | jemalloc.c | 1537 ialloc_prof_sample(tsd_t *tsd, size_t usize, szind_t ind, bool zero, argument 1544 if (usize <= SMALL_MAXCLASS) { 1549 arena_prof_promoted(tsd_tsdn(tsd), p, usize); 1551 p = ialloc(tsd, usize, ind, zero, slow_path); 1557 ialloc_prof(tsd_t *tsd, size_t usize, szind_t ind, bool zero, bool slow_path) argument 1562 tctx = prof_alloc_prep(tsd, usize, prof_active_get_unlocked(), true); 1564 p = ialloc_prof_sample(tsd, usize, ind, zero, tctx, slow_path); 1566 p = ialloc(tsd, usize, ind, zero, slow_path); 1571 prof_malloc(tsd_tsdn(tsd), p, usize, tctx); local 1583 ialloc_body(size_t size, bool zero, tsdn_t **tsdn, size_t *usize, argument 1615 ialloc_post_check(void *ret, tsdn_t *tsdn, size_t usize, const char *func, bool update_errno, bool slow_path) argument 1663 imemalign_prof_sample(tsd_t *tsd, size_t alignment, size_t usize, prof_tctx_t *tctx) argument 1683 imemalign_prof(tsd_t *tsd, size_t alignment, size_t usize) argument 1697 prof_malloc(tsd_tsdn(tsd), p, usize, tctx); local 1708 size_t usize; local 1840 irealloc_prof_sample(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, prof_tctx_t *tctx) argument 1859 irealloc_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize) argument 1885 size_t usize; local 1912 isfree(tsd_t *tsd, void *ptr, size_t usize, tcache_t *tcache, bool slow_path) argument 2122 imallocx_flags_decode(tsd_t *tsd, size_t size, int flags, size_t *usize, size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena) argument 2154 imallocx_flags(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, bool slow_path) argument 2168 imallocx_prof_sample(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, bool slow_path) argument 2190 imallocx_prof(tsd_t *tsd, size_t size, int flags, size_t *usize, bool slow_path) argument 2215 prof_malloc(tsd_tsdn(tsd), p, *usize, tctx); local 2222 imallocx_no_prof(tsd_t *tsd, size_t size, int flags, size_t *usize, bool slow_path) argument 2242 imallocx_body(size_t size, int flags, tsdn_t **tsdn, size_t *usize, bool slow_path) argument 2287 size_t usize; local 2306 irallocx_prof_sample(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, prof_tctx_t *tctx) argument 2329 irallocx_prof(tsd_t *tsd, void *old_ptr, size_t old_usize, size_t size, size_t alignment, size_t *usize, bool zero, tcache_t *tcache, arena_t *arena) argument 2376 size_t usize; local 2452 size_t usize; local 2465 size_t usize; local 2479 size_t usize_max, usize; local 2529 size_t usize, old_usize; local 2589 size_t usize; local 2637 size_t usize; local 2654 size_t usize; local 2683 size_t usize; local [all...] |
H A D | prof.c | 226 prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, argument 230 prof_tctx_set(tsdn, ptr, usize, tctx); 234 tctx->cnts.curbytes += usize; 237 tctx->cnts.accumbytes += usize; 244 prof_free_sampled_object(tsd_t *tsd, size_t usize, prof_tctx_t *tctx) argument 249 assert(tctx->cnts.curbytes >= usize); 251 tctx->cnts.curbytes -= usize;
|
H A D | arena.c | 811 arena_huge_malloc_stats_update(arena_t *arena, size_t usize) argument 813 szind_t index = size2index(usize) - nlclasses - NBINS; 818 arena->stats.allocated_huge += usize; 824 arena_huge_malloc_stats_update_undo(arena_t *arena, size_t usize) argument 826 szind_t index = size2index(usize) - nlclasses - NBINS; 831 arena->stats.allocated_huge -= usize; 837 arena_huge_dalloc_stats_update(arena_t *arena, size_t usize) argument 839 szind_t index = size2index(usize) - nlclasses - NBINS; 844 arena->stats.allocated_huge -= usize; 850 arena_huge_reset_stats_cancel(arena_t *arena, size_t usize) argument 861 arena_huge_dalloc_stats_update_undo(arena_t *arena, size_t usize) argument 874 arena_huge_ralloc_stats_update(arena_t *arena, size_t oldsize, size_t usize) argument 882 arena_huge_ralloc_stats_update_undo(arena_t *arena, size_t oldsize, size_t usize) argument 917 arena_chunk_alloc_huge_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, size_t usize, size_t alignment, size_t *sn, bool *zero, size_t csize) argument 941 arena_chunk_alloc_huge(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, size_t *sn, bool *zero) argument 970 arena_chunk_dalloc_huge(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t usize, size_t sn) argument 989 arena_chunk_ralloc_huge_similar(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize) argument 1007 arena_chunk_ralloc_huge_shrink(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize, size_t sn) argument 1033 arena_chunk_ralloc_huge_expand_hard(tsdn_t *tsdn, arena_t *arena, chunk_hooks_t *chunk_hooks, void *chunk, size_t oldsize, size_t usize, size_t *sn, bool *zero, void *nchunk, size_t udiff, size_t cdiff) argument 1062 arena_chunk_ralloc_huge_expand(tsdn_t *tsdn, arena_t *arena, void *chunk, size_t oldsize, size_t usize, bool *zero) argument 1879 size_t usize = isalloc(tsd_tsdn(tsd), ptr, local 1941 size_t usize; local 2478 arena_redzone_corruption(void *ptr, size_t usize, bool after, size_t offset, uint8_t byte) argument 2550 arena_quarantine_junk_small(void *ptr, size_t usize) argument 2569 size_t usize; local 2622 size_t usize; local 2702 arena_palloc_large(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero) argument 2790 arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 2994 arena_dalloc_junk_large(void *ptr, size_t usize) argument 3017 size_t usize = arena_mapbits_large_size_get(chunk, pageind) - local 3115 size_t usize, splitsize, size, flag_dirty, flag_unzeroed_mask; local 3196 arena_ralloc_junk_large(void *ptr, size_t old_usize, size_t usize) argument 3299 arena_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, bool zero, tcache_t *tcache) argument 3317 size_t usize; local [all...] |
/external/jemalloc/include/jemalloc/internal/ |
H A D | quarantine.h | 16 size_t usize; member in struct:quarantine_obj_s
|
H A D | prof.h | 284 void prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, 286 void prof_free_sampled_object(tsd_t *tsd, size_t usize, prof_tctx_t *tctx); 334 void prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, 336 void prof_tctx_reset(tsdn_t *tsdn, const void *ptr, size_t usize, 338 bool prof_sample_accum_update(tsd_t *tsd, size_t usize, bool commit, 340 prof_tctx_t *prof_alloc_prep(tsd_t *tsd, size_t usize, bool prof_active, 342 void prof_malloc(tsdn_t *tsdn, const void *ptr, size_t usize, 344 void prof_realloc(tsd_t *tsd, const void *ptr, size_t usize, 347 void prof_free(tsd_t *tsd, const void *ptr, size_t usize); 411 prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, prof_tctx_ argument 421 prof_tctx_reset(tsdn_t *tsdn, const void *ptr, size_t usize, const void *old_ptr, prof_tctx_t *old_tctx) argument 432 prof_sample_accum_update(tsd_t *tsd, size_t usize, bool update, prof_tdata_t **tdata_out) argument 462 prof_alloc_prep(tsd_t *tsd, size_t usize, bool prof_active, bool update) argument 483 prof_malloc(tsdn_t *tsdn, const void *ptr, size_t usize, prof_tctx_t *tctx) argument 497 prof_realloc(tsd_t *tsd, const void *ptr, size_t usize, prof_tctx_t *tctx, bool prof_active, bool updated, const void *old_ptr, size_t old_usize, prof_tctx_t *old_tctx) argument 525 prof_malloc_sample_object(tsd_tsdn(tsd), ptr, usize, tctx); local 527 prof_tctx_reset(tsd_tsdn(tsd), ptr, usize, old_ptr, old_tctx); local 534 prof_free(tsd_t *tsd, const void *ptr, size_t usize) argument [all...] |
H A D | jemalloc_internal.h | 659 size_t usize = (psz + delta_mask) & ~delta_mask; local 660 return (usize); 739 size_t usize = grp_size + mod_size; local 740 return (usize); 780 size_t usize = (size + delta_mask) & ~delta_mask; local 781 return (usize); 815 size_t usize; local 835 usize = s2u(ALIGNMENT_CEILING(size, alignment)); 836 if (usize < LARGE_MINCLASS) 837 return (usize); 1078 ipallocztm(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, bool is_metadata, arena_t *arena) argument 1098 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, tcache_t *tcache, arena_t *arena) argument 1106 ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero) argument 1130 u2rz(size_t usize) argument 1146 size_t usize = isalloc(tsdn, ptr, false); local 1207 size_t usize, copysize; local [all...] |
/external/jemalloc/test/integration/ |
H A D | allocated.c | 18 size_t sz, usize; local 70 usize = malloc_usable_size(p); 71 assert_u64_le(a0 + usize, a1, 87 assert_u64_le(d0 + usize, d1,
|
/external/jemalloc/test/unit/ |
H A D | quarantine.c | 62 arena_redzone_corruption_replacement(void *ptr, size_t usize, bool after, argument
|
H A D | junk.c | 41 arena_dalloc_junk_large_intercept(void *ptr, size_t usize) argument 45 arena_dalloc_junk_large_orig(ptr, usize); 46 for (i = 0; i < usize; i++) { 49 i, usize); 56 huge_dalloc_junk_intercept(void *ptr, size_t usize) argument 59 huge_dalloc_junk_orig(ptr, usize); 171 arena_ralloc_junk_large_intercept(void *ptr, size_t old_usize, size_t usize) argument 174 arena_ralloc_junk_large_orig(ptr, old_usize, usize); 176 assert_zu_eq(usize, shrink_size(large_maxclass), "Unexpected usize"); 203 arena_redzone_corruption_replacement(void *ptr, size_t usize, bool after, size_t offset, uint8_t byte) argument [all...] |
/external/tpm2/ |
H A D | TpmFail.c | 116 UINT32 usize; local 119 || UINT32_Unmarshal(&usize, buffer, size) != TPM_RC_SUCCESS 123 header->size = usize;
|
/external/v8/src/runtime/ |
H A D | runtime-atomics.cc | 580 uint32_t usize = NumberToUint32(*size); local 581 return isolate->heap()->ToBoolean(AtomicIsLockFree(usize));
|
/external/libcups/cups/ |
H A D | pwg-media.c | 272 char usize[12 + 1 + 12 + 3], /* Unit size: NNNNNNNNNNNNxNNNNNNNNNNNNuu */ local 314 name = usize; 358 uptr = usize; 359 (*format)(uptr, sizeof(usize) - (size_t)(uptr - usize), width); 362 (*format)(uptr, sizeof(usize) - (size_t)(uptr - usize), length); 366 * Safe because usize can hold up to 12 + 1 + 12 + 4 bytes. 375 snprintf(keyword, keysize, "%s_%s_%s", prefix, name, usize);
|
/external/pcre/dist2/src/ |
H A D | pcre2test.c | 4791 size_t bsize, usize; local 4801 usize = regerror(rc, &preg, (char *)pbuffer8, bsize); 4809 if (usize > bsize)
|
/external/elfutils/src/ |
H A D | readelf.c | 4354 uint8_t usize = *(uint8_t *) data++; local 4355 NEED (usize); 4358 print_block (usize, data); 4359 data += usize; 4385 usize = *(uint8_t *) data++; 4392 op_name, usize, uleb);
|