/external/kmod/shared/ |
H A D | hash.h | 9 unsigned int bucket; member in struct:hash_iter
|
H A D | hash.c | 69 struct hash_bucket *bucket, *bucket_end; local 74 bucket = hash->buckets; 75 bucket_end = bucket + hash->n_buckets; 76 for (; bucket < bucket_end; bucket++) { 79 entry = bucket->entries; 80 entry_end = entry + bucket->used; 84 free(bucket->entries); 151 struct hash_bucket *bucket = hash->buckets + pos; local 154 if (bucket 194 struct hash_bucket *bucket = hash->buckets + pos; local 239 const struct hash_bucket *bucket = hash->buckets + pos; local 258 struct hash_bucket *bucket = hash->buckets + pos; local [all...] |
/external/brotli/c/enc/ |
H A D | prefix.h | 36 size_t bucket = Log2FloorNonZero(dist) - 1; local 39 size_t prefix = (dist >> bucket) & 1; 40 size_t offset = (2 + prefix) << bucket; 41 size_t nbits = bucket - postfix_bits;
|
H A D | hash_longest_match_inc.h | 22 /* HashBytes is the function that chooses the bucket to place the address in. */ 36 /* Left-shift for computing hash bucket index from hash value. */ 43 /* Number of entries in a particular bucket. */ 213 uint32_t* BROTLI_RESTRICT bucket = local 218 size_t prev_ix = bucket[--i & self->block_mask_]; 248 bucket[num[key] & self->block_mask_] = (uint32_t)cur_ix;
|
H A D | hash_longest_match_quickly_inc.h | 21 /* HashBytes is the function that chooses the bucket to place 86 /* Wiggle the value with the bucket sweep range. */ 194 uint32_t *bucket = self->buckets_ + key; local 196 prev_ix = *bucket++; 197 for (i = 0; i < BUCKET_SWEEP; ++i, prev_ix = *bucket++) {
|
/external/libdrm/etnaviv/ |
H A D | etnaviv_bo_cache.c | 82 struct etna_bo_bucket *bucket = &cache->cache_bucket[i]; local 85 while (!LIST_IS_EMPTY(&bucket->list)) { 86 bo = LIST_ENTRY(struct etna_bo, bucket->list.next, list); 105 * way to the correct bucket size rather than looping.. 108 struct etna_bo_bucket *bucket = &cache->cache_bucket[i]; local 109 if (bucket->size >= size) { 110 return bucket; 125 static struct etna_bo *find_in_bucket(struct etna_bo_bucket *bucket, uint32_t flags) argument 130 while (!LIST_IS_EMPTY(&bucket->list)) { 131 bo = LIST_ENTRY(struct etna_bo, bucket 154 struct etna_bo_bucket *bucket; local 175 struct etna_bo_bucket *bucket = get_bucket(cache, bo->size); local [all...] |
/external/libdrm/freedreno/ |
H A D | freedreno_bo_cache.c | 53 * @coarse: if true, only power-of-two bucket sizes, otherwise 95 struct fd_bo_bucket *bucket = &cache->cache_bucket[i]; local 98 while (!LIST_IS_EMPTY(&bucket->list)) { 99 bo = LIST_ENTRY(struct fd_bo, bucket->list.next, list); 118 * way to the correct bucket size rather than looping.. 121 struct fd_bo_bucket *bucket = &cache->cache_bucket[i]; local 122 if (bucket->size >= size) { 123 return bucket; 138 static struct fd_bo *find_in_bucket(struct fd_bo_bucket *bucket, uint32_t flags) argument 150 if (!LIST_IS_EMPTY(&bucket 169 struct fd_bo_bucket *bucket; local 199 struct fd_bo_bucket *bucket = get_bucket(cache, bo->size); local [all...] |
/external/libdrm/tests/ |
H A D | hash.c | 87 static int count_entries(HashBucketPtr bucket) argument 91 for (; bucket; bucket = bucket->next) 107 HashBucketPtr bucket; local 113 bucket = table->buckets[i]; 114 update_dist(count_entries(bucket));
|
/external/fio/ |
H A D | filehash.c | 54 struct flist_head *bucket = &file_hash[hash(name)]; local 57 flist_for_each(n, bucket) {
|
/external/libdrm/ |
H A D | xf86drmHash.c | 127 HashBucketPtr bucket; local 134 for (bucket = table->buckets[i]; bucket;) { 135 next = bucket->next; 136 drmFree(bucket); 137 bucket = next; 144 /* Find the bucket and organize the list so that this bucket is at the 152 HashBucketPtr bucket; local 156 for (bucket 178 HashBucketPtr bucket; local 191 HashBucketPtr bucket; local 214 HashBucketPtr bucket; local [all...] |
/external/mesa3d/src/gallium/drivers/swr/rasterizer/core/ |
H A D | arena.h | 81 uint32_t bucket = GetBucketId(size); local 86 ArenaBlock* pPrevBlock = &m_cachedBlocks[bucket]; 92 if (pBlock == m_pLastCachedBlocks[bucket]) 94 m_pLastCachedBlocks[bucket] = pPrevBlock; 99 pPrevBlock = &m_oldCachedBlocks[bucket]; 105 if (pBlock == m_pOldLastCachedBlocks[bucket]) 107 m_pOldLastCachedBlocks[bucket] = pPrevBlock; 133 if (bucket && bucket < (CACHE_NUM_BUCKETS - 1)) 135 // Make all blocks in this bucket th [all...] |
/external/blktrace/ |
H A D | stats.h | 116 static inline void histlog2_account(__u32 *bucket, __u32 val, argument 120 bucket[index]++;
|
/external/llvm/lib/CodeGen/AsmPrinter/ |
H A D | DwarfAccelTable.cpp | 54 // Then compute the bucket size, minimum of 1 bucket. 87 // Figure out how many buckets we need, then compute the bucket 94 // Compute bucket contents and final ordering. 97 uint32_t bucket = Data[i]->HashValue % Header.bucket_count; local 98 Buckets[bucket].push_back(Data[i]); 162 // bucket. 180 // element in each bucket. This is done via a symbol subtraction from the 204 // the bucket. For the string case emit the dies and the various offsets. 205 // Terminate each HashData bucket wit [all...] |
/external/mesa3d/src/gallium/drivers/svga/ |
H A D | svga_screen_cache.h | 82 /** Head for the bucket lists. */ 111 struct list_head bucket[SVGA_HOST_SURFACE_CACHE_BUCKETS]; member in struct:svga_host_surface_cache
|
/external/mesa3d/src/gallium/drivers/swr/rasterizer/common/ |
H A D | rdtsc_buckets.cpp | 79 void BucketManager::PrintBucket(FILE* f, UINT level, uint64_t threadCycles, uint64_t parentCycles, const BUCKET& bucket) argument 93 // compute percent of total cycles used by this bucket 94 float percentTotal = (float)((double)bucket.elapsed / (double)threadCycles * 100.0); 96 // compute percent of parent cycles used by this bucket 97 float percentParent = (float)((double)bucket.elapsed / (double)parentCycles * 100.0); 100 uint64_t CPE = bucket.elapsed / bucket.count; 102 BUCKET_DESC &desc = mBuckets[bucket.id]; 113 bucket.elapsed, 115 bucket [all...] |
/external/tensorflow/tensorflow/contrib/training/python/training/ |
H A D | bucket_ops.py | 59 raise ValueError("Expected at least one tensor in bucket().") 63 def bucket(tensors, function 82 The tensors entering this function are put into the bucket given by 83 `which_bucket`. Each bucket has its own queue. When a bucket contains 119 to bucket. Nested lists are not supported. 122 the same size). If a list is passed in then each bucket will have a 128 and also (by default) the maximum number of elements within each bucket. 130 each bucket. If None, capacity is used (default). If specified, it must 132 as capacity for the i-th bucket queu [all...] |
/external/tensorflow/tensorflow/core/kernels/ |
H A D | bucketize_op_gpu.cu.cc | 60 int32 bucket = 0; local 63 int32 l = bucket; 67 bucket = ++l; 73 out[i] = bucket;
|
/external/freetype/src/cache/ |
H A D | ftccache.c | 87 /* get a top bucket for specified hash from cache, 130 * the bucket lists 144 /* split a single bucket */ 488 FTC_Node* bucket; local 501 bucket = pnode = FTC_NODE_TOP_FOR_HASH( cache, hash ); 520 /* Update bucket by modified linked list */ 521 bucket = pnode = FTC_NODE_TOP_FOR_HASH( cache, hash ); 537 if ( node != *bucket ) 540 node->link = *bucket; 541 *bucket 575 FTC_Node* bucket = cache->buckets + i; local [all...] |
/external/jemalloc/src/ |
H A D | ckh.c | 4 * hash bucket contains 2^n cells, for n >= 1, and 2 indicates that two hash 24 * | #cells/bucket | 32 * The number of cells per bucket is chosen such that a bucket fits in one cache 49 * Search bucket for key and return the cell number if found; SIZE_T_MAX 53 ckh_bucket_search(ckh_t *ckh, size_t bucket, const void *key) argument 59 cell = &ckh->tab[(bucket << LG_CKH_BUCKET_CELLS) + i]; 61 return ((bucket << LG_CKH_BUCKET_CELLS) + i); 73 size_t hashes[2], bucket, cell; local 79 /* Search primary bucket 92 ckh_try_bucket_insert(ckh_t *ckh, size_t bucket, const void *key, const void *data) argument 130 size_t hashes[2], bucket, tbucket; local 198 size_t hashes[2], bucket; local [all...] |
/external/libchrome/base/metrics/ |
H A D | persistent_sample_map.cc | 194 return false; // SparseHistogram only supports bucket with size 1. 207 Count* bucket = GetSampleCountStorage(min); local 208 if (bucket == nullptr) { 211 if (*bucket < count) { 213 *bucket = 0; 215 *bucket -= count;
|
/external/libchrome/base/trace_event/ |
H A D | heap_profiler_heap_dump_writer.cc | 30 // The pairs are grouped into |Bucket|s. A bucket is a group of (context, size) 33 // bucket that represents the entire heap. Then this bucket is recursively 34 // broken down into smaller buckets. Each bucket keeps track of whether further 63 // elements in this bucket, the stack frames 0 up to (but not including) the 67 // When true, the type name for all elements in this bucket must be equal. 76 // Groups the allocations in the bucket by |break_by|. The buckets in the 79 std::vector<Bucket> GetSubbuckets(const Bucket& bucket, argument 85 for (const auto& context_and_metrics : bucket.metrics_by_context) { 89 const StackFrame* cursor = begin + bucket 131 BreakDownBy(const Bucket& bucket, BreakDownMode break_by, size_t min_size_bytes) argument 192 AddEntryForBucket(const Bucket& bucket) argument 220 BreakDown(const Bucket& bucket) argument [all...] |
/external/libnetfilter_conntrack/src/conntrack/ |
H A D | labels.c | 25 struct labelmap_bucket *bucket; local 31 bucket = malloc(sizeof(*bucket)); 32 if (!bucket) { 36 bucket->name = name; 37 bucket->bit = b; 38 return bucket;
|
/external/libnl/lib/route/qdisc/ |
H A D | tbf.c | 108 nl_dump(p, "rate-bucket-size %1.f%s " 125 "bucket-size %.1f%s cell-size %.1f%s" 190 int bucket) 195 limit += bucket; 262 static inline int calc_cell_log(int cell, int bucket) argument 272 * @arg bucket Size of bucket in bytes. 276 void rtnl_qdisc_tbf_set_rate(struct rtnl_qdisc *qdisc, int rate, int bucket, argument 291 tbf->qt_rate_bucket = bucket; 293 tbf->qt_rate_txtime = nl_us2ticks(rtnl_tc_calc_txtime(bucket, rat 189 calc_limit(struct rtnl_ratespec *spec, int latency, int bucket) argument 359 rtnl_qdisc_tbf_set_peakrate(struct rtnl_qdisc *qdisc, int rate, int bucket, int cell) argument [all...] |
/external/libvpx/libvpx/ |
H A D | rate_hist.c | 33 struct hist_bucket bucket[RATE_BINS]; member in struct:rate_hist 62 hist->bucket[i].low = INT_MAX; 63 hist->bucket[i].high = 0; 64 hist->bucket[i].count = 0; 118 if (hist->bucket[idx].low > avg_bitrate) 119 hist->bucket[idx].low = (int)avg_bitrate; 120 if (hist->bucket[idx].high < avg_bitrate) 121 hist->bucket[idx].high = (int)avg_bitrate; 122 hist->bucket[idx].count++; 126 static int merge_hist_buckets(struct hist_bucket *bucket, in argument 194 show_histogram(const struct hist_bucket *bucket, int buckets, int total, int scale) argument 254 struct hist_bucket bucket[64]; local [all...] |
/external/mesa3d/src/gallium/drivers/nouveau/ |
H A D | nouveau_mm.c | 36 struct mm_bucket bucket[MM_NUM_BUCKETS]; member in struct:nouveau_mman 97 return &cache->bucket[MAX2(order, MM_MIN_ORDER) - MM_MIN_ORDER]; 167 struct mm_bucket *bucket; local 172 bucket = mm_bucket_by_size(cache, size); 173 if (!bucket) { 184 if (!LIST_IS_EMPTY(&bucket->used)) { 185 slab = LIST_ENTRY(struct mm_slab, bucket->used.next, head); 187 if (LIST_IS_EMPTY(&bucket->free)) { 190 slab = LIST_ENTRY(struct mm_slab, bucket->free.next, head); 193 LIST_ADD(&slab->head, &bucket 220 struct mm_bucket *bucket = mm_bucket_by_order(slab->cache, slab->order); local [all...] |