/external/chromium_org/third_party/mesa/src/src/gallium/state_trackers/d3d1x/d3d1xshader/src/ |
H A D | dxbc_dump.cpp | 36 struct dxbc_chunk_header* chunk = container.chunks[i]; local 38 memcpy(fourcc_str, &chunk->fourcc, 4); 40 out << "# DXBC chunk " << std::setw(2) << i << ": " << fourcc_str << " offset " << ((char*)chunk - (char*)container.data) << " size " << bswap_le32(chunk->size) << "\n";
|
/external/mesa3d/src/gallium/state_trackers/d3d1x/d3d1xshader/src/ |
H A D | dxbc_dump.cpp | 36 struct dxbc_chunk_header* chunk = container.chunks[i]; local 38 memcpy(fourcc_str, &chunk->fourcc, 4); 40 out << "# DXBC chunk " << std::setw(2) << i << ": " << fourcc_str << " offset " << ((char*)chunk - (char*)container.data) << " size " << bswap_le32(chunk->size) << "\n";
|
/external/chromium_org/content/browser/speech/ |
H A D | chunked_byte_buffer_unittest.cc | 27 // Append partially chunk 1. 32 // Complete chunk 1. 37 // Append fully chunk 2. 42 // Remove and check chunk 1. 43 scoped_ptr<ByteVector> chunk; local 44 chunk = buffer.PopChunk(); 45 EXPECT_TRUE(chunk != NULL); 46 EXPECT_EQ(4U, chunk->size()); 48 &(*chunk)[0], 49 chunk [all...] |
H A D | audio_buffer.cc | 61 scoped_refptr<AudioChunk> chunk(chunks_.front()); 63 return chunk; 67 scoped_refptr<AudioChunk> chunk(new AudioChunk(bytes_per_sample_)); 75 chunk->data_string_.reserve(resulting_length); 77 chunk->data_string_.append((*it)->data_string_); 80 return chunk;
|
/external/chromium_org/gpu/command_buffer/client/ |
H A D | mapped_memory.cc | 40 MemoryChunk* chunk = *iter; local 41 cmd_buf->DestroyTransferBuffer(chunk->shm_id()); 53 MemoryChunk* chunk = chunks_[ii]; local 54 chunk->FreeUnused(); 55 total_bytes_in_use += chunk->bytes_in_use(); 56 if (chunk->GetLargestFreeSizeWithoutWaiting() >= size) { 57 void* mem = chunk->Alloc(size); 59 *shm_id = chunk->shm_id(); 60 *shm_offset = chunk->GetOffset(mem); 72 MemoryChunk* chunk local 107 MemoryChunk* chunk = chunks_[ii]; local 118 MemoryChunk* chunk = chunks_[ii]; local 131 MemoryChunk* chunk = *iter; local [all...] |
/external/openssl/crypto/md4/ |
H A D | md4_one.c | 81 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD4_Update(&c,temp,chunk); 88 n -= chunk; 89 d += chunk;
|
/external/openssl/crypto/md5/ |
H A D | md5_one.c | 81 unsigned long chunk; 85 chunk = (n > sizeof(temp)) ? sizeof(temp) : n; 86 ebcdic2ascii(temp, d, chunk); 87 MD5_Update(&c,temp,chunk); 88 n -= chunk; 89 d += chunk;
|
/external/jemalloc/src/ |
H A D | arena.c | 56 static void arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, 58 static void arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, 168 * Break ties by chunk address. For fragmented chunks, report lower 191 arena_avail_adjac_pred(arena_chunk_t *chunk, size_t pageind) argument 198 ret = (arena_mapbits_allocated_get(chunk, pageind-1) == 0); 199 assert(ret == false || arena_mapbits_dirty_get(chunk, 200 pageind-1) != arena_mapbits_dirty_get(chunk, pageind)); 206 arena_avail_adjac_succ(arena_chunk_t *chunk, size_t pageind, size_t npages) argument 214 ret = (arena_mapbits_allocated_get(chunk, pageind+npages) == 0); 215 assert(ret == false || arena_mapbits_dirty_get(chunk, pagein 222 arena_avail_adjac(arena_chunk_t *chunk, size_t pageind, size_t npages) argument 230 arena_avail_insert(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages, bool maybe_adjac_pred, bool maybe_adjac_succ) argument 263 arena_avail_remove(arena_t *arena, arena_chunk_t *chunk, size_t pageind, size_t npages, bool maybe_adjac_pred, bool maybe_adjac_succ) argument 320 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local 344 arena_run_zero(arena_chunk_t *chunk, size_t run_ind, size_t npages) argument 354 arena_run_page_mark_zeroed(arena_chunk_t *chunk, size_t run_ind) argument 362 arena_run_page_validate_zeroed(arena_chunk_t *chunk, size_t run_ind) argument 386 arena_run_split_remove(arena_t *arena, arena_chunk_t *chunk, size_t run_ind, size_t flag_dirty, size_t need_pages) argument 430 arena_chunk_t *chunk; local 497 arena_chunk_t *chunk; local 542 arena_chunk_t *chunk; local 565 arena_chunk_t *chunk; local 623 arena_chunk_t *chunk; local 683 arena_chunk_t *chunk; local 701 arena_chunk_dalloc_internal(arena_t *arena, arena_chunk_t *chunk) argument 714 arena_chunk_dalloc_huge(arena_t *arena, void *chunk, size_t size) argument 732 arena_chunk_dalloc(arena_t *arena, arena_chunk_t *chunk) argument 784 arena_chunk_t *chunk; local 838 arena_chunk_t *chunk; local 896 chunks_dirty_iter_cb(arena_chunk_tree_t *tree, arena_chunk_t *chunk, void *arg) argument 927 arena_chunk_stash_dirty(arena_t *arena, arena_chunk_t *chunk, bool all, arena_chunk_mapelms_t *mapelms) argument 985 arena_chunk_purge_stashed(arena_t *arena, arena_chunk_t *chunk, arena_chunk_mapelms_t *mapelms) argument 1032 arena_chunk_unstash_purged(arena_t *arena, arena_chunk_t *chunk, arena_chunk_mapelms_t *mapelms) argument 1052 arena_chunk_purge(arena_t *arena, arena_chunk_t *chunk, bool all) argument 1101 arena_chunk_t *chunk; local 1189 arena_run_coalesce(arena_t *arena, arena_chunk_t *chunk, size_t *p_size, size_t *p_run_ind, size_t *p_run_pages, size_t flag_dirty) argument 1259 arena_chunk_t *chunk; local 1333 arena_run_trim_head(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize) argument 1365 arena_run_trim_tail(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, size_t oldsize, size_t newsize, bool dirty) argument 1402 arena_chunk_t *chunk; local 1420 arena_chunk_t *chunk = CHUNK_ADDR2BASE(run); local 1432 arena_chunk_t *chunk = (arena_chunk_t *)CHUNK_ADDR2BASE(run); local 1528 arena_chunk_t *chunk; local 1800 arena_chunk_t *chunk; local 1852 arena_chunk_t *chunk; local 1873 arena_dissociate_bin_run(arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 1896 arena_dalloc_bin_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 1947 arena_bin_lower_run(arena_t *arena, arena_chunk_t *chunk, arena_run_t *run, arena_bin_t *bin) argument 1968 arena_dalloc_bin_locked(arena_t *arena, arena_chunk_t *chunk, void *ptr, arena_chunk_map_t *mapelm) argument 2003 arena_dalloc_bin(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind, arena_chunk_map_t *mapelm) argument 2018 arena_dalloc_small(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t pageind) argument 2051 arena_dalloc_large_locked(arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 2071 arena_dalloc_large(arena_t *arena, arena_chunk_t *chunk, void *ptr) argument 2080 arena_ralloc_large_shrink(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size) argument 2110 arena_ralloc_large_grow(arena_t *arena, arena_chunk_t *chunk, void *ptr, size_t oldsize, size_t size, size_t extra, bool zero) argument 2209 arena_chunk_t *chunk; local [all...] |
H A D | chunk.c | 26 /* Various chunk-related settings. */ 39 static void chunk_dalloc_core(void *chunk, size_t size); 56 * the current chunk allocation request is on behalf of the 87 /* Insert the leading space as a smaller chunk. */ 94 /* Insert the trailing space as a smaller chunk. */ 98 * base_node_alloc() can cause a new base chunk to be 182 chunk_register(void *chunk, size_t size, bool base) argument 185 assert(chunk != NULL); 186 assert(CHUNK_ADDR2BASE(chunk) == chunk); 258 chunk_record(extent_tree_t *chunks_szad, extent_tree_t *chunks_ad, void *chunk, size_t size) argument 346 chunk_unmap(void *chunk, size_t size) argument 360 chunk_dalloc_core(void *chunk, size_t size) argument 382 chunk_dalloc_default(void *chunk, size_t size, unsigned arena_ind) argument [all...] |
/external/apache-xml/src/main/java/org/apache/xml/dtm/ref/ |
H A D | ChunkedIntArray.java | 94 int[] chunk = chunks.elementAt(chunkpos); 95 chunk[slotpos] = w0; 96 chunk[slotpos+1] = w1; 97 chunk[slotpos+2] = w2; 98 chunk[slotpos+3] = w3; 126 int[] chunk = chunks.elementAt(chunkpos); 127 return chunk[slotpos + offset]; 148 int[] chunk = chunks.elementAt(chunkpos); 153 ancestor = chunk[slotpos + 1]; 175 CIA; when only a single characters() chunk ha [all...] |
/external/guava/guava-tests/test/com/google/common/io/ |
H A D | LineBufferTest.java | 73 for (int chunk : CHUNK_SIZES) { 74 chunk = Math.max(1, Math.min(chunk, input.length())); 75 assertEquals(expectProcess, bufferHelper(input, chunk)); 76 assertEquals(expectRead, readUsingJava(input, chunk)); 77 assertEquals(expectRead, readUsingReader(input, chunk, true)); 78 assertEquals(expectRead, readUsingReader(input, chunk, false)); 82 private static List<String> bufferHelper(String input, int chunk) argument 93 int len = Math.min(chars.length, off + chunk) - off; 101 private static List<String> readUsingJava(String input, int chunk) argument 113 readUsingReader(String input, int chunk, boolean asReader) argument 128 getChunkedReadable(String input, int chunk) argument 138 getChunkedReader(String input, final int chunk) argument [all...] |
/external/libpng/contrib/libtests/ |
H A D | fakepng.c | 15 put_chunk(const unsigned char *chunk, uInt length) argument 21 fwrite(chunk, length, 1, stdout); 24 put_uLong(crc32(crc, chunk, length));
|
/external/jemalloc/include/jemalloc/internal/ |
H A D | arena.h | 57 /* Each element of the chunk map corresponds to one page within the chunk. */ 63 * chunk map overhead accounts for a percentage of memory, rather than 162 /* Arena chunk header. */ 164 /* Arena that owns the chunk. */ 186 * Map of pages within chunk that keeps track of free/large/small. The 187 * first map_bias entries are omitted, since the chunk header does not 189 * for common chunk sizes (e.g. 4 MiB). 337 * In order to avoid rapid chunk allocation/deallocation when an arena 338 * oscillates right on the cusp of needing a new chunk, cach 1027 arena_chunk_t *chunk; local 1049 arena_chunk_t *chunk; local 1099 arena_chunk_t *chunk; local 1139 arena_dalloc(arena_chunk_t *chunk, void *ptr, bool try_tcache) argument [all...] |
H A D | chunk_mmap.h | 15 bool chunk_dalloc_mmap(void *chunk, size_t size);
|
/external/nanohttpd/core/src/test/java/fi/iki/elonen/ |
H A D | HttpChunkedResponseTest.java | 43 int chunk = 0; field in class:HttpChunkedResponseTest.ChunkedInputStream 53 for (int i = 0; i < chunks[chunk].length(); ++i) { 54 buffer[i] = (byte) chunks[chunk].charAt(i); 56 return chunks[chunk++].length();
|
/external/chromium_org/v8/tools/testrunner/server/ |
H A D | compression.py | 74 chunk = self.sock.recv(8192) 77 if not chunk: return None 78 self._AppendData(chunk) 83 chunk = self.sock.recv(8192) 86 if not chunk: return None 87 self._AppendData(chunk) 104 chunk = self.data.read(length) 111 return chunk
|
/external/bison/lib/ |
H A D | obstack.c | 133 /* Initialize an obstack H for use. Specify chunk size SIZE (0 means default). 147 register struct _obstack_chunk *chunk; /* points to new chunk */ 174 chunk = h->chunk = CALL_CHUNKFUN (h, h -> chunk_size); 175 if (!chunk) 177 h->next_free = h->object_base = __PTR_ALIGN ((char *) chunk, chunk->contents, 179 h->chunk_limit = chunk->limit 180 = (char *) chunk 146 register struct _obstack_chunk *chunk; /* points to new chunk */ local 193 register struct _obstack_chunk *chunk; /* points to new chunk */ local [all...] |
/external/elfutils/0.153/libelf/ |
H A D | elf_getdata_rawchunk.c | 1 /* Return converted data from raw chunk of ELF file. 170 Elf_Data_Chunk *chunk = calloc (1, sizeof *chunk); local 171 if (chunk == NULL) 178 chunk->dummy_scn.elf = elf; 179 chunk->dummy_scn.flags = flags; 180 chunk->data.s = &chunk->dummy_scn; 181 chunk->data.d.d_buf = buffer; 182 chunk [all...] |
/external/fonttools/Lib/fontTools/ |
H A D | t1Lib.py | 169 raise T1Error('bad chunk code: ' + repr(code)) 186 chunk = f.read(chunklen) 187 assert len(chunk) == chunklen 188 data.append(chunk) 192 raise T1Error('bad chunk code: ' + repr(code)) 209 for isEncrypted, chunk in chunks: 210 if isEncrypted and isHex(chunk[:4]): 211 data.append(deHexString(chunk)) 213 data.append(chunk) 225 for isEncrypted, chunk i [all...] |
/external/fonttools/Tools/fontTools/ |
H A D | t1Lib.py | 169 raise T1Error('bad chunk code: ' + repr(code)) 186 chunk = f.read(chunklen) 187 assert len(chunk) == chunklen 188 data.append(chunk) 192 raise T1Error('bad chunk code: ' + repr(code)) 209 for isEncrypted, chunk in chunks: 210 if isEncrypted and isHex(chunk[:4]): 211 data.append(deHexString(chunk)) 213 data.append(chunk) 225 for isEncrypted, chunk i [all...] |
/external/chromium_org/chrome/browser/resources/chromeos/chromevox/host/chrome/ |
H A D | expanding_braille_translator.js | 117 function chunkTranslated(chunk, cells, textToBraille, brailleToText) { 118 chunk.cells = cells; 119 chunk.textToBraille = textToBraille; 120 chunk.brailleToText = brailleToText; 128 function(accum, chunk) { return accum + chunk.cells.byteLength}, 0); 138 for (var i = 0, chunk; chunk = chunks[i]; ++i) { 139 cells.set(new Uint8Array(chunk.cells), cellPos); 140 appendAdjusted(textToBraille, chunk [all...] |
/external/mockito/src/org/mockito/internal/verification/checkers/ |
H A D | NumberOfInvocationsInOrderChecker.java | 35 List<Invocation> chunk = finder.findMatchingChunk(invocations, wanted, wantedCount, context); 37 int actualCount = chunk.size(); 40 Location lastInvocation = finder.getLastLocation(chunk); 43 Location firstUndesired = chunk.get(wantedCount).getLocation(); 47 invocationMarker.markVerifiedInOrder(chunk, wanted, context);
|
/external/mockito/src/org/mockito/internal/invocation/ |
H A D | InvocationMarker.java | 25 public void markVerifiedInOrder(List<Invocation> chunk, CapturesArgumensFromInvocation wanted, InOrderContext context) {
argument 26 markVerified(chunk, wanted);
28 for (Invocation i : chunk) {
|
/external/chromium_org/chrome/browser/safe_browsing/ |
H A D | chunk_range.cc | 60 for (int chunk = range.start(); chunk <= range.stop(); ++chunk) { 61 chunks->push_back(chunk); 70 // Crack the string into chunk parts, then crack each part looking for a 83 // atoi error, since chunk numbers are guaranteed to never be 0. 104 const ChunkRange& chunk = ranges[mid]; local 105 if ((chunk.stop() >= chunk_number) && (chunk.start() <= chunk_number)) 109 if (chunk [all...] |
/external/jemalloc/android/test/ |
H A D | run_integration.sh | 9 $test_dir/chunk
|