/external/webrtc/webrtc/modules/desktop_capture/mac/ |
H A D | scoped_pixel_buffer_object.cc | 27 int size_in_bytes) { 33 glBufferDataARB(GL_PIXEL_PACK_BUFFER_ARB, size_in_bytes, NULL, 26 Init(CGLContextObj cgl_context, int size_in_bytes) argument
|
/external/libchrome/base/trace_event/ |
H A D | process_memory_maps.h | 34 uint64_t size_in_bytes; member in struct:base::trace_event::ProcessMemoryMaps::VMRegion
|
/external/webrtc/webrtc/modules/audio_device/ |
H A D | fine_audio_buffer.cc | 115 size_t size_in_bytes, 121 if (record_write_pos_ + size_in_bytes > required_record_buffer_size_bytes_) { 131 memcpy(record_cache_buffer_.get() + record_write_pos_, buffer, size_in_bytes); 132 record_write_pos_ += size_in_bytes; 133 record_cached_bytes_ += size_in_bytes; 114 DeliverRecordedData(const int8_t* buffer, size_t size_in_bytes, int playout_delay_ms, int record_delay_ms) argument
|
/external/stressapptest/src/ |
H A D | adler32memcpy.cc | 55 // 1. size_in_bytes is a multiple of 16. 57 // 3. size_in_bytes is less than 2^19 bytes. 86 bool CalculateAdlerChecksum(uint64 *data64, unsigned int size_in_bytes, argument 90 unsigned int count = size_in_bytes / sizeof(data); 125 unsigned int size_in_bytes, AdlerChecksum *checksum) { 128 unsigned int count = size_in_bytes / sizeof(data); 166 unsigned int size_in_bytes, AdlerChecksum *checksum) { 169 unsigned int count = size_in_bytes / sizeof(data); 227 unsigned int size_in_bytes, AdlerChecksum *checksum) { 264 if ((size_in_bytes >> 1 124 AdlerMemcpyC(uint64 *dstmem64, uint64 *srcmem64, unsigned int size_in_bytes, AdlerChecksum *checksum) argument 165 AdlerMemcpyWarmC(uint64 *dstmem64, uint64 *srcmem64, unsigned int size_in_bytes, AdlerChecksum *checksum) argument 226 AdlerMemcpyAsm(uint64 *dstmem64, uint64 *srcmem64, unsigned int size_in_bytes, AdlerChecksum *checksum) argument [all...] |
H A D | os.cc | 257 unsigned int size_in_bytes, 260 return AdlerMemcpyAsm(dstmem, srcmem, size_in_bytes, checksum); 262 return AdlerMemcpyWarmC(dstmem, srcmem, size_in_bytes, checksum); 256 AdlerMemcpyWarm(uint64 *dstmem, uint64 *srcmem, unsigned int size_in_bytes, AdlerChecksum *checksum) argument
|
H A D | worker.cc | 1012 unsigned int size_in_bytes, 1018 unsigned int count = size_in_bytes / sizeof(data); 1083 unsigned int size_in_bytes, 1088 os_->AdlerMemcpyWarm(dstmem64, srcmem64, size_in_bytes, &ignored_checksum); 1093 int length = size_in_bytes / kCacheLineSize; 1103 AdlerAddrCrcC(srcmem64, size_in_bytes, checksum, pe); 1105 TagAddrC(dstmem64, size_in_bytes); 1111 unsigned int size_in_bytes) { 1116 int length = size_in_bytes / wordsize_; 1127 unsigned int size_in_bytes, 1010 AdlerAddrMemcpyC(uint64 *dstmem64, uint64 *srcmem64, unsigned int size_in_bytes, AdlerChecksum *checksum, struct page_entry *pe) argument 1081 AdlerAddrMemcpyWarm(uint64 *dstmem64, uint64 *srcmem64, unsigned int size_in_bytes, AdlerChecksum *checksum, struct page_entry *pe) argument 1110 TagAddrC(uint64 *memwords, unsigned int size_in_bytes) argument 1126 AdlerAddrCrcC(uint64 *srcmem64, unsigned int size_in_bytes, AdlerChecksum *checksum, struct page_entry *pe) argument [all...] |
/external/compiler-rt/lib/asan/ |
H A D | asan_malloc_linux.cc | 38 static void *AllocateFromLocalPool(uptr size_in_bytes) { argument 39 uptr size_in_words = RoundUpTo(size_in_bytes, kWordSize) / kWordSize;
|
/external/parameter-framework/asio/include/asio/ |
H A D | buffer.hpp | 642 * @returns <tt>mutable_buffers_1(data, size_in_bytes)</tt>. 644 inline mutable_buffers_1 buffer(void* data, std::size_t size_in_bytes) argument 646 return mutable_buffers_1(mutable_buffer(data, size_in_bytes)); 651 * @returns <tt>const_buffers_1(data, size_in_bytes)</tt>. 654 std::size_t size_in_bytes) 656 return const_buffers_1(const_buffer(data, size_in_bytes)); 653 buffer(const void* data, std::size_t size_in_bytes) argument
|
/external/parameter-framework/asio-1.10.6/include/asio/ |
H A D | buffer.hpp | 642 * @returns <tt>mutable_buffers_1(data, size_in_bytes)</tt>. 644 inline mutable_buffers_1 buffer(void* data, std::size_t size_in_bytes) argument 646 return mutable_buffers_1(mutable_buffer(data, size_in_bytes)); 651 * @returns <tt>const_buffers_1(data, size_in_bytes)</tt>. 654 std::size_t size_in_bytes) 656 return const_buffers_1(const_buffer(data, size_in_bytes)); 653 buffer(const void* data, std::size_t size_in_bytes) argument
|
/external/v8/src/crankshaft/ |
H A D | hydrogen-escape-analysis.cc | 50 int size_in_bytes = allocate->size()->GetInteger32Constant(); local 51 if (HasNoEscapingUses(instr, size_in_bytes)) { 300 int size_in_bytes = allocate->size()->GetInteger32Constant(); local 301 number_of_values_ = size_in_bytes / kPointerSize;
|
/external/v8/src/heap/ |
H A D | spaces-inl.h | 382 HeapObject* PagedSpace::AllocateLinearly(int size_in_bytes) { argument 384 Address new_top = current_top + size_in_bytes; 393 int size_in_bytes, AllocationAlignment alignment) { 397 Address new_top = current_top + filler_size + size_in_bytes; 410 HeapObject* PagedSpace::AllocateLinearlyAligned(int* size_in_bytes, argument 415 Address new_top = current_top + filler_size + *size_in_bytes; 420 *size_in_bytes += filler_size; 431 int size_in_bytes, UpdateSkipList update_skip_list) { 432 HeapObject* object = AllocateLinearly(size_in_bytes); 435 object = free_list_.Allocate(size_in_bytes); 392 AllocateRawAligned( int size_in_bytes, AllocationAlignment alignment) argument 430 AllocateRawUnaligned( int size_in_bytes, UpdateSkipList update_skip_list) argument 459 AllocateRawUnalignedSynchronized( int size_in_bytes) argument 467 AllocateRawAligned(int size_in_bytes, AllocationAlignment alignment) argument 500 AllocateRaw(int size_in_bytes, AllocationAlignment alignment) argument 522 AllocateRawAligned(int size_in_bytes, AllocationAlignment alignment) argument 553 AllocateRawUnaligned(int size_in_bytes) argument 574 AllocateRaw(int size_in_bytes, AllocationAlignment alignment) argument 586 AllocateRawSynchronized( int size_in_bytes, AllocationAlignment alignment) argument [all...] |
H A D | heap-inl.h | 296 AllocationResult Heap::AllocateRaw(int size_in_bytes, AllocationSpace space, argument 310 bool large_object = size_in_bytes > kMaxRegularHeapObjectSize; 317 allocation = new_space_->AllocateRaw(size_in_bytes, alignment); 319 OnAllocationEvent(object, size_in_bytes); 328 allocation = lo_space_->AllocateRaw(size_in_bytes, NOT_EXECUTABLE); 330 allocation = old_space_->AllocateRaw(size_in_bytes, alignment); 333 if (size_in_bytes <= code_space()->AreaSize()) { 334 allocation = code_space_->AllocateRawUnaligned(size_in_bytes); 336 allocation = lo_space_->AllocateRaw(size_in_bytes, EXECUTABLE); 340 allocation = lo_space_->AllocateRaw(size_in_bytes, NOT_EXECUTABL 355 OnAllocationEvent(HeapObject* object, int size_in_bytes) argument 383 OnMoveEvent(HeapObject* target, HeapObject* source, int size_in_bytes) argument [all...] |
H A D | spaces.cc | 1687 void NewSpace::UpdateInlineAllocationLimit(int size_in_bytes) { argument 1691 Address new_top = allocation_info_.top() + size_in_bytes; 1699 Address new_top = allocation_info_.top() + size_in_bytes; 1735 bool NewSpace::EnsureAllocation(int size_in_bytes, argument 1740 int aligned_size_in_bytes = size_in_bytes + filler_size; 1764 InlineAllocationStep(new_top, new_top, soon_object, size_in_bytes); 2405 bool FreeListCategory::Free(FreeSpace* free_space, size_t size_in_bytes, argument 2411 available_ += size_in_bytes; 2460 size_t FreeList::Free(Address start, size_t size_in_bytes, FreeMode mode) { argument 2461 if (size_in_bytes 2537 FindNodeFor(size_t size_in_bytes, size_t* node_size) argument 2574 Allocate(size_t size_in_bytes) argument 2810 SweepAndRetryAllocation(int size_in_bytes) argument 2824 SweepAndRetryAllocation(int size_in_bytes) argument 2833 SlowAllocateRaw(int size_in_bytes) argument [all...] |
H A D | mark-compact.cc | 1773 inline AllocationResult AllocateInNewSpace(int size_in_bytes, argument 1777 heap_->new_space()->AllocateRawSynchronized(size_in_bytes, alignment); 1782 allocation = heap_->new_space()->AllocateRawSynchronized(size_in_bytes, 1792 inline AllocationResult AllocateInOldSpace(int size_in_bytes, argument 1795 compaction_spaces_->Get(OLD_SPACE)->AllocateRaw(size_in_bytes, 1804 inline AllocationResult AllocateInLab(int size_in_bytes, argument 1813 allocation = buffer_.AllocateRawAligned(size_in_bytes, alignment); 1819 allocation = buffer_.AllocateRawAligned(size_in_bytes, alignment);
|
H A D | spaces.h | 171 bool Free(FreeSpace* node, size_t size_in_bytes, FreeMode mode); 1667 // Adds a node on the free list. The block of size {size_in_bytes} starting 1673 size_t Free(Address start, size_t size_in_bytes, FreeMode mode); 1675 // Allocate a block of size {size_in_bytes} from the free list. The block is 1678 MUST_USE_RESULT HeapObject* Allocate(size_t size_in_bytes); 1774 FreeSpace* FindNodeFor(size_t size_in_bytes, size_t* node_size); 1789 FreeListCategoryType SelectFreeListCategoryType(size_t size_in_bytes) { argument 1790 if (size_in_bytes <= kTiniestListMax) { 1792 } else if (size_in_bytes <= kTinyListMax) { 1794 } else if (size_in_bytes < 1805 SelectFastAllocationFreeListCategoryType( size_t size_in_bytes) argument 2007 Free(Address start, size_t size_in_bytes) argument 2014 UnaccountedFree(Address start, size_t size_in_bytes) argument [all...] |
/external/swiftshader/third_party/LLVM/include/llvm/ADT/ |
H A D | SmallVector.h | 86 /// size_in_bytes - This returns size()*sizeof(T). 87 size_t size_in_bytes() const { function in class:llvm::SmallVectorBase
|
/external/v8/src/snapshot/ |
H A D | deserializer.cc | 762 int size_in_bytes = source_.GetInt(); local 764 source_.CopyRaw(raw_data_out, size_in_bytes); 832 int size_in_bytes = (data - kFixedRawDataStart) << kPointerSizeLog2; local 833 source_.CopyRaw(raw_data_out, size_in_bytes); 834 current = reinterpret_cast<Object**>(raw_data_out + size_in_bytes);
|
/external/llvm/include/llvm/ADT/ |
H A D | SmallVector.h | 48 size_t size_in_bytes() const { function in class:llvm::SmallVectorBase
|
/external/google-breakpad/src/processor/ |
H A D | minidump.cc | 211 static inline void Swap(uint16_t* data, size_t size_in_bytes) { argument 212 size_t data_length = size_in_bytes / sizeof(data[0]);
|
/external/mesa3d/src/gallium/drivers/r300/ |
H A D | r300_context.h | 362 unsigned size_in_bytes; member in struct:r300_texture_desc
|
/external/v8/src/compiler/ |
H A D | wasm-compiler.cc | 1097 size_t size_in_bytes) { 1098 switch (size_in_bytes) { 1096 ReverseBytesSupported(MachineOperatorBuilder* m, size_t size_in_bytes) argument
|
/external/vixl/src/aarch64/ |
H A D | simulator-aarch64.cc | 616 // Only the least-significant `size_in_bytes` bytes of the register are printed, 619 // For typical register updates, size_in_bytes should be set to kXRegSizeInBytes 621 // size_in_bytes are intended for use when the register hasn't actually been 628 int size_in_bytes) { 634 unsigned padding_chars = (kXRegSizeInBytes - size_in_bytes) * 2; 638 switch (size_in_bytes) { 668 bits &= kXRegMask >> ((kXRegSizeInBytes - size_in_bytes) * 8); 671 int chars = size_in_bytes * 2; 626 PrintRegisterRawHelper(unsigned code, Reg31Mode r31mode, int size_in_bytes) argument
|
/external/vixl/src/aarch32/ |
H A D | disasm-aarch32.cc | 67776 DisassembleA32Buffer(const uint32_t* buffer, size_t size_in_bytes) argument 67788 DisassembleT32Buffer(const uint16_t* buffer, size_t size_in_bytes) argument [all...] |
/external/clang/lib/CodeGen/ |
H A D | CGObjCMac.cpp | 2465 CharUnits size_in_bytes = local 2471 size_in_bytes += gap; 2475 residue_in_bytes = size_in_bytes % WordSizeInBytes; 2476 size_in_bytes -= residue_in_bytes; 2480 unsigned size_in_words = size_in_bytes.getQuantity() / WordSizeInBytes;
|
/external/v8/src/ |
H A D | code-stub-assembler.cc | 723 Node* CodeStubAssembler::AllocateRawUnaligned(Node* size_in_bytes, argument 735 Node* new_top = IntPtrAdd(top, size_in_bytes); 747 SmiTag(size_in_bytes), runtime_flags); 750 NoContextConstant(), SmiTag(size_in_bytes)); 769 Node* CodeStubAssembler::AllocateRawAligned(Node* size_in_bytes, argument 776 adjusted_size.Bind(size_in_bytes); 785 IntPtrAdd(size_in_bytes, IntPtrConstant(kPointerSize)); 800 Branch(IntPtrEqual(adjusted_size.value(), size_in_bytes), &doesnt_need_filler, 823 Node* CodeStubAssembler::Allocate(Node* size_in_bytes, AllocationFlags flags) { argument 837 return AllocateRawAligned(size_in_bytes, flag 844 Allocate(int size_in_bytes, AllocationFlags flags) argument 1871 AllocateUninitializedJSArray(ElementsKind kind, Node* array_map, Node* length, Node* allocation_site, Node* size_in_bytes) argument [all...] |