/external/chromium_org/gpu/command_buffer/client/ |
H A D | mapped_memory.cc | 16 MemoryChunk::MemoryChunk(int32 shm_id, function in class:gpu::MemoryChunk 24 MemoryChunk::~MemoryChunk() {} 40 MemoryChunk* chunk = *iter; 53 MemoryChunk* chunk = chunks_[ii]; 72 MemoryChunk* chunk = chunks_[ii]; 95 MemoryChunk* mc = new MemoryChunk(id, shm, helper_, poll_callback_); 107 MemoryChunk* chun [all...] |
H A D | mapped_memory.h | 22 class GPU_EXPORT MemoryChunk { class in namespace:gpu 24 MemoryChunk(int32_t shm_id, 28 ~MemoryChunk(); 115 DISALLOW_COPY_AND_ASSIGN(MemoryChunk); 176 MemoryChunk* chunk = chunks_[ii]; 188 typedef ScopedVector<MemoryChunk> MemoryChunkVector;
|
H A D | mapped_memory_unittest.cc | 94 // Test fixture for MemoryChunk test - Creates a MemoryChunk, using a 106 chunk_.reset(new MemoryChunk(kShmId, 121 scoped_ptr<MemoryChunk> chunk_;
|
/external/chromium_org/v8/src/heap/ |
H A D | incremental-marking-inl.h | 21 MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address()); 22 if (chunk->IsFlagSet(MemoryChunk::HAS_PROGRESS_BAR)) { 70 MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address()); 71 if (chunk->IsFlagSet(MemoryChunk::HAS_PROGRESS_BAR)) { 88 MemoryChunk::IncrementLiveBytesFromGC(obj->address(), -obj_size);
|
H A D | mark-compact-inl.h | 17 MemoryChunk* p = MemoryChunk::FromAddress(addr); 34 MemoryChunk::IncrementLiveBytesFromGC(obj->address(), obj->Size()); 46 MemoryChunk::IncrementLiveBytesFromGC(obj->address(), obj->Size());
|
H A D | spaces-inl.h | 21 void Bitmap::Clear(MemoryChunk* chunk) { 136 Page* Page::Initialize(Heap* heap, MemoryChunk* chunk, Executability executable, 157 void MemoryChunk::set_scan_on_scavenge(bool scan) { 169 MemoryChunk* MemoryChunk::FromAnyPointerAddress(Heap* heap, Address addr) { 170 MemoryChunk* maybe = reinterpret_cast<MemoryChunk*>( 178 MemoryChunk* chunk = MemoryChunk::FromAddress(o->address()); 189 void MemoryChunk [all...] |
H A D | incremental-marking.cc | 55 MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address()); 57 if (counter < (MemoryChunk::kWriteBarrierCounterGranularity / 2)) { 59 MemoryChunk::kWriteBarrierCounterGranularity - 62 MemoryChunk::kWriteBarrierCounterGranularity); 130 MemoryChunk::IncrementLiveBytesFromGC(heap_obj->address(), 143 MemoryChunk::IncrementLiveBytesFromGC(heap_object->address(), size); 153 MemoryChunk::IncrementLiveBytesFromGC(heap_object->address(), size); 171 MemoryChunk* chunk = MemoryChunk [all...] |
H A D | store-buffer.cc | 146 MemoryChunk* chunk; 155 Filter(MemoryChunk::SCAN_ON_SCAVENGE); 186 MemoryChunk* chunk; 191 MemoryChunk* previous_chunk = NULL; 194 MemoryChunk* containing_chunk = NULL; 198 containing_chunk = MemoryChunk::FromAnyPointerAddress(heap_, addr); 209 Filter(MemoryChunk::SCAN_ON_SCAVENGE); 217 MemoryChunk* previous_chunk = NULL; 220 MemoryChunk* containing_chunk = NULL; 224 containing_chunk = MemoryChunk [all...] |
H A D | spaces.cc | 128 MemoryChunk::kAlignment); 198 size_t aligned_requested = RoundUp(requested_size, MemoryChunk::kAlignment); 207 DCHECK(IsAddressAligned(current.start, MemoryChunk::kAlignment)); 234 DCHECK(IsAddressAligned(address, MemoryChunk::kAlignment)); 403 MemoryChunk* chunk = 404 MemoryChunk::Initialize(heap, start, Page::kPageSize, area_start, 410 chunk->SetFlag(in_to_space ? MemoryChunk::IN_TO_SPACE 411 : MemoryChunk::IN_FROM_SPACE); 412 DCHECK(!chunk->IsFlagSet(in_to_space ? MemoryChunk::IN_FROM_SPACE 413 : MemoryChunk [all...] |
H A D | store-buffer-inl.h | 47 (*callback_)(heap_, MemoryChunk::FromAnyPointerAddress(heap_, addr),
|
H A D | spaces.h | 99 class MemoryChunk; 196 static inline void Clear(MemoryChunk* chunk); 269 // MemoryChunk represents a memory region owned by a specific space. 273 class MemoryChunk { class in namespace:v8::internal 275 // Only works if the pointer is in the first kPageSize of the MemoryChunk. 276 static MemoryChunk* FromAddress(Address a) { 277 return reinterpret_cast<MemoryChunk*>(OffsetFrom(a) & ~kAlignmentMask); 279 static const MemoryChunk* FromAddress(const byte* a) { 280 return reinterpret_cast<const MemoryChunk*>(OffsetFrom(a) & 285 static inline MemoryChunk* FromAnyPointerAddres [all...] |
H A D | incremental-marking.h | 130 inline void SetOldSpacePageFlags(MemoryChunk* chunk) { 185 static void SetOldSpacePageFlags(MemoryChunk* chunk, bool is_marking,
|
H A D | heap-inl.h | 279 MemoryChunk* memory_chunk = MemoryChunk::FromAddress(object_address); 373 return page->IsFlagSet(MemoryChunk::NEW_SPACE_BELOW_AGE_MARK) && 444 MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address());
|
H A D | mark-compact.h | 173 MemoryChunk::IncrementLiveBytesFromGC(object->address(), -object->Size()); 899 explicit MarkBitCellIterator(MemoryChunk* chunk) : chunk_(chunk) { 930 MemoryChunk* chunk_;
|
H A D | heap.cc | 1302 void Heap::ScavengeStoreBufferCallback(Heap* heap, MemoryChunk* page, 1308 void StoreBufferRebuilder::Callback(MemoryChunk* page, StoreBufferEvent event) { 1362 DCHECK((Page::kPageSize - MemoryChunk::kBodyOffset) % (2 * kPointerSize) == 1943 MemoryChunk::IncrementLiveBytesFromGC(target->address(), size); 2332 MemoryChunk* chunk = MemoryChunk::FromAddress(obj->address()); 3253 MemoryChunk::IncrementLiveBytesFromGC(address, by); 3255 MemoryChunk::IncrementLiveBytesFromMutator(address, by); 3440 MemoryChunk::FromAddress(address)->owner()->identity() != LO_SPACE) { 6019 void Heap::QueueMemoryChunkForFree(MemoryChunk* chun [all...] |
H A D | mark-compact.cc | 653 DCHECK(MemoryChunk::FromAddress(old_start) == 654 MemoryChunk::FromAddress(new_start)); 925 p->ClearFlag(MemoryChunk::RESCAN_ON_EVACUATION); 1890 MemoryChunk::IncrementLiveBytesFromGC(object->address(), object->Size()); 1902 MemoryChunk* p) { 1935 MemoryChunk::IncrementLiveBytesFromGC(object->address(), object->Size()); 3176 MemoryChunk::SWEEPING_DONE); 3379 p->set_parallel_sweeping(MemoryChunk::SWEEPING_FINALIZE); 3397 uint32_t start_index = MemoryChunk::FastAddressToMarkbitIndex(code_start); 3399 MemoryChunk [all...] |
H A D | heap.h | 354 void Callback(MemoryChunk* page, StoreBufferEvent event); 365 MemoryChunk* current_page_; 1225 void QueueMemoryChunkForFree(MemoryChunk* chunk); 1849 static void ScavengeStoreBufferCallback(Heap* heap, MemoryChunk* page, 2019 MemoryChunk* chunks_queued_for_free_;
|
/external/chromium_org/v8/test/cctest/ |
H A D | test-spaces.cc | 165 : MemoryChunk::kObjectStartOffset; 170 MemoryChunk* memory_chunk = memory_allocator->AllocateChunk(reserve_area_size, 175 MemoryChunk::kAlignment : v8::base::OS::CommitPageSize(); 240 TEST(MemoryChunk) {
|
H A D | test-constantpool.cc | 256 MemoryChunk::FromAddress(array->address())->set_scan_on_scavenge(true);
|
/external/chromium_org/v8/src/ia32/ |
H A D | macro-assembler-ia32.cc | 138 DCHECK(MemoryChunk::IN_FROM_SPACE < 8); 139 DCHECK(MemoryChunk::IN_TO_SPACE < 8); 140 int mask = (1 << MemoryChunk::IN_FROM_SPACE) 141 | (1 << MemoryChunk::IN_TO_SPACE); 143 test_b(Operand(scratch, MemoryChunk::kFlagsOffset), 483 MemoryChunk::kPointersToHereAreInterestingMask, 546 MemoryChunk::kPointersToHereAreInterestingMask, 553 MemoryChunk::kPointersFromHereAreInterestingMask, 3074 test_b(Operand(scratch, MemoryChunk::kFlagsOffset), 3077 test(Operand(scratch, MemoryChunk [all...] |
/external/chromium_org/v8/src/x87/ |
H A D | macro-assembler-x87.cc | 138 DCHECK(MemoryChunk::IN_FROM_SPACE < 8); 139 DCHECK(MemoryChunk::IN_TO_SPACE < 8); 140 int mask = (1 << MemoryChunk::IN_FROM_SPACE) 141 | (1 << MemoryChunk::IN_TO_SPACE); 143 test_b(Operand(scratch, MemoryChunk::kFlagsOffset), 406 MemoryChunk::kPointersToHereAreInterestingMask, 465 MemoryChunk::kPointersToHereAreInterestingMask, 472 MemoryChunk::kPointersFromHereAreInterestingMask, 3034 test_b(Operand(scratch, MemoryChunk::kFlagsOffset), 3037 test(Operand(scratch, MemoryChunk [all...] |
/external/chromium_org/v8/src/ |
H A D | globals.h | 310 class MemoryChunk; 501 MemoryChunk* page,
|
/external/chromium_org/v8/src/arm/ |
H A D | macro-assembler-arm.cc | 540 MemoryChunk::kPointersToHereAreInterestingMask, 614 MemoryChunk::kPointersToHereAreInterestingMask, 620 MemoryChunk::kPointersFromHereAreInterestingMask, 3650 ldr(scratch, MemOperand(scratch, MemoryChunk::kFlagsOffset)); 3688 ldr(ip, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 3699 ldr(ip, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize + kPointerSize)); 3760 ldr(load_scratch, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 3821 ldr(ip, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 3823 str(ip, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 3826 ldr(ip, MemOperand(bitmap_scratch, MemoryChunk [all...] |
/external/chromium_org/v8/src/arm64/ |
H A D | macro-assembler-arm64.cc | 4337 Ldr(scratch, MemOperand(scratch, MemoryChunk::kFlagsOffset)); 4347 Ldr(scratch, MemOperand(scratch, MemoryChunk::kFlagsOffset)); 4445 MemoryChunk::kPointersToHereAreInterestingMask, 4514 MemoryChunk::kPointersToHereAreInterestingMask, 4519 MemoryChunk::kPointersFromHereAreInterestingMask, 4595 Ldr(bitmap_scratch, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 4708 Ldr(load_scratch, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 4771 Ldr(load_scratch, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 4773 Str(load_scratch, MemOperand(bitmap_scratch, MemoryChunk::kHeaderSize)); 4776 Ldr(load_scratch, MemOperand(bitmap_scratch, MemoryChunk [all...] |
/external/chromium_org/v8/src/x64/ |
H A D | macro-assembler-x64.cc | 419 MemoryChunk::kPointersToHereAreInterestingMask, 481 MemoryChunk::kPointersToHereAreInterestingMask, 489 MemoryChunk::kPointersFromHereAreInterestingMask, 5048 testb(Operand(scratch, MemoryChunk::kFlagsOffset), 5051 testl(Operand(scratch, MemoryChunk::kFlagsOffset), Immediate(mask)); 5085 andp(rcx, Operand(bitmap_scratch, MemoryChunk::kHeaderSize)); 5157 testp(Operand(bitmap_scratch, MemoryChunk::kHeaderSize), mask_scratch); 5166 testp(Operand(bitmap_scratch, MemoryChunk::kHeaderSize), mask_scratch); 5225 orp(Operand(bitmap_scratch, MemoryChunk::kHeaderSize), mask_scratch); 5228 addl(Operand(bitmap_scratch, MemoryChunk [all...] |