/external/valgrind/massif/tests/ |
H A D | basic2.post.exp | 36 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 39 00.00% (0B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 42 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 45 98.04% (3,600B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 49 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 52 98.04% (7,600B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 56 n time(B) total(B) useful-heap( [all...] |
H A D | ignored.post.exp | 3 Massif arguments: --stacks=no --time-unit=B --heap-admin=0 --massif-out-file=massif.out --ignore-fn=ignore1 --ignore-fn=ignore2 --ignore-fn=__part_load_locale --ignore-fn=__time_load_locale --ignore-fn=dwarf2_unwind_dyld_add_image_hook --ignore-fn=get_or_create_key_element 36 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 42 100.00% (800B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 48 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B)
|
H A D | one.post.exp | 3 Massif arguments: --stacks=no --time-unit=B --heap-admin=0 --massif-out-file=massif.out --ignore-fn=__part_load_locale --ignore-fn=__time_load_locale --ignore-fn=dwarf2_unwind_dyld_add_image_hook --ignore-fn=get_or_create_key_element 36 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B)
|
H A D | custom_alloc.post.exp | 3 Massif arguments: --stacks=no --time-unit=B --heap-admin=16 --massif-out-file=massif.out --ignore-fn=__part_load_locale --ignore-fn=__time_load_locale --ignore-fn=dwarf2_unwind_dyld_add_image_hook --ignore-fn=get_or_create_key_element 36 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 41 96.15% (400B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 46 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 51 98.04% (800B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 58 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B)
|
H A D | realloc.post.exp | 3 Massif arguments: --stacks=no --heap-admin=0 --time-unit=B --threshold=0 --massif-out-file=massif.out --ignore-fn=__part_load_locale --ignore-fn=__time_load_locale --ignore-fn=dwarf2_unwind_dyld_add_image_hook --ignore-fn=get_or_create_key_element 36 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 42 100.00% (800B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 48 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 53 100.00% (1,200B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 63 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B)
|
H A D | zero1.post.exp | 3 Massif arguments: --stacks=no --heap-admin=0 --time-unit=B --massif-out-file=massif.out --ignore-fn=__part_load_locale --ignore-fn=__time_load_locale --ignore-fn=dwarf2_unwind_dyld_add_image_hook --ignore-fn=get_or_create_key_element 36 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 48 00.00% (0B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 52 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 64 00.00% (0B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 68 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B)
|
H A D | zero2.post.exp | 3 Massif arguments: --stacks=no --heap-admin=0 --time-unit=B --massif-out-file=massif.out --ignore-fn=__part_load_locale --ignore-fn=__time_load_locale --ignore-fn=dwarf2_unwind_dyld_add_image_hook --ignore-fn=get_or_create_key_element 36 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 48 00.00% (0B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 52 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B) 64 00.00% (0B) (heap allocation functions) malloc/new/new[], --alloc-fns, etc. 68 n time(B) total(B) useful-heap(B) extra-heap(B) stacks(B)
|
/external/v8/src/heap/ |
H A D | scavenger-inl.h | 8 #include "src/heap/scavenger.h" 14 DCHECK(object->GetIsolate()->heap()->InFromSpace(object)); 16 // We use the first word (where the map pointer usually is) of a heap 26 DCHECK(object->GetIsolate()->heap()->InFromSpace(*p)); 40 SlotCallbackResult Scavenger::CheckAndScavengeObject(Heap* heap, argument 44 if (heap->InFromSpace(object)) { 55 if (heap->InToSpace(object)) { 65 void StaticScavengeVisitor::VisitPointer(Heap* heap, HeapObject* obj, argument 68 if (!heap->InNewSpace(object)) return;
|
H A D | scavenger.h | 8 #include "src/heap/objects-visiting.h" 9 #include "src/heap/slot-set.h" 19 explicit Scavenger(Heap* heap) : heap_(heap) {} argument 26 // ensure the precondition that the object is (a) a heap object and (b) in 27 // the heap's from space. 29 static inline SlotCallbackResult CheckAndScavengeObject(Heap* heap, 36 // of the heap (i.e. incremental marking, logging and profiling). 40 Heap* heap() { return heap_; } function in class:v8::internal::Scavenger 52 explicit ScavengeVisitor(Heap* heap) argument [all...] |
H A D | incremental-marking-job.h | 25 static void Step(Heap* heap); 37 void Start(Heap* heap); 41 void ScheduleTask(Heap* heap);
|
H A D | memory-reducer.cc | 5 #include "src/heap/memory-reducer.h" 8 #include "src/heap/gc-tracer.h" 9 #include "src/heap/heap-inl.h" 24 : CancelableTask(memory_reducer->heap()->isolate()), 29 Heap* heap = memory_reducer_->heap(); local 31 double time_ms = heap->MonotonicallyIncreasingTimeInMs(); 32 heap->tracer()->SampleAllocation(time_ms, heap [all...] |
H A D | objects-visiting-inl.h | 8 #include "src/heap/array-buffer-tracker.h" 9 #include "src/heap/mark-compact.h" 10 #include "src/heap/objects-visiting.h" 217 Heap* heap, HeapObject* object, Address entry_address) { 219 heap->mark_compact_collector()->RecordCodeEntrySlot(object, entry_address, 221 StaticVisitor::MarkObject(heap, code); 227 Heap* heap, RelocInfo* rinfo) { 231 heap->mark_compact_collector()->RecordRelocSlot(host, rinfo, object); 236 StaticVisitor::MarkObject(heap, object); 242 void StaticMarkingVisitor<StaticVisitor>::VisitCell(Heap* heap, argument 216 VisitCodeEntry( Heap* heap, HeapObject* object, Address entry_address) argument 226 VisitEmbeddedPointer( Heap* heap, RelocInfo* rinfo) argument 255 VisitDebugTarget(Heap* heap, RelocInfo* rinfo) argument 267 VisitCodeTarget(Heap* heap, RelocInfo* rinfo) argument 277 VisitCodeAgeSequence( Heap* heap, RelocInfo* rinfo) argument 306 Heap* heap = map->GetHeap(); local 329 Heap* heap = map->GetHeap(); local 358 Heap* heap = array->GetHeap(); local 371 Heap* heap = map->GetHeap(); local 384 Heap* heap = map->GetHeap(); local 414 Heap* heap = map->GetHeap(); local 426 Heap* heap = map->GetHeap(); local 453 Heap* heap = map->GetHeap(); local 477 MarkMapContents(Heap* heap, Map* map) argument 511 HasSourceCode(Heap* heap, SharedFunctionInfo* info) argument 519 IsFlushable(Heap* heap, JSFunction* function) argument 544 IsFlushable( Heap* heap, SharedFunctionInfo* shared_info) argument [all...] |
H A D | objects-visiting.cc | 5 #include "src/heap/objects-visiting.h" 7 #include "src/heap/heap-inl.h" 8 #include "src/heap/mark-compact-inl.h" 9 #include "src/heap/objects-visiting-inl.h" 222 static bool MustRecordSlots(Heap* heap) { argument 223 return heap->gc_state() == Heap::MARK_COMPACT && 224 heap->mark_compact_collector()->is_compacting(); 233 Object* VisitWeakList(Heap* heap, Object* list, WeakObjectRetainer* retainer) { argument 234 Object* undefined = heap 282 ClearWeakList(Heap* heap, Object* list) argument 340 VisitLiveObject(Heap* heap, Context* context, WeakObjectRetainer* retainer) argument 363 DoWeakList(Heap* heap, Context* context, WeakObjectRetainer* retainer, int index) argument 379 VisitPhantomObject(Heap* heap, Context* context) argument [all...] |
H A D | array-buffer-tracker.h | 33 inline static void RegisterNew(Heap* heap, JSArrayBuffer* buffer); 34 inline static void Unregister(Heap* heap, JSArrayBuffer* buffer); 38 static void FreeDeadInNewSpace(Heap* heap); 68 explicit LocalArrayBufferTracker(Heap* heap) : heap_(heap) {} argument
|
H A D | mark-compact.cc | 5 #include "src/heap/mark-compact.h" 17 #include "src/heap/array-buffer-tracker.h" 18 #include "src/heap/gc-tracer.h" 19 #include "src/heap/incremental-marking.h" 20 #include "src/heap/mark-compact-inl.h" 21 #include "src/heap/object-stats.h" 22 #include "src/heap/objects-visiting-inl.h" 23 #include "src/heap/objects-visiting.h" 24 #include "src/heap/page-parallel-job.h" 25 #include "src/heap/space 50 MarkCompactCollector(Heap* heap) argument 97 VerifyMarking(Heap* heap, Address bottom, Address top) argument 150 VerifyMarking(Heap* heap) argument 212 VerifyEvacuation(Heap* heap, PagedSpace* space) argument 223 VerifyEvacuation(Heap* heap) argument 1060 Heap* heap = isolate_->heap(); local 1076 Initialize(Heap* heap) argument 1080 VisitPointer(Heap* heap, HeapObject* object, Object** p) argument 1089 MarkRecursively(Heap* heap, HeapObject* object) argument 1105 INLINE(static void VisitPointer(Heap* heap, HeapObject* object, Object** p)) argument 1109 INLINE(static void VisitPointers(Heap* heap, HeapObject* object, Object** start, Object** end)) argument 1124 INLINE(static void MarkObject(Heap* heap, HeapObject* object)) argument 1130 INLINE(static bool MarkObjectWithoutPush(Heap* heap, HeapObject* object)) argument 1156 Heap* heap = obj->GetHeap(); local 1165 INLINE(static bool VisitUnmarkedObjects(Heap* heap, HeapObject* object, Object** start, Object** end)) argument 1189 UpdateRegExpCodeAgeAndFlush(Heap* heap, JSRegExp* re, bool is_one_byte) argument 1245 Heap* heap = map->GetHeap(); local 1350 RootMarkingVisitor(Heap* heap) argument 1403 StringTableCleaner(Heap* heap, HeapObject* table) argument 1613 EvacuateVisitorBase(Heap* heap, CompactionSpaceCollection* compaction_spaces) argument 1714 EvacuateNewSpaceVisitor(Heap* heap, CompactionSpaceCollection* compaction_spaces, base::HashMap* local_pretenuring_feedback) argument 1854 EvacuateNewSpacePageVisitor( Heap* heap, base::HashMap* local_pretenuring_feedback) argument 1897 EvacuateOldSpaceVisitor(Heap* heap, CompactionSpaceCollection* compaction_spaces) argument 1916 EvacuateRecordOnlyVisitor(Heap* heap) argument 1952 IsUnmarkedHeapObjectWithHeap(Heap* heap, Object** p) argument 2232 ObjectStatsVisitor(Heap* heap, ObjectStats* live_stats, ObjectStats* dead_stats) argument 2293 CheckAndMarkObject( Heap* heap, Address slot_address) argument 2312 IsUnmarkedObject(Heap* heap, Object** p) argument 2813 Heap* heap = this->heap(); local 2987 UpdateReferenceInExternalStringTableEntry(Heap* heap, Object** p) argument 3074 inline Heap* heap() { return collector_->heap(); } function in class:v8::MarkCompactCollector::Evacuator 3105 Heap* heap = page->heap(); local 3226 ProcessPageInParallel(Heap* heap, PerTaskData evacuator, MemoryChunk* chunk, PerPageData) argument 3231 FinalizePageSequentially(Heap* heap, MemoryChunk* chunk, bool success, PerPageData data) argument 3645 ProcessPageInParallel(Heap* heap, PerTaskData, MemoryChunk* chunk, PerPageData) argument 3656 UpdateUntypedPointers(Heap* heap, MemoryChunk* chunk) argument 3668 UpdateTypedPointers(Heap* heap, MemoryChunk* chunk) argument 3690 CheckAndUpdateOldToNewSlot(Heap* heap, Address slot_address) argument 3748 UpdatePointersInParallel(Heap* heap, base::Semaphore* semaphore) argument 3763 ProcessPageInParallel(Heap* heap, PerTaskData visitor, MemoryChunk* chunk, PerPageData limits) argument 3780 ProcessPageInParallelVisitAll(Heap* heap, PerTaskData visitor, MemoryChunk* chunk, PerPageData limits) argument 3792 ProcessPageInParallelVisitLive(Heap* heap, PerTaskData visitor, MemoryChunk* chunk, PerPageData limits) argument 3805 UpdateToSpacePointersInParallel(Heap* heap, base::Semaphore* semaphore) argument 3836 Heap* heap = this->heap(); local [all...] |
H A D | scavenge-job.h | 10 #include "src/heap/gc-tracer.h" 40 void ScheduleIdleTaskIfNeeded(Heap* heap, int bytes_allocated); 45 void RescheduleIdleTask(Heap* heap); 73 void ScheduleIdleTask(Heap* heap);
|
H A D | incremental-marking.cc | 5 #include "src/heap/incremental-marking.h" 10 #include "src/heap/gc-idle-time-handler.h" 11 #include "src/heap/gc-tracer.h" 12 #include "src/heap/heap-inl.h" 13 #include "src/heap/mark-compact-inl.h" 14 #include "src/heap/object-stats.h" 15 #include "src/heap/objects-visiting-inl.h" 16 #include "src/heap/objects-visiting.h" 23 IncrementalMarking::IncrementalMarking(Heap* heap) argument 132 TransferMark(Heap* heap, HeapObject* from, HeapObject* to) argument 186 Heap* heap = map->GetHeap(); local 235 INLINE(static void VisitPointer(Heap* heap, HeapObject* object, Object** p)) argument 243 INLINE(static void VisitPointers(Heap* heap, HeapObject* object, Object** start, Object** end)) argument 255 INLINE(static void MarkObject(Heap* heap, Object* obj)) argument 261 INLINE(static bool MarkObjectWithoutPush(Heap* heap, Object* obj)) argument 419 PatchIncrementalMarkingRecordWriteStubs( Heap* heap, RecordWriteStub::Mode mode) argument 840 MarkGrey(Heap* heap, HeapObject* object) argument [all...] |
/external/libdrm/intel/ |
H A D | mm.h | 41 struct mem_block *heap; member in struct:mem_block 49 * return: a heap pointer if OK, NULL if error 60 * startSearch = linear offset from start of heap to begin search 63 drm_private extern struct mem_block *mmAllocMem(struct mem_block *heap,
|
/external/syslinux/efi/ |
H A D | mem.c | 9 void *efi_malloc(size_t size, enum heap heap, malloc_tag_t tag) argument
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | heap_simulator_test.cc | 43 // HeapCallRecorder is a dummy heap algorithm that simply records its calls. 70 // HeapSimulatorTracker runs the heap simulator, recording the sequence of calls 71 // made to the underlying heap algorithm. Tests compare the actual call 493 // Base class for heap algorithm tests. 520 // Create a dummy LogicalBuffer to pass to the heap algorithm. 536 NoFragmentationStatsHeap heap; local 537 EXPECT_EQ(0, heap.Finish().heap_size); 541 NoFragmentationStatsHeap heap; local 542 heap.Alloc(buffer_a_, 10); 543 heap 554 NoFragmentationStatsHeap heap; local [all...] |
/external/bzip2/ |
H A D | huffman.c | 36 zz = z; tmp = heap[zz]; \ 37 while (weight[tmp] < weight[heap[zz >> 1]]) { \ 38 heap[zz] = heap[zz >> 1]; \ 41 heap[zz] = tmp; \ 47 zz = z; tmp = heap[zz]; \ 52 weight[heap[yy+1]] < weight[heap[yy]]) \ 54 if (weight[tmp] < weight[heap[yy]]) break; \ 55 heap[z 75 Int32 heap [ BZ_MAX_ALPHA_SIZE + 2 ]; local [all...] |
/external/python/cpython2/Lib/test/ |
H A D | test_heapq.py | 33 heap = [] 35 self.check_invariant(heap) 39 self.module.heappush(heap, item) 40 self.check_invariant(heap) 42 while heap: 43 item = self.module.heappop(heap) 44 self.check_invariant(heap) 59 def check_invariant(self, heap): 60 # Check the heap invariant. 61 for pos, item in enumerate(heap) [all...] |
/external/python/cpython3/Lib/test/ |
H A D | test_heapq.py | 33 heap = [] 35 self.check_invariant(heap) 39 self.module.heappush(heap, item) 40 self.check_invariant(heap) 42 while heap: 43 item = self.module.heappop(heap) 44 self.check_invariant(heap) 59 def check_invariant(self, heap): 60 # Check the heap invariant. 61 for pos, item in enumerate(heap) [all...] |
/external/v8/src/extensions/ |
H A D | statistics-extension.cc | 8 #include "src/heap/heap-inl.h" 63 Heap* heap = isolate->heap(); local 70 heap->CollectAllGarbage(Heap::kNoGCFlags, 120 {heap->memory_allocator()->Size(), "total_committed_bytes"}, 121 {heap->new_space()->Size(), "new_space_live_bytes"}, 122 {heap->new_space()->Available(), "new_space_available_bytes"}, 123 {heap->new_space()->CommittedMemory(), "new_space_commited_bytes"}, 124 {heap [all...] |
/external/compiler-rt/test/asan/TestCases/Linux/ |
H A D | swapcontext_test.cc | 80 char *heap = new char[kStackSize + 1]; local 81 ret += Run(argc - 1, 0, heap); 84 ret += Run(argc - 1, 1, heap); 88 delete [] heap;
|