/art/runtime/gc/collector/ |
H A D | immune_region.cc | 19 #include "gc/space/space-inl.h" 35 bool ImmuneRegion::AddContinuousSpace(space::ContinuousSpace* space) { argument 37 if (space->GetLiveBitmap() != space->GetMarkBitmap()) { 38 CHECK(space->IsContinuousMemMapAllocSpace()); 39 space->AsContinuousMemMapAllocSpace()->BindLiveToMarkBitmap(); 41 mirror::Object* space_begin = reinterpret_cast<mirror::Object*>(space->Begin()); 42 mirror::Object* space_limit = reinterpret_cast<mirror::Object*>(space [all...] |
H A D | partial_mark_sweep.cc | 20 #include "gc/space/space.h" 37 // For partial GCs we need to bind the bitmap of the zygote space so that all objects in the 38 // zygote space are viewed as marked. 39 for (const auto& space : GetHeap()->GetContinuousSpaces()) { 40 if (space->GetGcRetentionPolicy() == space::kGcRetentionPolicyFullCollect) { 41 CHECK(space->IsZygoteSpace()); 42 CHECK(immune_region_.AddContinuousSpace(space)) << "Failed to add space " << *spac [all...] |
H A D | sticky_mark_sweep.cc | 18 #include "gc/space/large_object_space.h" 19 #include "gc/space/space.h" 38 // know what was allocated since the last GC. A side-effect of binding the allocation space mark 40 for (const auto& space : GetHeap()->GetContinuousSpaces()) { 41 if (space->IsContinuousMemMapAllocSpace() && 42 space->GetGcRetentionPolicy() == space::kGcRetentionPolicyAlwaysCollect) { 43 DCHECK(space->IsContinuousMemMapAllocSpace()); 44 space [all...] |
H A D | immune_region.h | 28 namespace space { namespace in namespace:art::gc 30 } // namespace space 43 bool AddContinuousSpace(space::ContinuousSpace* space) 45 bool ContainsSpace(const space::ContinuousSpace* space) const;
|
H A D | garbage_collector.cc | 25 #include "gc/space/large_object_space.h" 26 #include "gc/space/space-inl.h" 105 // Swap the live and mark bitmaps for each alloc space. This is needed since sweep re-swaps 109 for (const auto& space : GetHeap()->GetContinuousSpaces()) { 111 if (space->GetGcRetentionPolicy() == space::kGcRetentionPolicyAlwaysCollect || 113 space->GetGcRetentionPolicy() == space::kGcRetentionPolicyFullCollect)) { 114 accounting::ContinuousSpaceBitmap* live_bitmap = space 125 space::LargeObjectSpace* space = disc_space->AsLargeObjectSpace(); local [all...] |
H A D | mark_compact.cc | 28 #include "gc/space/bump_pointer_space.h" 29 #include "gc/space/bump_pointer_space-inl.h" 30 #include "gc/space/image_space.h" 31 #include "gc/space/large_object_space.h" 32 #include "gc/space/space-inl.h" 62 for (const auto& space : GetHeap()->GetContinuousSpaces()) { 63 if (space->GetGcRetentionPolicy() == space::kGcRetentionPolicyNeverCollect || 64 space 596 SetSpace(space::BumpPointerSpace* space) argument [all...] |
H A D | mark_sweep.cc | 35 #include "gc/space/image_space.h" 36 #include "gc/space/large_object_space.h" 37 #include "gc/space/space-inl.h" 87 for (const auto& space : GetHeap()->GetContinuousSpaces()) { 88 if (space->GetGcRetentionPolicy() == space::kGcRetentionPolicyNeverCollect) { 89 CHECK(immune_region_.AddContinuousSpace(space)) << "Failed to add space " << *space; [all...] |
H A D | semi_space.cc | 34 #include "gc/space/bump_pointer_space.h" 35 #include "gc/space/bump_pointer_space-inl.h" 36 #include "gc/space/image_space.h" 37 #include "gc/space/large_object_space.h" 38 #include "gc/space/space-inl.h" 65 for (const auto& space : GetHeap()->GetContinuousSpaces()) { 66 if (space->GetGcRetentionPolicy() == space::kGcRetentionPolicyNeverCollect || 67 space [all...] |
H A D | semi_space.h | 51 namespace space { namespace in namespace:art::gc 54 } // namespace space 83 // Sets which space we will be copying objects to. 84 void SetToSpace(space::ContinuousMemMapAllocSpace* to_space); 86 // Set the space where we copy objects from. 87 void SetFromSpace(space::ContinuousMemMapAllocSpace* from_space); 183 // Returns true if we should sweep the space. 184 virtual bool ShouldSweepSpace(space::ContinuousSpace* space) const; 203 // Current space, w [all...] |
/art/runtime/gc/accounting/ |
H A D | mod_union_table-inl.h | 22 #include "gc/space/space.h" 28 // A mod-union table to record image references to the Zygote and alloc space. 32 space::ContinuousSpace* space) 33 : ModUnionTableReferenceCache(name, heap, space) {} 31 ModUnionTableToZygoteAllocspace(const std::string& name, Heap* heap, space::ContinuousSpace* space) argument
|
H A D | remembered_set.h | 34 namespace space { namespace in namespace:art::gc 36 } // namespace space 49 explicit RememberedSet(const std::string& name, Heap* heap, space::ContinuousSpace* space) argument 50 : name_(name), heap_(heap), space_(space) {} 55 // Mark through all references to the target space. 58 space::ContinuousSpace* target_space, void* arg) 64 space::ContinuousSpace* GetSpace() { 78 space::ContinuousSpace* const space_;
|
H A D | mod_union_table.h | 38 namespace space { namespace in namespace:art::gc 41 } // namespace space 56 explicit ModUnionTable(const std::string& name, Heap* heap, space::ContinuousSpace* space) argument 59 space_(space) { 64 // Clear cards which map to a memory range of a space. This doesn't immediately update the 81 space::ContinuousSpace* GetSpace() { 94 space::ContinuousSpace* const space_; 97 // Reference caching implementation. Caches references pointing to alloc space(s) for each card. 101 space 100 ModUnionTableReferenceCache(const std::string& name, Heap* heap, space::ContinuousSpace* space) argument 136 ModUnionTableCardCache(const std::string& name, Heap* heap, space::ContinuousSpace* space) argument [all...] |
H A D | card_table.cc | 22 #include "gc/space/space.h" 74 // We allocated up to a bytes worth of extra space to allow biased_begin's byte value to equal 93 void CardTable::ClearSpaceCards(space::ContinuousSpace* space) { argument 95 byte* card_start = CardFromAddr(space->Begin()); 96 byte* card_end = CardFromAddr(space->End()); // Make sure to round up.
|
/art/runtime/gc/space/ |
H A D | space_test.h | 32 namespace space { namespace in namespace:art::gc 41 void AddSpace(ContinuousSpace* space, bool revoke = true) { argument 46 heap->AddSpace(space); 47 heap->SetSpaceAsDefault(space); 63 mirror::Object* Alloc(space::MallocSpace* alloc_space, Thread* self, size_t bytes, 75 mirror::Object* AllocWithGrowth(space::MallocSpace* alloc_space, Thread* self, size_t bytes, 119 void SizeFootPrintGrowthLimitAndTrimBody(MallocSpace* space, intptr_t object_size, 132 std::unique_ptr<Space> space(create_space("test", 16 * MB, 32 * MB, 32 * MB, nullptr)); 133 EXPECT_TRUE(space.get() != nullptr); 137 std::unique_ptr<Space> space(create_spac 375 SizeFootPrintGrowthLimitAndTrimBody(MallocSpace* space, intptr_t object_size, int round, size_t growth_limit) argument [all...] |
H A D | dlmalloc_space_base_test.cc | 24 namespace space { namespace in namespace:art::gc 34 } // namespace space
|
H A D | dlmalloc_space_random_test.cc | 23 namespace space { namespace in namespace:art::gc 33 } // namespace space
|
H A D | dlmalloc_space_static_test.cc | 23 namespace space { namespace in namespace:art::gc 33 } // namespace space
|
H A D | rosalloc_space_base_test.cc | 21 namespace space { namespace in namespace:art::gc 32 } // namespace space
|
H A D | rosalloc_space_random_test.cc | 21 namespace space { namespace in namespace:art::gc 32 } // namespace space
|
H A D | rosalloc_space_static_test.cc | 21 namespace space { namespace in namespace:art::gc 32 } // namespace space
|
H A D | space.cc | 17 #include "space.h" 27 namespace space { namespace in namespace:art::gc 36 std::ostream& operator<<(std::ostream& os, const Space& space) { argument 37 space.Dump(os); 87 // If the bitmaps are bound then sweeping this space clearly won't do anything. 136 AllocSpace::SweepCallbackContext::SweepCallbackContext(bool swap_bitmaps, space::Space* space) argument 137 : swap_bitmaps(swap_bitmaps), space(space), self(Thread::Current()) { 140 } // namespace space [all...] |
/art/runtime/base/ |
H A D | stringprintf.cc | 25 char space[1024]; local 32 int result = vsnprintf(space, sizeof(space), format, backup_ap); 35 if (result < static_cast<int>(sizeof(space))) { 38 dst->append(space, result);
|
/art/runtime/native/ |
H A D | dalvik_system_VMDebug.cc | 23 #include "gc/space/bump_pointer_space.h" 24 #include "gc/space/dlmalloc_space.h" 25 #include "gc/space/large_object_space.h" 26 #include "gc/space/space-inl.h" 27 #include "gc/space/zygote_space.h" 265 // We export the VM internal per-heap-space size/alloc/free metrics 266 // for the zygote space, alloc space (application heap), and the large 267 // object space fo [all...] |
/art/runtime/gc/ |
H A D | heap.h | 79 namespace space { namespace in namespace:art::gc 91 } // namespace space 131 // Primitive arrays larger than this size are put in the large object space. 221 // Transition the garbage collector during runtime, may copy objects from one space to another. 247 // Faster alternative to IsHeapAddress since finding if an object is in the large object space is 290 // Removes the growth limit on the alloc space so it may grow to its maximum capacity. Used to 304 // Set the heap's private space pointers to be the same as the space based on it's type. Public 306 void SetSpaceAsDefault(space::ContinuousSpace* continuous_space) 308 void AddSpace(space [all...] |
H A D | heap.cc | 46 #include "gc/space/bump_pointer_space.h" 47 #include "gc/space/dlmalloc_space-inl.h" 48 #include "gc/space/image_space.h" 49 #include "gc/space/large_object_space.h" 50 #include "gc/space/rosalloc_space-inl.h" 51 #include "gc/space/space-inl.h" 52 #include "gc/space/zygote_space.h" 85 // Whether or not we use the free list large object space. Only use it if USE_ART_LOW_4G_ALLOCATOR 99 // Define space nam 608 space::Space* space = nullptr; local 623 stream << " in space " << *space; local 741 AddSpace(space::Space* space) argument 782 RemoveSpace(space::Space* space) argument 941 space::AllocSpace* space = nullptr; local 1208 stream << space << " " << *space << "\\n"; local 1217 stream << space << " " << *space << "\\n"; local 1854 BuildBins(space::ContinuousSpace* space) argument 2361 space::MallocSpace* space = ref_space->AsMallocSpace(); variable 2677 FindModUnionTableFromSpace(space::Space* space) argument 2685 FindRememberedSetFromSpace(space::Space* space) argument 2888 space::Space* space = FindContinuousSpaceFromObject(obj, true); local 3249 space::Space* space = remembered_set->GetSpace(); local 3256 RemoveRememberedSet(space::Space* space) argument [all...] |