Copying.cpp revision fe9052edaf6bebbccaac5a9fb607012778d0dd74
1/* 2 * Copyright (C) 2009 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#include <errno.h> 18#include <limits.h> 19#include <sys/mman.h> 20 21#include "Dalvik.h" 22#include "alloc/Heap.h" 23#include "alloc/HeapBitmap.h" 24#include "alloc/HeapInternal.h" 25#include "alloc/HeapSource.h" 26#include "alloc/Verify.h" 27 28/* 29 * A "mostly copying", generational, garbage collector. 30 * 31 * TODO: we allocate our own contiguous tract of page frames to back 32 * object allocations. To cooperate with other heaps active in the 33 * virtual machine we need to move the responsibility of allocating 34 * pages someplace outside of this code. 35 * 36 * The other major data structures that maintain the state of the heap 37 * are the block space table and the block queue. 38 * 39 * The block space table records the state of a block. We must track 40 * whether a block is: 41 * 42 * - Free or allocated in some space. 43 * 44 * - If the block holds part of a large object allocation, whether the 45 * block is the initial or a continued block of the allocation. 46 * 47 * - Whether the block is pinned, that is to say whether at least one 48 * object in the block must remain stationary. Only needed during a 49 * GC. 50 * 51 * - Which space the object belongs to. At present this means 52 * from-space or to-space. 53 * 54 * The block queue is used during garbage collection. Unlike Cheney's 55 * algorithm, from-space and to-space are not contiguous. Therefore, 56 * one cannot maintain the state of the copy with just two pointers. 57 * The block queue exists to thread lists of blocks from the various 58 * spaces together. 59 * 60 * Additionally, we record the free space frontier of the heap, as 61 * well as the address of the first object within a block, which is 62 * required to copy objects following a large object (not currently 63 * implemented). This is stored in the heap source structure. This 64 * should be moved elsewhere to support in-line allocations from Java 65 * threads. 66 * 67 * Allocation requests are satisfied by reserving storage from one or 68 * more contiguous blocks. Objects that are small enough to fit 69 * inside a block are packed together within a block. Objects that 70 * are larger than a block are allocated from contiguous sequences of 71 * blocks. When half the available blocks are filled, a garbage 72 * collection occurs. We "flip" spaces (exchange from- and to-space), 73 * copy live objects into to space, and perform pointer adjustment. 74 * 75 * Copying is made more complicated by the requirement that some 76 * objects must not be moved. This property is known as "pinning". 77 * These objects must be dealt with specially. We use Bartlett's 78 * scheme; blocks containing such objects are grayed (promoted) at the 79 * start of a garbage collection. By virtue of this trick, tracing 80 * from the roots proceeds as usual but all objects on those pages are 81 * considered promoted and therefore not moved. 82 * 83 * TODO: there is sufficient information within the garbage collector 84 * to implement Attardi's scheme for evacuating unpinned objects from 85 * a page that is otherwise pinned. This would eliminate false 86 * retention caused by the large pinning granularity. 87 * 88 * We need a scheme for medium and large objects. Ignore that for 89 * now, we can return to this later. 90 * 91 * Eventually we need to worry about promoting objects out of the 92 * copy-collected heap (tenuring) into a less volatile space. Copying 93 * may not always be the best policy for such spaces. We should 94 * consider a variant of mark, sweep, compact. 95 * 96 * The block scheme allows us to use VM page faults to maintain a 97 * write barrier. Consider having a special leaf state for a page. 98 * 99 * Bibliography: 100 * 101 * C. J. Cheney. 1970. A non-recursive list compacting 102 * algorithm. CACM. 13-11 pp677--678. 103 * 104 * Joel F. Bartlett. 1988. Compacting Garbage Collection with 105 * Ambiguous Roots. Digital Equipment Corporation. 106 * 107 * Joel F. Bartlett. 1989. Mostly-Copying Garbage Collection Picks Up 108 * Generations and C++. Digital Equipment Corporation. 109 * 110 * G. May Yip. 1991. Incremental, Generational Mostly-Copying Garbage 111 * Collection in Uncooperative Environments. Digital Equipment 112 * Corporation. 113 * 114 * Giuseppe Attardi, Tito Flagella. 1994. A Customisable Memory 115 * Management Framework. TR-94-010 116 * 117 * Giuseppe Attardi, Tito Flagella, Pietro Iglio. 1998. A customisable 118 * memory management framework for C++. Software -- Practice and 119 * Experience. 28(11), 1143-1183. 120 * 121 */ 122 123#define ARRAYSIZE(x) (sizeof(x) / sizeof(x[0])) 124 125#if 0 126#define LOG_ALLOC LOGI 127#define LOG_PIN LOGI 128#define LOG_PROM LOGI 129#define LOG_REF LOGI 130#define LOG_SCAV LOGI 131#define LOG_TRAN LOGI 132#define LOG_VER LOGI 133#else 134#define LOG_ALLOC(...) ((void)0) 135#define LOG_PIN(...) ((void)0) 136#define LOG_PROM(...) ((void)0) 137#define LOG_REF(...) ((void)0) 138#define LOG_SCAV(...) ((void)0) 139#define LOG_TRAN(...) ((void)0) 140#define LOG_VER(...) ((void)0) 141#endif 142 143static void enqueueBlock(HeapSource *heapSource, size_t block); 144static void scavengeReference(Object **obj); 145static bool toSpaceContains(const void *addr); 146static bool fromSpaceContains(const void *addr); 147static size_t sumHeapBitmap(const HeapBitmap *bitmap); 148static size_t objectSize(const Object *obj); 149static void scavengeDataObject(Object *obj); 150static void scavengeBlockQueue(); 151 152/* 153 * We use 512-byte blocks. 154 */ 155enum { BLOCK_SHIFT = 9 }; 156enum { BLOCK_SIZE = 1 << BLOCK_SHIFT }; 157 158/* 159 * Space identifiers, stored into the blockSpace array. 160 */ 161enum { 162 BLOCK_FREE = 0, 163 BLOCK_FROM_SPACE = 1, 164 BLOCK_TO_SPACE = 2, 165 BLOCK_CONTINUED = 7 166}; 167 168/* 169 * Alignment for all allocations, in bytes. 170 */ 171enum { ALLOC_ALIGNMENT = 8 }; 172 173/* 174 * Sentinel value for the queue end. 175 */ 176#define QUEUE_TAIL (~(size_t)0) 177 178struct HeapSource { 179 180 /* The base address of backing store. */ 181 u1 *blockBase; 182 183 /* Total number of blocks available for allocation. */ 184 size_t totalBlocks; 185 size_t allocBlocks; 186 187 /* 188 * The scavenger work queue. Implemented as an array of index 189 * values into the queue. 190 */ 191 size_t *blockQueue; 192 193 /* 194 * Base and limit blocks. Basically the shifted start address of 195 * the block. We convert blocks to a relative number when 196 * indexing in the block queue. TODO: make the block queue base 197 * relative rather than the index into the block queue. 198 */ 199 size_t baseBlock, limitBlock; 200 201 size_t queueHead; 202 size_t queueTail; 203 size_t queueSize; 204 205 /* The space of the current block 0 (free), 1 or 2. */ 206 char *blockSpace; 207 208 /* Start of free space in the current block. */ 209 u1 *allocPtr; 210 /* Exclusive limit of free space in the current block. */ 211 u1 *allocLimit; 212 213 HeapBitmap allocBits; 214 215 /* 216 * The starting size of the heap. This value is the same as the 217 * value provided to the -Xms flag. 218 */ 219 size_t minimumSize; 220 221 /* 222 * The maximum size of the heap. This value is the same as the 223 * -Xmx flag. 224 */ 225 size_t maximumSize; 226 227 /* 228 * The current, committed size of the heap. At present, this is 229 * equivalent to the maximumSize. 230 */ 231 size_t currentSize; 232 233 size_t bytesAllocated; 234}; 235 236static unsigned long alignDown(unsigned long x, unsigned long n) 237{ 238 return x & -n; 239} 240 241static unsigned long alignUp(unsigned long x, unsigned long n) 242{ 243 return alignDown(x + (n - 1), n); 244} 245 246static void describeBlocks(const HeapSource *heapSource) 247{ 248 for (size_t i = 0; i < heapSource->totalBlocks; ++i) { 249 if ((i % 32) == 0) putchar('\n'); 250 printf("%d ", heapSource->blockSpace[i]); 251 } 252 putchar('\n'); 253} 254 255/* 256 * Virtual memory interface. 257 */ 258 259static void *virtualAlloc(size_t length) 260{ 261 int flags = MAP_PRIVATE | MAP_ANONYMOUS; 262 int prot = PROT_READ | PROT_WRITE; 263 void *addr = mmap(NULL, length, prot, flags, -1, 0); 264 if (addr == MAP_FAILED) { 265 LOGE_HEAP("mmap: %s", strerror(errno)); 266 addr = NULL; 267 } 268 return addr; 269} 270 271static void virtualFree(void *addr, size_t length) 272{ 273 assert(addr != NULL); 274 assert((uintptr_t)addr % SYSTEM_PAGE_SIZE == 0); 275 int res = munmap(addr, length); 276 if (res == -1) { 277 LOGE_HEAP("munmap: %s", strerror(errno)); 278 } 279} 280 281#ifndef NDEBUG 282static int isValidAddress(const HeapSource *heapSource, const u1 *addr) 283{ 284 size_t block; 285 286 block = (uintptr_t)addr >> BLOCK_SHIFT; 287 return heapSource->baseBlock <= block && 288 heapSource->limitBlock > block; 289} 290#endif 291 292/* 293 * Iterate over the block map looking for a contiguous run of free 294 * blocks. 295 */ 296static void *allocateBlocks(HeapSource *heapSource, size_t blocks) 297{ 298 size_t allocBlocks = heapSource->allocBlocks; 299 size_t totalBlocks = heapSource->totalBlocks; 300 /* Check underflow. */ 301 assert(blocks != 0); 302 /* Check overflow. */ 303 if (allocBlocks + blocks > totalBlocks / 2) { 304 return NULL; 305 } 306 /* Scan block map. */ 307 for (size_t i = 0; i < totalBlocks; ++i) { 308 /* Check fit. */ 309 for (size_t j = 0; j < blocks; ++j) { /* runs over totalBlocks */ 310 if (heapSource->blockSpace[i+j] != BLOCK_FREE) { 311 break; 312 } 313 } 314 /* No fit? */ 315 if (j != blocks) { 316 i += j; 317 continue; 318 } 319 /* Fit, allocate. */ 320 heapSource->blockSpace[i] = BLOCK_TO_SPACE; /* why to-space? */ 321 for (size_t j = 1; j < blocks; ++j) { 322 heapSource->blockSpace[i+j] = BLOCK_CONTINUED; 323 } 324 heapSource->allocBlocks += blocks; 325 void *addr = &heapSource->blockBase[i*BLOCK_SIZE]; 326 memset(addr, 0, blocks*BLOCK_SIZE); 327 /* Collecting? */ 328 if (heapSource->queueHead != QUEUE_TAIL) { 329 LOG_ALLOC("allocateBlocks allocBlocks=%zu,block#=%zu", heapSource->allocBlocks, i); 330 /* 331 * This allocated was on behalf of the transporter when it 332 * shaded a white object gray. We enqueue the block so 333 * the scavenger can further shade the gray objects black. 334 */ 335 enqueueBlock(heapSource, i); 336 } 337 338 return addr; 339 } 340 /* Insufficient space, fail. */ 341 LOGE("Insufficient space, %zu blocks, %zu blocks allocated and %zu bytes allocated", 342 heapSource->totalBlocks, 343 heapSource->allocBlocks, 344 heapSource->bytesAllocated); 345 return NULL; 346} 347 348/* Converts an absolute address to a relative block number. */ 349static size_t addressToBlock(const HeapSource *heapSource, const void *addr) 350{ 351 assert(heapSource != NULL); 352 assert(isValidAddress(heapSource, addr)); 353 return (((uintptr_t)addr) >> BLOCK_SHIFT) - heapSource->baseBlock; 354} 355 356/* Converts a relative block number to an absolute address. */ 357static u1 *blockToAddress(const HeapSource *heapSource, size_t block) 358{ 359 u1 *addr; 360 361 addr = (u1 *) (((uintptr_t) heapSource->baseBlock + block) * BLOCK_SIZE); 362 assert(isValidAddress(heapSource, addr)); 363 return addr; 364} 365 366static void clearBlock(HeapSource *heapSource, size_t block) 367{ 368 assert(heapSource != NULL); 369 assert(block < heapSource->totalBlocks); 370 u1 *addr = heapSource->blockBase + block*BLOCK_SIZE; 371 memset(addr, 0xCC, BLOCK_SIZE); 372 for (size_t i = 0; i < BLOCK_SIZE; i += 8) { 373 dvmHeapBitmapClearObjectBit(&heapSource->allocBits, addr + i); 374 } 375} 376 377static void clearFromSpace(HeapSource *heapSource) 378{ 379 assert(heapSource != NULL); 380 size_t i = 0; 381 size_t count = 0; 382 while (i < heapSource->totalBlocks) { 383 if (heapSource->blockSpace[i] != BLOCK_FROM_SPACE) { 384 ++i; 385 continue; 386 } 387 heapSource->blockSpace[i] = BLOCK_FREE; 388 clearBlock(heapSource, i); 389 ++i; 390 ++count; 391 while (i < heapSource->totalBlocks && 392 heapSource->blockSpace[i] == BLOCK_CONTINUED) { 393 heapSource->blockSpace[i] = BLOCK_FREE; 394 clearBlock(heapSource, i); 395 ++i; 396 ++count; 397 } 398 } 399 LOG_SCAV("freed %zu blocks (%zu bytes)", count, count*BLOCK_SIZE); 400} 401 402/* 403 * Appends the given block to the block queue. The block queue is 404 * processed in-order by the scavenger. 405 */ 406static void enqueueBlock(HeapSource *heapSource, size_t block) 407{ 408 assert(heapSource != NULL); 409 assert(block < heapSource->totalBlocks); 410 if (heapSource->queueHead != QUEUE_TAIL) { 411 heapSource->blockQueue[heapSource->queueTail] = block; 412 } else { 413 heapSource->queueHead = block; 414 } 415 heapSource->blockQueue[block] = QUEUE_TAIL; 416 heapSource->queueTail = block; 417 ++heapSource->queueSize; 418} 419 420/* 421 * Grays all objects within the block corresponding to the given 422 * address. 423 */ 424static void promoteBlockByAddr(HeapSource *heapSource, const void *addr) 425{ 426 size_t block; 427 428 block = addressToBlock(heapSource, (const u1 *)addr); 429 if (heapSource->blockSpace[block] != BLOCK_TO_SPACE) { 430 // LOG_PROM("promoting block %zu %d @ %p", block, heapSource->blockSpace[block], obj); 431 heapSource->blockSpace[block] = BLOCK_TO_SPACE; 432 enqueueBlock(heapSource, block); 433 /* TODO(cshapiro): count continued blocks?*/ 434 heapSource->allocBlocks += 1; 435 } else { 436 // LOG_PROM("NOT promoting block %zu %d @ %p", block, heapSource->blockSpace[block], obj); 437 } 438} 439 440GcHeap *dvmHeapSourceStartup(size_t startSize, size_t absoluteMaxSize) 441{ 442 GcHeap* gcHeap; 443 HeapSource *heapSource; 444 445 assert(startSize <= absoluteMaxSize); 446 447 heapSource = malloc(sizeof(*heapSource)); 448 assert(heapSource != NULL); 449 memset(heapSource, 0, sizeof(*heapSource)); 450 451 heapSource->minimumSize = alignUp(startSize, BLOCK_SIZE); 452 heapSource->maximumSize = alignUp(absoluteMaxSize, BLOCK_SIZE); 453 454 heapSource->currentSize = heapSource->maximumSize; 455 456 /* Allocate underlying storage for blocks. */ 457 heapSource->blockBase = virtualAlloc(heapSource->maximumSize); 458 assert(heapSource->blockBase != NULL); 459 heapSource->baseBlock = (uintptr_t) heapSource->blockBase >> BLOCK_SHIFT; 460 heapSource->limitBlock = ((uintptr_t) heapSource->blockBase + heapSource->maximumSize) >> BLOCK_SHIFT; 461 462 heapSource->allocBlocks = 0; 463 heapSource->totalBlocks = (heapSource->limitBlock - heapSource->baseBlock); 464 465 assert(heapSource->totalBlocks = heapSource->maximumSize / BLOCK_SIZE); 466 467 { 468 size_t size = sizeof(heapSource->blockQueue[0]); 469 heapSource->blockQueue = malloc(heapSource->totalBlocks*size); 470 assert(heapSource->blockQueue != NULL); 471 memset(heapSource->blockQueue, 0xCC, heapSource->totalBlocks*size); 472 heapSource->queueHead = QUEUE_TAIL; 473 } 474 475 /* Byte indicating space residence or free status of block. */ 476 { 477 size_t size = sizeof(heapSource->blockSpace[0]); 478 heapSource->blockSpace = malloc(heapSource->totalBlocks*size); 479 assert(heapSource->blockSpace != NULL); 480 memset(heapSource->blockSpace, 0, heapSource->totalBlocks*size); 481 } 482 483 dvmHeapBitmapInit(&heapSource->allocBits, 484 heapSource->blockBase, 485 heapSource->maximumSize, 486 "blockBase"); 487 488 /* Initialize allocation pointers. */ 489 heapSource->allocPtr = allocateBlocks(heapSource, 1); 490 heapSource->allocLimit = heapSource->allocPtr + BLOCK_SIZE; 491 492 gcHeap = malloc(sizeof(*gcHeap)); 493 assert(gcHeap != NULL); 494 memset(gcHeap, 0, sizeof(*gcHeap)); 495 gcHeap->heapSource = heapSource; 496 497 return gcHeap; 498} 499 500/* 501 * Perform any required heap initializations after forking from the 502 * zygote process. This is a no-op for the time being. Eventually 503 * this will demarcate the shared region of the heap. 504 */ 505bool dvmHeapSourceStartupAfterZygote() 506{ 507 return true; 508} 509 510bool dvmHeapSourceStartupBeforeFork() 511{ 512 assert(!"implemented"); 513 return false; 514} 515 516void dvmHeapSourceShutdown(GcHeap **gcHeap) 517{ 518 if (*gcHeap == NULL || (*gcHeap)->heapSource == NULL) 519 return; 520 free((*gcHeap)->heapSource->blockQueue); 521 free((*gcHeap)->heapSource->blockSpace); 522 virtualFree((*gcHeap)->heapSource->blockBase, 523 (*gcHeap)->heapSource->maximumSize); 524 free((*gcHeap)->heapSource); 525 (*gcHeap)->heapSource = NULL; 526 free(*gcHeap); 527 *gcHeap = NULL; 528} 529 530size_t dvmHeapSourceGetValue(HeapSourceValueSpec spec, 531 size_t perHeapStats[], 532 size_t arrayLen) 533{ 534 HeapSource *heapSource; 535 size_t value; 536 537 heapSource = gDvm.gcHeap->heapSource; 538 switch (spec) { 539 case HS_FOOTPRINT: 540 value = heapSource->maximumSize; 541 break; 542 case HS_ALLOWED_FOOTPRINT: 543 value = heapSource->maximumSize; 544 break; 545 case HS_BYTES_ALLOCATED: 546 value = heapSource->bytesAllocated; 547 break; 548 case HS_OBJECTS_ALLOCATED: 549 value = sumHeapBitmap(&heapSource->allocBits); 550 break; 551 default: 552 assert(!"implemented"); 553 value = 0; 554 } 555 if (perHeapStats) { 556 *perHeapStats = value; 557 } 558 return value; 559} 560 561/* 562 * Performs a shallow copy of the allocation bitmap into the given 563 * vector of heap bitmaps. 564 */ 565void dvmHeapSourceGetObjectBitmaps(HeapBitmap objBits[], HeapBitmap markBits[], 566 size_t numHeaps) 567{ 568 assert(!"implemented"); 569} 570 571HeapBitmap *dvmHeapSourceGetLiveBits() 572{ 573 return &gDvm.gcHeap->heapSource->allocBits; 574} 575 576/* 577 * Allocate the specified number of bytes from the heap. The 578 * allocation cursor points into a block of free storage. If the 579 * given allocation fits in the remaining space of the block, we 580 * advance the cursor and return a pointer to the free storage. If 581 * the allocation cannot fit in the current block but is smaller than 582 * a block we request a new block and allocate from it instead. If 583 * the allocation is larger than a block we must allocate from a span 584 * of contiguous blocks. 585 */ 586void *dvmHeapSourceAlloc(size_t length) 587{ 588 HeapSource *heapSource; 589 unsigned char *addr; 590 size_t aligned, available, blocks; 591 592 heapSource = gDvm.gcHeap->heapSource; 593 assert(heapSource != NULL); 594 assert(heapSource->allocPtr != NULL); 595 assert(heapSource->allocLimit != NULL); 596 597 aligned = alignUp(length, ALLOC_ALIGNMENT); 598 available = heapSource->allocLimit - heapSource->allocPtr; 599 600 /* Try allocating inside the current block. */ 601 if (aligned <= available) { 602 addr = heapSource->allocPtr; 603 heapSource->allocPtr += aligned; 604 heapSource->bytesAllocated += aligned; 605 dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr); 606 return addr; 607 } 608 609 /* Try allocating in a new block. */ 610 if (aligned <= BLOCK_SIZE) { 611 addr = allocateBlocks(heapSource, 1); 612 if (addr != NULL) { 613 heapSource->allocLimit = addr + BLOCK_SIZE; 614 heapSource->allocPtr = addr + aligned; 615 heapSource->bytesAllocated += aligned; 616 dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr); 617 /* TODO(cshapiro): pad out the current block. */ 618 } 619 return addr; 620 } 621 622 /* Try allocating in a span of blocks. */ 623 blocks = alignUp(aligned, BLOCK_SIZE) / BLOCK_SIZE; 624 625 addr = allocateBlocks(heapSource, blocks); 626 /* Propagate failure upward. */ 627 if (addr != NULL) { 628 heapSource->bytesAllocated += aligned; 629 dvmHeapBitmapSetObjectBit(&heapSource->allocBits, addr); 630 /* TODO(cshapiro): pad out free space in the last block. */ 631 } 632 return addr; 633} 634 635void *dvmHeapSourceAllocAndGrow(size_t size) 636{ 637 return dvmHeapSourceAlloc(size); 638} 639 640/* TODO: refactor along with dvmHeapSourceAlloc */ 641void *allocateGray(size_t size) 642{ 643 HeapSource *heapSource; 644 void *addr; 645 size_t block; 646 647 /* TODO: add a check that we are in a GC. */ 648 heapSource = gDvm.gcHeap->heapSource; 649 addr = dvmHeapSourceAlloc(size); 650 assert(addr != NULL); 651 block = addressToBlock(heapSource, (const u1 *)addr); 652 if (heapSource->queueHead == QUEUE_TAIL) { 653 /* 654 * Forcibly append the underlying block to the queue. This 655 * condition occurs when referents are transported following 656 * the initial trace. 657 */ 658 enqueueBlock(heapSource, block); 659 LOG_PROM("forced promoting block %zu %d @ %p", block, heapSource->blockSpace[block], addr); 660 } 661 return addr; 662} 663 664bool dvmHeapSourceContainsAddress(const void *ptr) 665{ 666 HeapSource *heapSource = gDvm.gcHeap->heapSource; 667 return dvmHeapBitmapCoversAddress(&heapSource->allocBits, ptr); 668} 669 670/* 671 * Returns true if the given address is within the heap and points to 672 * the header of a live object. 673 */ 674bool dvmHeapSourceContains(const void *addr) 675{ 676 HeapSource *heapSource; 677 HeapBitmap *bitmap; 678 679 heapSource = gDvm.gcHeap->heapSource; 680 bitmap = &heapSource->allocBits; 681 if (!dvmHeapBitmapCoversAddress(bitmap, addr)) { 682 return false; 683 } else { 684 return dvmHeapBitmapIsObjectBitSet(bitmap, addr); 685 } 686} 687 688bool dvmHeapSourceGetPtrFlag(const void *ptr, HeapSourcePtrFlag flag) 689{ 690 assert(!"implemented"); 691 return false; 692} 693 694size_t dvmHeapSourceChunkSize(const void *ptr) 695{ 696 assert(!"implemented"); 697 return 0; 698} 699 700size_t dvmHeapSourceFootprint() 701{ 702 assert(!"implemented"); 703 return 0; 704} 705 706/* 707 * Returns the "ideal footprint" which appears to be the number of 708 * bytes currently committed to the heap. This starts out at the 709 * start size of the heap and grows toward the maximum size. 710 */ 711size_t dvmHeapSourceGetIdealFootprint() 712{ 713 return gDvm.gcHeap->heapSource->currentSize; 714} 715 716float dvmGetTargetHeapUtilization() 717{ 718 return 0.5f; 719} 720 721void dvmSetTargetHeapUtilization(float newTarget) 722{ 723 assert(newTarget > 0.0f && newTarget < 1.0f); 724} 725 726/* 727 * Expands the size of the heap after a collection. At present we 728 * commit the pages for maximum size of the heap so this routine is 729 * just a no-op. Eventually, we will either allocate or commit pages 730 * on an as-need basis. 731 */ 732void dvmHeapSourceGrowForUtilization() 733{ 734 /* do nothing */ 735} 736 737void dvmHeapSourceWalk(void (*callback)(const void *chunkptr, size_t chunklen, 738 const void *userptr, size_t userlen, 739 void *arg), 740 void *arg) 741{ 742 assert(!"implemented"); 743} 744 745size_t dvmHeapSourceGetNumHeaps() 746{ 747 return 1; 748} 749 750bool dvmTrackExternalAllocation(size_t n) 751{ 752 /* do nothing */ 753 return true; 754} 755 756void dvmTrackExternalFree(size_t n) 757{ 758 /* do nothing */ 759} 760 761size_t dvmGetExternalBytesAllocated() 762{ 763 assert(!"implemented"); 764 return 0; 765} 766 767void dvmHeapSourceFlip() 768{ 769 HeapSource *heapSource = gDvm.gcHeap->heapSource; 770 771 /* Reset the block queue. */ 772 heapSource->allocBlocks = 0; 773 heapSource->queueSize = 0; 774 heapSource->queueHead = QUEUE_TAIL; 775 776 /* TODO(cshapiro): pad the current (prev) block. */ 777 778 heapSource->allocPtr = NULL; 779 heapSource->allocLimit = NULL; 780 781 /* Whiten all allocated blocks. */ 782 for (size_t i = 0; i < heapSource->totalBlocks; ++i) { 783 if (heapSource->blockSpace[i] == BLOCK_TO_SPACE) { 784 heapSource->blockSpace[i] = BLOCK_FROM_SPACE; 785 } 786 } 787} 788 789static void room(size_t *alloc, size_t *avail, size_t *total) 790{ 791 HeapSource *heapSource = gDvm.gcHeap->heapSource; 792 *total = heapSource->totalBlocks*BLOCK_SIZE; 793 *alloc = heapSource->allocBlocks*BLOCK_SIZE; 794 *avail = *total - *alloc; 795} 796 797static bool isSpaceInternal(u1 *addr, int space) 798{ 799 HeapSource *heapSource; 800 u1 *base, *limit; 801 size_t offset; 802 char space2; 803 804 heapSource = gDvm.gcHeap->heapSource; 805 base = heapSource->blockBase; 806 assert(addr >= base); 807 limit = heapSource->blockBase + heapSource->maximumSize; 808 assert(addr < limit); 809 offset = addr - base; 810 space2 = heapSource->blockSpace[offset >> BLOCK_SHIFT]; 811 return space == space2; 812} 813 814static bool fromSpaceContains(const void *addr) 815{ 816 return isSpaceInternal((u1 *)addr, BLOCK_FROM_SPACE); 817} 818 819static bool toSpaceContains(const void *addr) 820{ 821 return isSpaceInternal((u1 *)addr, BLOCK_TO_SPACE); 822} 823 824/* 825 * Notifies the collector that the object at the given address must 826 * remain stationary during the current collection. 827 */ 828static void pinObject(const Object *obj) 829{ 830 promoteBlockByAddr(gDvm.gcHeap->heapSource, obj); 831} 832 833static size_t sumHeapBitmap(const HeapBitmap *bitmap) 834{ 835 size_t sum = 0; 836 for (size_t i = 0; i < bitmap->bitsLen >> 2; ++i) { 837 sum += CLZ(bitmap->bits[i]); 838 } 839 return sum; 840} 841 842/* 843 * Miscellaneous functionality. 844 */ 845 846static int isForward(const void *addr) 847{ 848 return (uintptr_t)addr & 0x1; 849} 850 851static void setForward(const void *toObj, void *fromObj) 852{ 853 *(unsigned long *)fromObj = (uintptr_t)toObj | 0x1; 854} 855 856static void* getForward(const void *fromObj) 857{ 858 return (void *)((uintptr_t)fromObj & ~0x1); 859} 860 861/* Beware, uses the same encoding as a forwarding pointers! */ 862static int isPermanentString(const StringObject *obj) { 863 return (uintptr_t)obj & 0x1; 864} 865 866static void* getPermanentString(const StringObject *obj) 867{ 868 return (void *)((uintptr_t)obj & ~0x1); 869} 870 871 872/* 873 * Scavenging and transporting routines follow. A transporter grays 874 * an object. A scavenger blackens an object. We define these 875 * routines for each fundamental object type. Dispatch is performed 876 * in scavengeObject. 877 */ 878 879/* 880 * Class object scavenging. 881 */ 882static void scavengeClassObject(ClassObject *obj) 883{ 884 LOG_SCAV("scavengeClassObject(obj=%p)", obj); 885 assert(obj != NULL); 886 assert(obj->obj.clazz != NULL); 887 assert(obj->obj.clazz->descriptor != NULL); 888 assert(!strcmp(obj->obj.clazz->descriptor, "Ljava/lang/Class;")); 889 assert(obj->descriptor != NULL); 890 LOG_SCAV("scavengeClassObject: descriptor='%s',vtableCount=%zu", 891 obj->descriptor, obj->vtableCount); 892 /* Delegate class object and instance field scavenging. */ 893 scavengeDataObject((Object *)obj); 894 /* Scavenge the array element class object. */ 895 if (IS_CLASS_FLAG_SET(obj, CLASS_ISARRAY)) { 896 scavengeReference((Object **)(void *)&obj->elementClass); 897 } 898 /* Scavenge the superclass. */ 899 scavengeReference((Object **)(void *)&obj->super); 900 /* Scavenge the class loader. */ 901 scavengeReference(&obj->classLoader); 902 /* Scavenge static fields. */ 903 for (int i = 0; i < obj->sfieldCount; ++i) { 904 char ch = obj->sfields[i].field.signature[0]; 905 if (ch == '[' || ch == 'L') { 906 scavengeReference((Object **)(void *)&obj->sfields[i].value.l); 907 } 908 } 909 /* Scavenge interface class objects. */ 910 for (int i = 0; i < obj->interfaceCount; ++i) { 911 scavengeReference((Object **) &obj->interfaces[i]); 912 } 913} 914 915/* 916 * Array object scavenging. 917 */ 918static size_t scavengeArrayObject(ArrayObject *array) 919{ 920 LOG_SCAV("scavengeArrayObject(array=%p)", array); 921 /* Scavenge the class object. */ 922 assert(toSpaceContains(array)); 923 assert(array != NULL); 924 assert(array->obj.clazz != NULL); 925 scavengeReference((Object **) array); 926 size_t length = dvmArrayObjectSize(array); 927 /* Scavenge the array contents. */ 928 if (IS_CLASS_FLAG_SET(array->obj.clazz, CLASS_ISOBJECTARRAY)) { 929 Object **contents = (Object **)array->contents; 930 for (size_t i = 0; i < array->length; ++i) { 931 scavengeReference(&contents[i]); 932 } 933 } 934 return length; 935} 936 937/* 938 * Reference object scavenging. 939 */ 940 941static int getReferenceFlags(const Object *obj) 942{ 943 int flags; 944 945 flags = CLASS_ISREFERENCE | 946 CLASS_ISWEAKREFERENCE | 947 CLASS_ISPHANTOMREFERENCE; 948 return GET_CLASS_FLAG_GROUP(obj->clazz, flags); 949} 950 951static int isSoftReference(const Object *obj) 952{ 953 return getReferenceFlags(obj) == CLASS_ISREFERENCE; 954} 955 956static int isWeakReference(const Object *obj) 957{ 958 return getReferenceFlags(obj) & CLASS_ISWEAKREFERENCE; 959} 960 961#ifndef NDEBUG 962static bool isPhantomReference(const Object *obj) 963{ 964 return getReferenceFlags(obj) & CLASS_ISPHANTOMREFERENCE; 965} 966#endif 967 968/* 969 * Returns true if the reference was registered with a reference queue 970 * but has not yet been appended to it. 971 */ 972static bool isReferenceEnqueuable(const Object *ref) 973{ 974 Object *queue, *queueNext; 975 976 queue = dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queue); 977 queueNext = dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queueNext); 978 if (queue == NULL || queueNext != NULL) { 979 /* 980 * There is no queue, or the reference has already 981 * been enqueued. The Reference.enqueue() method 982 * will do nothing even if we call it. 983 */ 984 return false; 985 } 986 987 /* 988 * We need to call enqueue(), but if we called it from 989 * here we'd probably deadlock. Schedule a call. 990 */ 991 return true; 992} 993 994/* 995 * Schedules a reference to be appended to its reference queue. 996 */ 997static void enqueueReference(Object *ref) 998{ 999 assert(ref != NULL); 1000 assert(dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queue) != NULL); 1001 assert(dvmGetFieldObject(ref, gDvm.offJavaLangRefReference_queueNext) == NULL); 1002 if (!dvmHeapAddRefToLargeTable(&gDvm.gcHeap->referenceOperations, ref)) { 1003 LOGE("no room for any more reference operations"); 1004 dvmAbort(); 1005 } 1006} 1007 1008/* 1009 * Sets the referent field of a reference object to NULL. 1010 */ 1011static void clearReference(Object *obj) 1012{ 1013 dvmSetFieldObject(obj, gDvm.offJavaLangRefReference_referent, NULL); 1014} 1015 1016/* 1017 * Clears reference objects with white referents. 1018 */ 1019void clearWhiteReferences(Object **list) 1020{ 1021 size_t referentOffset, queueNextOffset; 1022 bool doSignal; 1023 1024 queueNextOffset = gDvm.offJavaLangRefReference_queueNext; 1025 referentOffset = gDvm.offJavaLangRefReference_referent; 1026 doSignal = false; 1027 while (*list != NULL) { 1028 Object *ref = *list; 1029 JValue *field = dvmFieldPtr(ref, referentOffset); 1030 Object *referent = field->l; 1031 *list = dvmGetFieldObject(ref, queueNextOffset); 1032 dvmSetFieldObject(ref, queueNextOffset, NULL); 1033 assert(referent != NULL); 1034 if (isForward(referent->clazz)) { 1035 field->l = referent = getForward(referent->clazz); 1036 continue; 1037 } 1038 if (fromSpaceContains(referent)) { 1039 /* Referent is white, clear it. */ 1040 clearReference(ref); 1041 if (isReferenceEnqueuable(ref)) { 1042 enqueueReference(ref); 1043 doSignal = true; 1044 } 1045 } 1046 } 1047 /* 1048 * If we cleared a reference with a reference queue we must notify 1049 * the heap worker to append the reference. 1050 */ 1051 if (doSignal) { 1052 dvmSignalHeapWorker(false); 1053 } 1054 assert(*list == NULL); 1055} 1056 1057/* 1058 * Blackens referents subject to the soft reference preservation 1059 * policy. 1060 */ 1061void preserveSoftReferences(Object **list) 1062{ 1063 Object *ref; 1064 Object *prev, *next; 1065 size_t referentOffset, queueNextOffset; 1066 unsigned counter; 1067 bool white; 1068 1069 queueNextOffset = gDvm.offJavaLangRefReference_queueNext; 1070 referentOffset = gDvm.offJavaLangRefReference_referent; 1071 counter = 0; 1072 prev = next = NULL; 1073 ref = *list; 1074 while (ref != NULL) { 1075 JValue *field = dvmFieldPtr(ref, referentOffset); 1076 Object *referent = field->l; 1077 next = dvmGetFieldObject(ref, queueNextOffset); 1078 assert(referent != NULL); 1079 if (isForward(referent->clazz)) { 1080 /* Referent is black. */ 1081 field->l = referent = getForward(referent->clazz); 1082 white = false; 1083 } else { 1084 white = fromSpaceContains(referent); 1085 } 1086 if (!white && ((++counter) & 1)) { 1087 /* Referent is white and biased toward saving, gray it. */ 1088 scavengeReference((Object **)(void *)&field->l); 1089 white = true; 1090 } 1091 if (white) { 1092 /* Referent is black, unlink it. */ 1093 if (prev != NULL) { 1094 dvmSetFieldObject(ref, queueNextOffset, NULL); 1095 dvmSetFieldObject(prev, queueNextOffset, next); 1096 } 1097 } else { 1098 /* Referent is white, skip over it. */ 1099 prev = ref; 1100 } 1101 ref = next; 1102 } 1103 /* 1104 * Restart the trace with the newly gray references added to the 1105 * root set. 1106 */ 1107 scavengeBlockQueue(); 1108} 1109 1110void processFinalizableReferences() 1111{ 1112 HeapRefTable newPendingRefs; 1113 LargeHeapRefTable *finRefs = gDvm.gcHeap->finalizableRefs; 1114 Object **ref; 1115 Object **lastRef; 1116 size_t totalPendCount; 1117 1118 /* 1119 * All strongly, reachable objects are black. 1120 * Any white finalizable objects need to be finalized. 1121 */ 1122 1123 /* Create a table that the new pending refs will 1124 * be added to. 1125 */ 1126 if (!dvmHeapInitHeapRefTable(&newPendingRefs)) { 1127 //TODO: mark all finalizable refs and hope that 1128 // we can schedule them next time. Watch out, 1129 // because we may be expecting to free up space 1130 // by calling finalizers. 1131 LOG_REF("no room for pending finalizations"); 1132 dvmAbort(); 1133 } 1134 1135 /* 1136 * Walk through finalizableRefs and move any white references to 1137 * the list of new pending refs. 1138 */ 1139 totalPendCount = 0; 1140 while (finRefs != NULL) { 1141 Object **gapRef; 1142 size_t newPendCount = 0; 1143 1144 gapRef = ref = finRefs->refs.table; 1145 lastRef = finRefs->refs.nextEntry; 1146 while (ref < lastRef) { 1147 if (fromSpaceContains(*ref)) { 1148 if (!dvmHeapAddToHeapRefTable(&newPendingRefs, *ref)) { 1149 //TODO: add the current table and allocate 1150 // a new, smaller one. 1151 LOG_REF("no room for any more pending finalizations: %zd", 1152 dvmHeapNumHeapRefTableEntries(&newPendingRefs)); 1153 dvmAbort(); 1154 } 1155 newPendCount++; 1156 } else { 1157 /* This ref is black, so will remain on finalizableRefs. 1158 */ 1159 if (newPendCount > 0) { 1160 /* Copy it up to fill the holes. 1161 */ 1162 *gapRef++ = *ref; 1163 } else { 1164 /* No holes yet; don't bother copying. 1165 */ 1166 gapRef++; 1167 } 1168 } 1169 ref++; 1170 } 1171 finRefs->refs.nextEntry = gapRef; 1172 //TODO: if the table is empty when we're done, free it. 1173 totalPendCount += newPendCount; 1174 finRefs = finRefs->next; 1175 } 1176 LOG_REF("%zd finalizers triggered.", totalPendCount); 1177 if (totalPendCount == 0) { 1178 /* No objects required finalization. 1179 * Free the empty temporary table. 1180 */ 1181 dvmClearReferenceTable(&newPendingRefs); 1182 return; 1183 } 1184 1185 /* Add the new pending refs to the main list. 1186 */ 1187 if (!dvmHeapAddTableToLargeTable(&gDvm.gcHeap->pendingFinalizationRefs, 1188 &newPendingRefs)) 1189 { 1190 LOG_REF("can't insert new pending finalizations"); 1191 dvmAbort(); 1192 } 1193 1194 //TODO: try compacting the main list with a memcpy loop 1195 1196 /* Blacken the refs we just moved; we don't want them or their 1197 * children to get swept yet. 1198 */ 1199 ref = newPendingRefs.table; 1200 lastRef = newPendingRefs.nextEntry; 1201 assert(ref < lastRef); 1202 HPROF_SET_GC_SCAN_STATE(HPROF_ROOT_FINALIZING, 0); 1203 while (ref < lastRef) { 1204 scavengeReference(ref); 1205 ref++; 1206 } 1207 HPROF_CLEAR_GC_SCAN_STATE(); 1208 scavengeBlockQueue(); 1209 dvmSignalHeapWorker(false); 1210} 1211 1212/* 1213 * If a reference points to from-space and has been forwarded, we snap 1214 * the pointer to its new to-space address. If the reference points 1215 * to an unforwarded from-space address we must enqueue the reference 1216 * for later processing. TODO: implement proper reference processing 1217 * and move the referent scavenging elsewhere. 1218 */ 1219static void scavengeReferenceObject(Object *obj) 1220{ 1221 Object *referent; 1222 Object **queue; 1223 size_t referentOffset, queueNextOffset; 1224 1225 assert(obj != NULL); 1226 LOG_SCAV("scavengeReferenceObject(obj=%p),'%s'", obj, obj->clazz->descriptor); 1227 scavengeDataObject(obj); 1228 referentOffset = gDvm.offJavaLangRefReference_referent; 1229 referent = dvmGetFieldObject(obj, referentOffset); 1230 if (referent == NULL || toSpaceContains(referent)) { 1231 return; 1232 } 1233 if (isSoftReference(obj)) { 1234 queue = &gDvm.gcHeap->softReferences; 1235 } else if (isWeakReference(obj)) { 1236 queue = &gDvm.gcHeap->weakReferences; 1237 } else { 1238 assert(isPhantomReference(obj)); 1239 queue = &gDvm.gcHeap->phantomReferences; 1240 } 1241 queueNextOffset = gDvm.offJavaLangRefReference_queueNext; 1242 dvmSetFieldObject(obj, queueNextOffset, *queue); 1243 *queue = obj; 1244 LOG_SCAV("scavengeReferenceObject: enqueueing %p", obj); 1245} 1246 1247/* 1248 * Data object scavenging. 1249 */ 1250static void scavengeDataObject(Object *obj) 1251{ 1252 // LOG_SCAV("scavengeDataObject(obj=%p)", obj); 1253 assert(obj != NULL); 1254 assert(obj->clazz != NULL); 1255 assert(obj->clazz->objectSize != 0); 1256 assert(toSpaceContains(obj)); 1257 /* Scavenge the class object. */ 1258 ClassObject *clazz = obj->clazz; 1259 scavengeReference((Object **) obj); 1260 /* Scavenge instance fields. */ 1261 if (clazz->refOffsets != CLASS_WALK_SUPER) { 1262 size_t refOffsets = clazz->refOffsets; 1263 while (refOffsets != 0) { 1264 size_t rshift = CLZ(refOffsets); 1265 size_t offset = CLASS_OFFSET_FROM_CLZ(rshift); 1266 Object **ref = (Object **)((u1 *)obj + offset); 1267 scavengeReference(ref); 1268 refOffsets &= ~(CLASS_HIGH_BIT >> rshift); 1269 } 1270 } else { 1271 for (; clazz != NULL; clazz = clazz->super) { 1272 InstField *field = clazz->ifields; 1273 for (int i = 0; i < clazz->ifieldRefCount; ++i, ++field) { 1274 size_t offset = field->byteOffset; 1275 Object **ref = (Object **)((u1 *)obj + offset); 1276 scavengeReference(ref); 1277 } 1278 } 1279 } 1280} 1281 1282static Object *transportObject(const Object *fromObj) 1283{ 1284 Object *toObj; 1285 size_t allocSize, copySize; 1286 1287 LOG_TRAN("transportObject(fromObj=%p) allocBlocks=%zu", 1288 fromObj, 1289 gDvm.gcHeap->heapSource->allocBlocks); 1290 assert(fromObj != NULL); 1291 assert(fromSpaceContains(fromObj)); 1292 allocSize = copySize = objectSize(fromObj); 1293 if (LW_HASH_STATE(fromObj->lock) != LW_HASH_STATE_UNHASHED) { 1294 /* 1295 * The object has been hashed or hashed and moved. We must 1296 * reserve an additional word for a hash code. 1297 */ 1298 allocSize += sizeof(u4); 1299 } 1300 if (LW_HASH_STATE(fromObj->lock) == LW_HASH_STATE_HASHED_AND_MOVED) { 1301 /* 1302 * The object has its hash code allocated. Ensure the hash 1303 * code is copied along with the instance data. 1304 */ 1305 copySize += sizeof(u4); 1306 } 1307 /* TODO(cshapiro): don't copy, re-map large data objects. */ 1308 assert(copySize <= allocSize); 1309 toObj = allocateGray(allocSize); 1310 assert(toObj != NULL); 1311 assert(toSpaceContains(toObj)); 1312 memcpy(toObj, fromObj, copySize); 1313 if (LW_HASH_STATE(fromObj->lock) == LW_HASH_STATE_HASHED) { 1314 /* 1315 * The object has had its hash code exposed. Append it to the 1316 * instance and set a bit so we know to look for it there. 1317 */ 1318 *(u4 *)(((char *)toObj) + copySize) = (u4)fromObj >> 3; 1319 toObj->lock |= LW_HASH_STATE_HASHED_AND_MOVED << LW_HASH_STATE_SHIFT; 1320 } 1321 LOG_TRAN("transportObject: from %p/%zu to %p/%zu (%zu,%zu) %s", 1322 fromObj, addressToBlock(gDvm.gcHeap->heapSource,fromObj), 1323 toObj, addressToBlock(gDvm.gcHeap->heapSource,toObj), 1324 copySize, allocSize, copySize < allocSize ? "DIFFERENT" : ""); 1325 return toObj; 1326} 1327 1328/* 1329 * Generic reference scavenging. 1330 */ 1331 1332/* 1333 * Given a reference to an object, the scavenge routine will gray the 1334 * reference. Any objects pointed to by the scavenger object will be 1335 * transported to new space and a forwarding pointer will be installed 1336 * in the header of the object. 1337 */ 1338 1339/* 1340 * Blacken the given pointer. If the pointer is in from space, it is 1341 * transported to new space. If the object has a forwarding pointer 1342 * installed it has already been transported and the referent is 1343 * snapped to the new address. 1344 */ 1345static void scavengeReference(Object **obj) 1346{ 1347 ClassObject *clazz; 1348 Object *fromObj, *toObj; 1349 1350 assert(obj); 1351 1352 if (*obj == NULL) return; 1353 1354 assert(dvmIsValidObject(*obj)); 1355 1356 /* The entire block is black. */ 1357 if (toSpaceContains(*obj)) { 1358 LOG_SCAV("scavengeReference skipping pinned object @ %p", *obj); 1359 return; 1360 } 1361 LOG_SCAV("scavengeReference(*obj=%p)", *obj); 1362 1363 assert(fromSpaceContains(*obj)); 1364 1365 clazz = (*obj)->clazz; 1366 1367 if (isForward(clazz)) { 1368 // LOG_SCAV("forwarding %p @ %p to %p", *obj, obj, (void *)((uintptr_t)clazz & ~0x1)); 1369 *obj = (Object *)getForward(clazz); 1370 return; 1371 } 1372 fromObj = *obj; 1373 if (clazz == NULL) { 1374 // LOG_SCAV("scavangeReference %p has a NULL class object", fromObj); 1375 assert(!"implemented"); 1376 toObj = NULL; 1377 } else { 1378 toObj = transportObject(fromObj); 1379 } 1380 setForward(toObj, fromObj); 1381 *obj = (Object *)toObj; 1382} 1383 1384/* 1385 * Generic object scavenging. 1386 */ 1387static void scavengeObject(Object *obj) 1388{ 1389 ClassObject *clazz; 1390 1391 assert(obj != NULL); 1392 assert(obj->clazz != NULL); 1393 assert(!((uintptr_t)obj->clazz & 0x1)); 1394 clazz = obj->clazz; 1395 if (dvmIsTheClassClass(clazz)) { 1396 scavengeClassObject((ClassObject *)obj); 1397 } else if (IS_CLASS_FLAG_SET(clazz, CLASS_ISARRAY)) { 1398 scavengeArrayObject((ArrayObject *)obj); 1399 } else if (IS_CLASS_FLAG_SET(clazz, CLASS_ISREFERENCE)) { 1400 scavengeReferenceObject(obj); 1401 } else { 1402 scavengeDataObject(obj); 1403 } 1404} 1405 1406/* 1407 * External root scavenging routines. 1408 */ 1409 1410static void pinHashTableEntries(HashTable *table) 1411{ 1412 LOG_PIN(">>> pinHashTableEntries(table=%p)", table); 1413 if (table == NULL) { 1414 return; 1415 } 1416 dvmHashTableLock(table); 1417 for (int i = 0; i < table->tableSize; ++i) { 1418 HashEntry *entry = &table->pEntries[i]; 1419 void *obj = entry->data; 1420 if (obj == NULL || obj == HASH_TOMBSTONE) { 1421 continue; 1422 } 1423 pinObject(entry->data); 1424 } 1425 dvmHashTableUnlock(table); 1426 LOG_PIN("<<< pinHashTableEntries(table=%p)", table); 1427} 1428 1429static void pinPrimitiveClasses() 1430{ 1431 size_t length = ARRAYSIZE(gDvm.primitiveClass); 1432 for (size_t i = 0; i < length; i++) { 1433 if (gDvm.primitiveClass[i] != NULL) { 1434 pinObject((Object *)gDvm.primitiveClass[i]); 1435 } 1436 } 1437} 1438 1439/* 1440 * Scavenge interned strings. Permanent interned strings will have 1441 * been pinned and are therefore ignored. Non-permanent strings that 1442 * have been forwarded are snapped. All other entries are removed. 1443 */ 1444static void scavengeInternedStrings() 1445{ 1446 HashTable *table = gDvm.internedStrings; 1447 if (table == NULL) { 1448 return; 1449 } 1450 dvmHashTableLock(table); 1451 for (int i = 0; i < table->tableSize; ++i) { 1452 HashEntry *entry = &table->pEntries[i]; 1453 Object *obj = (Object *)entry->data; 1454 if (obj == NULL || obj == HASH_TOMBSTONE) { 1455 continue; 1456 } else if (!isPermanentString((StringObject *)obj)) { 1457 // LOG_SCAV("entry->data=%p", entry->data); 1458 LOG_SCAV(">>> string obj=%p", entry->data); 1459 /* TODO(cshapiro): detach white string objects */ 1460 scavengeReference((Object **)(void *)&entry->data); 1461 LOG_SCAV("<<< string obj=%p", entry->data); 1462 } 1463 } 1464 dvmHashTableUnlock(table); 1465} 1466 1467static void pinInternedStrings() 1468{ 1469 HashTable *table = gDvm.internedStrings; 1470 if (table == NULL) { 1471 return; 1472 } 1473 dvmHashTableLock(table); 1474 for (int i = 0; i < table->tableSize; ++i) { 1475 HashEntry *entry = &table->pEntries[i]; 1476 Object *obj = (Object *)entry->data; 1477 if (obj == NULL || obj == HASH_TOMBSTONE) { 1478 continue; 1479 } else if (isPermanentString((StringObject *)obj)) { 1480 obj = (Object *)getPermanentString((StringObject*)obj); 1481 LOG_PROM(">>> pin string obj=%p", obj); 1482 pinObject(obj); 1483 LOG_PROM("<<< pin string obj=%p", obj); 1484 } 1485 } 1486 dvmHashTableUnlock(table); 1487} 1488 1489/* 1490 * At present, reference tables contain references that must not be 1491 * moved by the collector. Instead of scavenging each reference in 1492 * the table we pin each referenced object. 1493 */ 1494static void pinReferenceTable(const ReferenceTable *table) 1495{ 1496 assert(table != NULL); 1497 assert(table->table != NULL); 1498 assert(table->nextEntry != NULL); 1499 for (Object **entry = table->table; entry < table->nextEntry; ++entry) { 1500 assert(entry != NULL); 1501 assert(!isForward(*entry)); 1502 pinObject(*entry); 1503 } 1504} 1505 1506static void scavengeLargeHeapRefTable(LargeHeapRefTable *table) 1507{ 1508 for (; table != NULL; table = table->next) { 1509 Object **ref = table->refs.table; 1510 for (; ref < table->refs.nextEntry; ++ref) { 1511 scavengeReference(ref); 1512 } 1513 } 1514} 1515 1516/* This code was copied from Thread.c */ 1517static void scavengeThreadStack(Thread *thread) 1518{ 1519 const u4 *framePtr; 1520#if WITH_EXTRA_GC_CHECKS > 1 1521 bool first = true; 1522#endif 1523 1524 framePtr = (const u4 *)thread->interpSave.curFrame; 1525 while (framePtr != NULL) { 1526 const StackSaveArea *saveArea; 1527 const Method *method; 1528 1529 saveArea = SAVEAREA_FROM_FP(framePtr); 1530 method = saveArea->method; 1531 if (method != NULL && !dvmIsNativeMethod(method)) { 1532#ifdef COUNT_PRECISE_METHODS 1533 /* the GC is running, so no lock required */ 1534 if (dvmPointerSetAddEntry(gDvm.preciseMethods, method)) 1535 LOG_SCAV("PGC: added %s.%s %p", 1536 method->clazz->descriptor, method->name, method); 1537#endif 1538#if WITH_EXTRA_GC_CHECKS > 1 1539 /* 1540 * May also want to enable the memset() in the "invokeMethod" 1541 * goto target in the portable interpreter. That sets the stack 1542 * to a pattern that makes referring to uninitialized data 1543 * very obvious. 1544 */ 1545 1546 if (first) { 1547 /* 1548 * First frame, isn't native, check the "alternate" saved PC 1549 * as a sanity check. 1550 * 1551 * It seems like we could check the second frame if the first 1552 * is native, since the PCs should be the same. It turns out 1553 * this doesn't always work. The problem is that we could 1554 * have calls in the sequence: 1555 * interp method #2 1556 * native method 1557 * interp method #1 1558 * 1559 * and then GC while in the native method after returning 1560 * from interp method #2. The currentPc on the stack is 1561 * for interp method #1, but thread->currentPc2 is still 1562 * set for the last thing interp method #2 did. 1563 * 1564 * This can also happen in normal execution: 1565 * - sget-object on not-yet-loaded class 1566 * - class init updates currentPc2 1567 * - static field init is handled by parsing annotations; 1568 * static String init requires creation of a String object, 1569 * which can cause a GC 1570 * 1571 * Essentially, any pattern that involves executing 1572 * interpreted code and then causes an allocation without 1573 * executing instructions in the original method will hit 1574 * this. These are rare enough that the test still has 1575 * some value. 1576 */ 1577 if (saveArea->xtra.currentPc != thread->currentPc2) { 1578 LOGW("PGC: savedPC(%p) != current PC(%p), %s.%s ins=%p", 1579 saveArea->xtra.currentPc, thread->currentPc2, 1580 method->clazz->descriptor, method->name, method->insns); 1581 if (saveArea->xtra.currentPc != NULL) 1582 LOGE(" pc inst = 0x%04x", *saveArea->xtra.currentPc); 1583 if (thread->currentPc2 != NULL) 1584 LOGE(" pc2 inst = 0x%04x", *thread->currentPc2); 1585 dvmDumpThread(thread, false); 1586 } 1587 } else { 1588 /* 1589 * It's unusual, but not impossible, for a non-first frame 1590 * to be at something other than a method invocation. For 1591 * example, if we do a new-instance on a nonexistent class, 1592 * we'll have a lot of class loader activity on the stack 1593 * above the frame with the "new" operation. Could also 1594 * happen while we initialize a Throwable when an instruction 1595 * fails. 1596 * 1597 * So there's not much we can do here to verify the PC, 1598 * except to verify that it's a GC point. 1599 */ 1600 } 1601 assert(saveArea->xtra.currentPc != NULL); 1602#endif 1603 1604 const RegisterMap* pMap; 1605 const u1* regVector; 1606 1607 Method* nonConstMethod = (Method*) method; // quiet gcc 1608 pMap = dvmGetExpandedRegisterMap(nonConstMethod); 1609 1610 //LOG_SCAV("PGC: %s.%s", method->clazz->descriptor, method->name); 1611 1612 if (pMap != NULL) { 1613 /* found map, get registers for this address */ 1614 int addr = saveArea->xtra.currentPc - method->insns; 1615 regVector = dvmRegisterMapGetLine(pMap, addr); 1616 /* 1617 if (regVector == NULL) { 1618 LOG_SCAV("PGC: map but no entry for %s.%s addr=0x%04x", 1619 method->clazz->descriptor, method->name, addr); 1620 } else { 1621 LOG_SCAV("PGC: found map for %s.%s 0x%04x (t=%d)", 1622 method->clazz->descriptor, method->name, addr, 1623 thread->threadId); 1624 } 1625 */ 1626 } else { 1627 /* 1628 * No map found. If precise GC is disabled this is 1629 * expected -- we don't create pointers to the map data even 1630 * if it's present -- but if it's enabled it means we're 1631 * unexpectedly falling back on a conservative scan, so it's 1632 * worth yelling a little. 1633 */ 1634 if (gDvm.preciseGc) { 1635 LOG_SCAV("PGC: no map for %s.%s", method->clazz->descriptor, method->name); 1636 } 1637 regVector = NULL; 1638 } 1639 if (regVector == NULL) { 1640 /* 1641 * There are no roots to scavenge. Skip over the entire frame. 1642 */ 1643 framePtr += method->registersSize; 1644 } else { 1645 /* 1646 * Precise scan. v0 is at the lowest address on the 1647 * interpreted stack, and is the first bit in the register 1648 * vector, so we can walk through the register map and 1649 * memory in the same direction. 1650 * 1651 * A '1' bit indicates a live reference. 1652 */ 1653 u2 bits = 1 << 1; 1654 for (int i = method->registersSize - 1; i >= 0; i--) { 1655 u4 rval = *framePtr; 1656 1657 bits >>= 1; 1658 if (bits == 1) { 1659 /* set bit 9 so we can tell when we're empty */ 1660 bits = *regVector++ | 0x0100; 1661 } 1662 1663 if (rval != 0 && (bits & 0x01) != 0) { 1664 /* 1665 * Non-null, register marked as live reference. This 1666 * should always be a valid object. 1667 */ 1668#if WITH_EXTRA_GC_CHECKS > 0 1669 if ((rval & 0x3) != 0 || !dvmIsValidObject((Object*) rval)) { 1670 /* this is very bad */ 1671 LOGE("PGC: invalid ref in reg %d: 0x%08x", 1672 method->registersSize-1 - i, rval); 1673 } else 1674#endif 1675 { 1676 1677 // LOG_SCAV("stack reference %u@%p", *framePtr, framePtr); 1678 /* dvmMarkObjectNonNull((Object *)rval); */ 1679 scavengeReference((Object **) framePtr); 1680 } 1681 } else { 1682 /* 1683 * Null or non-reference, do nothing at all. 1684 */ 1685#if WITH_EXTRA_GC_CHECKS > 1 1686 if (dvmIsValidObject((Object*) rval)) { 1687 /* this is normal, but we feel chatty */ 1688 LOGD("PGC: ignoring valid ref in reg %d: 0x%08x", 1689 method->registersSize-1 - i, rval); 1690 } 1691#endif 1692 } 1693 ++framePtr; 1694 } 1695 dvmReleaseRegisterMapLine(pMap, regVector); 1696 } 1697 } 1698 /* else this is a break frame and there is nothing to gray, or 1699 * this is a native method and the registers are just the "ins", 1700 * copied from various registers in the caller's set. 1701 */ 1702 1703#if WITH_EXTRA_GC_CHECKS > 1 1704 first = false; 1705#endif 1706 1707 /* Don't fall into an infinite loop if things get corrupted. 1708 */ 1709 assert((uintptr_t)saveArea->prevFrame > (uintptr_t)framePtr || 1710 saveArea->prevFrame == NULL); 1711 framePtr = saveArea->prevFrame; 1712 } 1713} 1714 1715static void scavengeThread(Thread *thread) 1716{ 1717 // LOG_SCAV("scavengeThread(thread=%p)", thread); 1718 1719 // LOG_SCAV("Scavenging threadObj=%p", thread->threadObj); 1720 scavengeReference(&thread->threadObj); 1721 1722 // LOG_SCAV("Scavenging exception=%p", thread->exception); 1723 scavengeReference(&thread->exception); 1724 1725 scavengeThreadStack(thread); 1726} 1727 1728static void scavengeThreadList() 1729{ 1730 Thread *thread; 1731 1732 dvmLockThreadList(dvmThreadSelf()); 1733 thread = gDvm.threadList; 1734 while (thread) { 1735 scavengeThread(thread); 1736 thread = thread->next; 1737 } 1738 dvmUnlockThreadList(); 1739} 1740 1741static void pinThreadStack(const Thread *thread) 1742{ 1743 const u4 *framePtr; 1744 const StackSaveArea *saveArea; 1745 Method *method; 1746 const char *shorty; 1747 Object *obj; 1748 1749 saveArea = NULL; 1750 framePtr = (const u4 *)thread->interpSave.curFrame; 1751 for (; framePtr != NULL; framePtr = saveArea->prevFrame) { 1752 saveArea = SAVEAREA_FROM_FP(framePtr); 1753 method = (Method *)saveArea->method; 1754 if (method != NULL && dvmIsNativeMethod(method)) { 1755 /* 1756 * This is native method, pin its arguments. 1757 * 1758 * For purposes of graying references, we don't need to do 1759 * anything here, because all of the native "ins" were copied 1760 * from registers in the caller's stack frame and won't be 1761 * changed (an interpreted method can freely use registers 1762 * with parameters like any other register, but natives don't 1763 * work that way). 1764 * 1765 * However, we need to ensure that references visible to 1766 * native methods don't move around. We can do a precise scan 1767 * of the arguments by examining the method signature. 1768 */ 1769 LOG_PIN("+++ native scan %s.%s", 1770 method->clazz->descriptor, method->name); 1771 assert(method->registersSize == method->insSize); 1772 if (!dvmIsStaticMethod(method)) { 1773 /* grab the "this" pointer */ 1774 obj = (Object *)*framePtr++; 1775 if (obj == NULL) { 1776 /* 1777 * This can happen for the "fake" entry frame inserted 1778 * for threads created outside the VM. There's no actual 1779 * call so there's no object. If we changed the fake 1780 * entry method to be declared "static" then this 1781 * situation should never occur. 1782 */ 1783 } else { 1784 assert(dvmIsValidObject(obj)); 1785 pinObject(obj); 1786 } 1787 } 1788 shorty = method->shorty+1; // skip return value 1789 for (int i = method->registersSize - 1; i >= 0; i--, framePtr++) { 1790 switch (*shorty++) { 1791 case 'L': 1792 obj = (Object *)*framePtr; 1793 if (obj != NULL) { 1794 assert(dvmIsValidObject(obj)); 1795 pinObject(obj); 1796 } 1797 break; 1798 case 'D': 1799 case 'J': 1800 framePtr++; 1801 break; 1802 default: 1803 /* 32-bit non-reference value */ 1804 obj = (Object *)*framePtr; // debug, remove 1805 if (dvmIsValidObject(obj)) { // debug, remove 1806 /* if we see a lot of these, our scan might be off */ 1807 LOG_PIN("+++ did NOT pin obj %p", obj); 1808 } 1809 break; 1810 } 1811 } 1812 } else if (method != NULL && !dvmIsNativeMethod(method)) { 1813 const RegisterMap* pMap = dvmGetExpandedRegisterMap(method); 1814 const u1* regVector = NULL; 1815 1816 LOGI("conservative : %s.%s", method->clazz->descriptor, method->name); 1817 1818 if (pMap != NULL) { 1819 int addr = saveArea->xtra.currentPc - method->insns; 1820 regVector = dvmRegisterMapGetLine(pMap, addr); 1821 } 1822 if (regVector == NULL) { 1823 /* 1824 * No register info for this frame, conservatively pin. 1825 */ 1826 for (int i = 0; i < method->registersSize; ++i) { 1827 u4 regValue = framePtr[i]; 1828 if (regValue != 0 && (regValue & 0x3) == 0 && dvmIsValidObject((Object *)regValue)) { 1829 pinObject((Object *)regValue); 1830 } 1831 } 1832 } 1833 } 1834 /* 1835 * Don't fall into an infinite loop if things get corrupted. 1836 */ 1837 assert((uintptr_t)saveArea->prevFrame > (uintptr_t)framePtr || 1838 saveArea->prevFrame == NULL); 1839 } 1840} 1841 1842static void pinThread(const Thread *thread) 1843{ 1844 assert(thread != NULL); 1845 LOG_PIN("pinThread(thread=%p)", thread); 1846 1847 LOG_PIN("Pin native method arguments"); 1848 pinThreadStack(thread); 1849 1850 LOG_PIN("Pin internalLocalRefTable"); 1851 pinReferenceTable(&thread->internalLocalRefTable); 1852 1853 LOG_PIN("Pin jniLocalRefTable"); 1854 pinReferenceTable(&thread->jniLocalRefTable); 1855 1856 /* Can the check be pushed into the promote routine? */ 1857 if (thread->jniMonitorRefTable.table) { 1858 LOG_PIN("Pin jniMonitorRefTable"); 1859 pinReferenceTable(&thread->jniMonitorRefTable); 1860 } 1861} 1862 1863static void pinThreadList() 1864{ 1865 Thread *thread; 1866 1867 dvmLockThreadList(dvmThreadSelf()); 1868 thread = gDvm.threadList; 1869 while (thread) { 1870 pinThread(thread); 1871 thread = thread->next; 1872 } 1873 dvmUnlockThreadList(); 1874} 1875 1876/* 1877 * Heap block scavenging. 1878 */ 1879 1880/* 1881 * Scavenge objects in the current block. Scavenging terminates when 1882 * the pointer reaches the highest address in the block or when a run 1883 * of zero words that continues to the highest address is reached. 1884 */ 1885static void scavengeBlock(HeapSource *heapSource, size_t block) 1886{ 1887 u1 *cursor; 1888 u1 *end; 1889 size_t size; 1890 1891 LOG_SCAV("scavengeBlock(heapSource=%p,block=%zu)", heapSource, block); 1892 1893 assert(heapSource != NULL); 1894 assert(block < heapSource->totalBlocks); 1895 assert(heapSource->blockSpace[block] == BLOCK_TO_SPACE); 1896 1897 cursor = blockToAddress(heapSource, block); 1898 end = cursor + BLOCK_SIZE; 1899 LOG_SCAV("scavengeBlock start=%p, end=%p", cursor, end); 1900 1901 /* Parse and scavenge the current block. */ 1902 size = 0; 1903 while (cursor < end) { 1904 u4 word = *(u4 *)cursor; 1905 if (word != 0) { 1906 scavengeObject((Object *)cursor); 1907 size = objectSize((Object *)cursor); 1908 size = alignUp(size, ALLOC_ALIGNMENT); 1909 cursor += size; 1910 } else { 1911 /* Check for padding. */ 1912 while (*(u4 *)cursor == 0) { 1913 cursor += 4; 1914 if (cursor == end) break; 1915 } 1916 /* Punt if something went wrong. */ 1917 assert(cursor == end); 1918 } 1919 } 1920} 1921 1922static size_t objectSize(const Object *obj) 1923{ 1924 size_t size; 1925 1926 assert(obj != NULL); 1927 assert(obj->clazz != NULL); 1928 if (obj->clazz == gDvm.classJavaLangClass) { 1929 size = dvmClassObjectSize((ClassObject *)obj); 1930 } else if (IS_CLASS_FLAG_SET(obj->clazz, CLASS_ISARRAY)) { 1931 size = dvmArrayObjectSize((ArrayObject *)obj); 1932 } else { 1933 assert(obj->clazz->objectSize != 0); 1934 size = obj->clazz->objectSize; 1935 } 1936 if (LW_HASH_STATE(obj->lock) == LW_HASH_STATE_HASHED_AND_MOVED) { 1937 size += sizeof(u4); 1938 } 1939 return size; 1940} 1941 1942static void verifyBlock(HeapSource *heapSource, size_t block) 1943{ 1944 u1 *cursor; 1945 u1 *end; 1946 size_t size; 1947 1948 // LOG_VER("verifyBlock(heapSource=%p,block=%zu)", heapSource, block); 1949 1950 assert(heapSource != NULL); 1951 assert(block < heapSource->totalBlocks); 1952 assert(heapSource->blockSpace[block] == BLOCK_TO_SPACE); 1953 1954 cursor = blockToAddress(heapSource, block); 1955 end = cursor + BLOCK_SIZE; 1956 // LOG_VER("verifyBlock start=%p, end=%p", cursor, end); 1957 1958 /* Parse and scavenge the current block. */ 1959 size = 0; 1960 while (cursor < end) { 1961 u4 word = *(u4 *)cursor; 1962 if (word != 0) { 1963 dvmVerifyObject((Object *)cursor); 1964 size = objectSize((Object *)cursor); 1965 size = alignUp(size, ALLOC_ALIGNMENT); 1966 cursor += size; 1967 } else { 1968 /* Check for padding. */ 1969 while (*(unsigned long *)cursor == 0) { 1970 cursor += 4; 1971 if (cursor == end) break; 1972 } 1973 /* Punt if something went wrong. */ 1974 assert(cursor == end); 1975 } 1976 } 1977} 1978 1979static void describeBlockQueue(const HeapSource *heapSource) 1980{ 1981 size_t block, count; 1982 char space; 1983 1984 block = heapSource->queueHead; 1985 count = 0; 1986 LOG_SCAV(">>> describeBlockQueue(heapSource=%p)", heapSource); 1987 /* Count the number of blocks enqueued. */ 1988 while (block != QUEUE_TAIL) { 1989 block = heapSource->blockQueue[block]; 1990 ++count; 1991 } 1992 LOG_SCAV("blockQueue %zu elements, enqueued %zu", 1993 count, heapSource->queueSize); 1994 block = heapSource->queueHead; 1995 while (block != QUEUE_TAIL) { 1996 space = heapSource->blockSpace[block]; 1997 LOG_SCAV("block=%zu@%p,space=%zu", block, blockToAddress(heapSource,block), space); 1998 block = heapSource->blockQueue[block]; 1999 } 2000 2001 LOG_SCAV("<<< describeBlockQueue(heapSource=%p)", heapSource); 2002} 2003 2004/* 2005 * Blackens promoted objects. 2006 */ 2007static void scavengeBlockQueue() 2008{ 2009 HeapSource *heapSource; 2010 size_t block; 2011 2012 LOG_SCAV(">>> scavengeBlockQueue()"); 2013 heapSource = gDvm.gcHeap->heapSource; 2014 describeBlockQueue(heapSource); 2015 while (heapSource->queueHead != QUEUE_TAIL) { 2016 block = heapSource->queueHead; 2017 LOG_SCAV("Dequeueing block %zu", block); 2018 scavengeBlock(heapSource, block); 2019 heapSource->queueHead = heapSource->blockQueue[block]; 2020 LOG_SCAV("New queue head is %zu", heapSource->queueHead); 2021 } 2022 LOG_SCAV("<<< scavengeBlockQueue()"); 2023} 2024 2025/* 2026 * Scan the block list and verify all blocks that are marked as being 2027 * in new space. This should be parametrized so we can invoke this 2028 * routine outside of the context of a collection. 2029 */ 2030static void verifyNewSpace() 2031{ 2032 HeapSource *heapSource = gDvm.gcHeap->heapSource; 2033 size_t c0 = 0, c1 = 0, c2 = 0, c7 = 0; 2034 for (size_t i = 0; i < heapSource->totalBlocks; ++i) { 2035 switch (heapSource->blockSpace[i]) { 2036 case BLOCK_FREE: ++c0; break; 2037 case BLOCK_TO_SPACE: ++c1; break; 2038 case BLOCK_FROM_SPACE: ++c2; break; 2039 case BLOCK_CONTINUED: ++c7; break; 2040 default: assert(!"reached"); 2041 } 2042 } 2043 LOG_VER("Block Demographics: " 2044 "Free=%zu,ToSpace=%zu,FromSpace=%zu,Continued=%zu", 2045 c0, c1, c2, c7); 2046 for (size_t i = 0; i < heapSource->totalBlocks; ++i) { 2047 if (heapSource->blockSpace[i] != BLOCK_TO_SPACE) { 2048 continue; 2049 } 2050 verifyBlock(heapSource, i); 2051 } 2052} 2053 2054void describeHeap() 2055{ 2056 HeapSource *heapSource = gDvm.gcHeap->heapSource; 2057 describeBlocks(heapSource); 2058} 2059 2060/* 2061 * The collection interface. Collection has a few distinct phases. 2062 * The first is flipping AKA condemning AKA whitening the heap. The 2063 * second is to promote all objects which are pointed to by pinned or 2064 * ambiguous references. The third phase is tracing from the stacks, 2065 * registers and various globals. Lastly, a verification of the heap 2066 * is performed. The last phase should be optional. 2067 */ 2068void dvmScavengeRoots() /* Needs a new name badly */ 2069{ 2070 GcHeap *gcHeap; 2071 2072 { 2073 size_t alloc, unused, total; 2074 2075 room(&alloc, &unused, &total); 2076 LOG_SCAV("BEFORE GC: %zu alloc, %zu free, %zu total.", 2077 alloc, unused, total); 2078 } 2079 2080 gcHeap = gDvm.gcHeap; 2081 dvmHeapSourceFlip(); 2082 2083 /* 2084 * Promote blocks with stationary objects. 2085 */ 2086 pinThreadList(); 2087 pinReferenceTable(&gDvm.jniGlobalRefTable); 2088 pinReferenceTable(&gDvm.jniPinRefTable); 2089 pinHashTableEntries(gDvm.loadedClasses); 2090 pinHashTableEntries(gDvm.dbgRegistry); 2091 pinPrimitiveClasses(); 2092 pinInternedStrings(); 2093 2094 // describeBlocks(gcHeap->heapSource); 2095 2096 /* 2097 * Create first, open new-space page right here. 2098 */ 2099 2100 /* Reset allocation to an unallocated block. */ 2101 gDvm.gcHeap->heapSource->allocPtr = allocateBlocks(gDvm.gcHeap->heapSource, 1); 2102 gDvm.gcHeap->heapSource->allocLimit = gDvm.gcHeap->heapSource->allocPtr + BLOCK_SIZE; 2103 /* 2104 * Hack: promote the empty block allocated above. If the 2105 * promotions that occurred above did not actually gray any 2106 * objects, the block queue may be empty. We must force a 2107 * promotion to be safe. 2108 */ 2109 promoteBlockByAddr(gDvm.gcHeap->heapSource, gDvm.gcHeap->heapSource->allocPtr); 2110 2111 /* 2112 * Scavenge blocks and relocate movable objects. 2113 */ 2114 2115 LOG_SCAV("Scavenging gDvm.threadList"); 2116 scavengeThreadList(); 2117 2118 LOG_SCAV("Scavenging gDvm.gcHeap->referenceOperations"); 2119 scavengeLargeHeapRefTable(gcHeap->referenceOperations); 2120 2121 LOG_SCAV("Scavenging gDvm.gcHeap->pendingFinalizationRefs"); 2122 scavengeLargeHeapRefTable(gcHeap->pendingFinalizationRefs); 2123 2124 LOG_SCAV("Scavenging random global stuff"); 2125 scavengeReference(&gDvm.outOfMemoryObj); 2126 scavengeReference(&gDvm.internalErrorObj); 2127 scavengeReference(&gDvm.noClassDefFoundErrorObj); 2128 2129 // LOG_SCAV("Scavenging gDvm.internedString"); 2130 scavengeInternedStrings(); 2131 2132 LOG_SCAV("Root scavenge has completed."); 2133 2134 scavengeBlockQueue(); 2135 2136 // LOG_SCAV("Re-snap global class pointers."); 2137 // scavengeGlobals(); 2138 2139 LOG_SCAV("New space scavenge has completed."); 2140 2141 /* 2142 * Process reference objects in strength order. 2143 */ 2144 2145 LOG_REF("Processing soft references..."); 2146 preserveSoftReferences(&gDvm.gcHeap->softReferences); 2147 clearWhiteReferences(&gDvm.gcHeap->softReferences); 2148 2149 LOG_REF("Processing weak references..."); 2150 clearWhiteReferences(&gDvm.gcHeap->weakReferences); 2151 2152 LOG_REF("Finding finalizations..."); 2153 processFinalizableReferences(); 2154 2155 LOG_REF("Processing f-reachable soft references..."); 2156 clearWhiteReferences(&gDvm.gcHeap->softReferences); 2157 2158 LOG_REF("Processing f-reachable weak references..."); 2159 clearWhiteReferences(&gDvm.gcHeap->weakReferences); 2160 2161 LOG_REF("Processing phantom references..."); 2162 clearWhiteReferences(&gDvm.gcHeap->phantomReferences); 2163 2164 /* 2165 * Verify the stack and heap. 2166 */ 2167 dvmVerifyRoots(); 2168 verifyNewSpace(); 2169 2170 //describeBlocks(gcHeap->heapSource); 2171 2172 clearFromSpace(gcHeap->heapSource); 2173 2174 { 2175 size_t alloc, rem, total; 2176 2177 room(&alloc, &rem, &total); 2178 LOG_SCAV("AFTER GC: %zu alloc, %zu free, %zu total.", alloc, rem, total); 2179 } 2180} 2181 2182/* 2183 * Interface compatibility routines. 2184 */ 2185 2186void dvmClearWhiteRefs(Object **list) 2187{ 2188 /* do nothing */ 2189 assert(*list == NULL); 2190} 2191 2192void dvmHandleSoftRefs(Object **list) 2193{ 2194 /* do nothing */ 2195 assert(*list == NULL); 2196} 2197 2198bool dvmHeapBeginMarkStep(GcMode mode) 2199{ 2200 /* do nothing */ 2201 return true; 2202} 2203 2204void dvmHeapFinishMarkStep() 2205{ 2206 /* do nothing */ 2207} 2208 2209void dvmHeapMarkRootSet() 2210{ 2211 /* do nothing */ 2212} 2213 2214void dvmHeapScanMarkedObjects() 2215{ 2216 dvmScavengeRoots(); 2217} 2218 2219void dvmHeapScheduleFinalizations() 2220{ 2221 /* do nothing */ 2222} 2223 2224void dvmHeapSweepUnmarkedObjects(GcMode mode, int *numFreed, size_t *sizeFreed) 2225{ 2226 *numFreed = 0; 2227 *sizeFreed = 0; 2228 /* do nothing */ 2229} 2230 2231void dvmMarkDirtyObjects() 2232{ 2233 assert(!"implemented"); 2234} 2235 2236void dvmHeapSourceThreadShutdown() 2237{ 2238 /* do nothing */ 2239} 2240