1/* Copyright (c) 2006, Google Inc.
2 * All rights reserved.
3 *
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions are
6 * met:
7 *
8 *     * Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 *     * Redistributions in binary form must reproduce the above
11 * copyright notice, this list of conditions and the following disclaimer
12 * in the documentation and/or other materials provided with the
13 * distribution.
14 *     * Neither the name of Google Inc. nor the names of its
15 * contributors may be used to endorse or promote products derived from
16 * this software without specific prior written permission.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
20 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
21 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
22 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
23 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
24 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
25 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
26 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
27 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
28 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
29 *
30 * ---
31 * Author: Maxim Lifantsev
32 */
33
34//
35// Background and key design points of MemoryRegionMap.
36//
37// MemoryRegionMap is a low-level module with quite atypical requirements that
38// result in some degree of non-triviality of the implementation and design.
39//
40// MemoryRegionMap collects info about *all* memory regions created with
41// mmap, munmap, mremap, sbrk.
42// They key word above is 'all': all that are happening in a process
43// during its lifetime frequently starting even before global object
44// constructor execution.
45//
46// This is needed by the primary client of MemoryRegionMap:
47// HeapLeakChecker uses the regions and the associated stack traces
48// to figure out what part of the memory is the heap:
49// if MemoryRegionMap were to miss some (early) regions, leak checking would
50// stop working correctly.
51//
52// To accomplish the goal of functioning before/during global object
53// constructor execution MemoryRegionMap is done as a singleton service
54// that relies on own on-demand initialized static constructor-less data,
55// and only relies on other low-level modules that can also function properly
56// even before global object constructors run.
57//
58// Accomplishing the goal of collecting data about all mmap, munmap, mremap,
59// sbrk occurrences is a more involved: conceptually to do this one needs to
60// record some bits of data in particular about any mmap or sbrk call,
61// but to do that one needs to allocate memory for that data at some point,
62// but all memory allocations in the end themselves come from an mmap
63// or sbrk call (that's how the address space of the process grows).
64//
65// Also note that we need to do all the above recording from
66// within an mmap/sbrk hook which is sometimes/frequently is made by a memory
67// allocator, including the allocator MemoryRegionMap itself must rely on.
68// In the case of heap-checker usage this includes even the very first
69// mmap/sbrk call happening in the program: heap-checker gets activated due to
70// a link-time installed mmap/sbrk hook and it initializes MemoryRegionMap
71// and asks it to record info about this very first call right from that
72// very first hook invocation.
73//
74// MemoryRegionMap is doing its memory allocations via LowLevelAlloc:
75// unlike more complex standard memory allocator, LowLevelAlloc cooperates with
76// MemoryRegionMap by not holding any of its own locks while it calls mmap
77// to get memory, thus we are able to call LowLevelAlloc from
78// our mmap/sbrk hooks without causing a deadlock in it.
79// For the same reason of deadlock prevention the locking in MemoryRegionMap
80// itself is write-recursive which is an exception to Google's mutex usage.
81//
82// We still need to break the infinite cycle of mmap calling our hook,
83// which asks LowLevelAlloc for memory to record this mmap,
84// which (sometimes) causes mmap, which calls our hook, and so on.
85// We do this as follows: on a recursive call of MemoryRegionMap's
86// mmap/sbrk/mremap hook we record the data about the allocation in a
87// static fixed-sized stack (saved_regions and saved_buckets), when the
88// recursion unwinds but before returning from the outer hook call we unwind
89// this stack and move the data from saved_regions and saved_buckets to its
90// permanent place in the RegionSet and "bucket_table" respectively,
91// which can cause more allocations and mmap-s and recursion and unwinding,
92// but the whole process ends eventually due to the fact that for the small
93// allocations we are doing LowLevelAlloc reuses one mmap call and parcels out
94// the memory it created to satisfy several of our allocation requests.
95//
96
97// ========================================================================= //
98
99#include <config.h>
100
101#ifdef HAVE_UNISTD_H
102#include <unistd.h>
103#endif
104#ifdef HAVE_INTTYPES_H
105#include <inttypes.h>
106#endif
107#ifdef HAVE_MMAP
108#include <sys/mman.h>
109#elif !defined(MAP_FAILED)
110#define MAP_FAILED -1  // the only thing we need from mman.h
111#endif
112#ifdef HAVE_PTHREAD
113#include <pthread.h>   // for pthread_t, pthread_self()
114#endif
115#include <stddef.h>
116
117#include <algorithm>
118#include <set>
119
120#include "memory_region_map.h"
121
122#include "base/logging.h"
123#include "base/low_level_alloc.h"
124#include "malloc_hook-inl.h"
125
126#include <gperftools/stacktrace.h>
127#include <gperftools/malloc_hook.h>
128
129// MREMAP_FIXED is a linux extension.  How it's used in this file,
130// setting it to 0 is equivalent to saying, "This feature isn't
131// supported", which is right.
132#ifndef MREMAP_FIXED
133# define MREMAP_FIXED  0
134#endif
135
136using std::max;
137
138// ========================================================================= //
139
140int MemoryRegionMap::client_count_ = 0;
141int MemoryRegionMap::max_stack_depth_ = 0;
142MemoryRegionMap::RegionSet* MemoryRegionMap::regions_ = NULL;
143LowLevelAlloc::Arena* MemoryRegionMap::arena_ = NULL;
144SpinLock MemoryRegionMap::lock_(SpinLock::LINKER_INITIALIZED);
145SpinLock MemoryRegionMap::owner_lock_(  // ACQUIRED_AFTER(lock_)
146    SpinLock::LINKER_INITIALIZED);
147int MemoryRegionMap::recursion_count_ = 0;  // GUARDED_BY(owner_lock_)
148pthread_t MemoryRegionMap::lock_owner_tid_;  // GUARDED_BY(owner_lock_)
149int64 MemoryRegionMap::map_size_ = 0;
150int64 MemoryRegionMap::unmap_size_ = 0;
151HeapProfileBucket** MemoryRegionMap::bucket_table_ = NULL;  // GUARDED_BY(lock_)
152int MemoryRegionMap::num_buckets_ = 0;  // GUARDED_BY(lock_)
153int MemoryRegionMap::saved_buckets_count_ = 0;  // GUARDED_BY(lock_)
154HeapProfileBucket MemoryRegionMap::saved_buckets_[20];  // GUARDED_BY(lock_)
155
156// GUARDED_BY(lock_)
157const void* MemoryRegionMap::saved_buckets_keys_[20][kMaxStackDepth];
158
159// ========================================================================= //
160
161// Simple hook into execution of global object constructors,
162// so that we do not call pthread_self() when it does not yet work.
163static bool libpthread_initialized = false;
164static bool initializer = (libpthread_initialized = true, true);
165
166static inline bool current_thread_is(pthread_t should_be) {
167  // Before main() runs, there's only one thread, so we're always that thread
168  if (!libpthread_initialized) return true;
169  // this starts working only sometime well into global constructor execution:
170  return pthread_equal(pthread_self(), should_be);
171}
172
173// ========================================================================= //
174
175// Constructor-less place-holder to store a RegionSet in.
176union MemoryRegionMap::RegionSetRep {
177  char rep[sizeof(RegionSet)];
178  void* align_it;  // do not need a better alignment for 'rep' than this
179  RegionSet* region_set() { return reinterpret_cast<RegionSet*>(rep); }
180};
181
182// The bytes where MemoryRegionMap::regions_ will point to.
183// We use RegionSetRep with noop c-tor so that global construction
184// does not interfere.
185static MemoryRegionMap::RegionSetRep regions_rep;
186
187// ========================================================================= //
188
189// Has InsertRegionLocked been called recursively
190// (or rather should we *not* use regions_ to record a hooked mmap).
191static bool recursive_insert = false;
192
193void MemoryRegionMap::Init(int max_stack_depth, bool use_buckets) {
194  RAW_VLOG(10, "MemoryRegionMap Init");
195  RAW_CHECK(max_stack_depth >= 0, "");
196  // Make sure we don't overflow the memory in region stacks:
197  RAW_CHECK(max_stack_depth <= kMaxStackDepth,
198            "need to increase kMaxStackDepth?");
199  Lock();
200  client_count_ += 1;
201  max_stack_depth_ = max(max_stack_depth_, max_stack_depth);
202  if (client_count_ > 1) {
203    // not first client: already did initialization-proper
204    Unlock();
205    RAW_VLOG(10, "MemoryRegionMap Init increment done");
206    return;
207  }
208  // Set our hooks and make sure they were installed:
209  RAW_CHECK(MallocHook::AddMmapHook(&MmapHook), "");
210  RAW_CHECK(MallocHook::AddMremapHook(&MremapHook), "");
211  RAW_CHECK(MallocHook::AddSbrkHook(&SbrkHook), "");
212  RAW_CHECK(MallocHook::AddMunmapHook(&MunmapHook), "");
213  // We need to set recursive_insert since the NewArena call itself
214  // will already do some allocations with mmap which our hooks will catch
215  // recursive_insert allows us to buffer info about these mmap calls.
216  // Note that Init() can be (and is) sometimes called
217  // already from within an mmap/sbrk hook.
218  recursive_insert = true;
219  arena_ = LowLevelAlloc::NewArena(0, LowLevelAlloc::DefaultArena());
220  recursive_insert = false;
221  HandleSavedRegionsLocked(&InsertRegionLocked);  // flush the buffered ones
222    // Can't instead use HandleSavedRegionsLocked(&DoInsertRegionLocked) before
223    // recursive_insert = false; as InsertRegionLocked will also construct
224    // regions_ on demand for us.
225  if (use_buckets) {
226    const int table_bytes = kHashTableSize * sizeof(*bucket_table_);
227    recursive_insert = true;
228    bucket_table_ = static_cast<HeapProfileBucket**>(
229        MyAllocator::Allocate(table_bytes));
230    recursive_insert = false;
231    memset(bucket_table_, 0, table_bytes);
232    num_buckets_ = 0;
233  }
234  if (regions_ == NULL)  // init regions_
235    InitRegionSetLocked();
236  Unlock();
237  RAW_VLOG(10, "MemoryRegionMap Init done");
238}
239
240bool MemoryRegionMap::Shutdown() {
241  RAW_VLOG(10, "MemoryRegionMap Shutdown");
242  Lock();
243  RAW_CHECK(client_count_ > 0, "");
244  client_count_ -= 1;
245  if (client_count_ != 0) {  // not last client; need not really shutdown
246    Unlock();
247    RAW_VLOG(10, "MemoryRegionMap Shutdown decrement done");
248    return true;
249  }
250  if (bucket_table_ != NULL) {
251    for (int i = 0; i < kHashTableSize; i++) {
252      for (HeapProfileBucket* curr = bucket_table_[i]; curr != 0; /**/) {
253        HeapProfileBucket* bucket = curr;
254        curr = curr->next;
255        MyAllocator::Free(bucket->stack, 0);
256        MyAllocator::Free(bucket, 0);
257      }
258    }
259    MyAllocator::Free(bucket_table_, 0);
260    num_buckets_ = 0;
261    bucket_table_ = NULL;
262  }
263  RAW_CHECK(MallocHook::RemoveMmapHook(&MmapHook), "");
264  RAW_CHECK(MallocHook::RemoveMremapHook(&MremapHook), "");
265  RAW_CHECK(MallocHook::RemoveSbrkHook(&SbrkHook), "");
266  RAW_CHECK(MallocHook::RemoveMunmapHook(&MunmapHook), "");
267  if (regions_) regions_->~RegionSet();
268  regions_ = NULL;
269  bool deleted_arena = LowLevelAlloc::DeleteArena(arena_);
270  if (deleted_arena) {
271    arena_ = 0;
272  } else {
273    RAW_LOG(WARNING, "Can't delete LowLevelAlloc arena: it's being used");
274  }
275  Unlock();
276  RAW_VLOG(10, "MemoryRegionMap Shutdown done");
277  return deleted_arena;
278}
279
280bool MemoryRegionMap::IsRecordingLocked() {
281  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
282  return client_count_ > 0;
283}
284
285// Invariants (once libpthread_initialized is true):
286//   * While lock_ is not held, recursion_count_ is 0 (and
287//     lock_owner_tid_ is the previous owner, but we don't rely on
288//     that).
289//   * recursion_count_ and lock_owner_tid_ are only written while
290//     both lock_ and owner_lock_ are held. They may be read under
291//     just owner_lock_.
292//   * At entry and exit of Lock() and Unlock(), the current thread
293//     owns lock_ iff pthread_equal(lock_owner_tid_, pthread_self())
294//     && recursion_count_ > 0.
295void MemoryRegionMap::Lock() {
296  {
297    SpinLockHolder l(&owner_lock_);
298    if (recursion_count_ > 0 && current_thread_is(lock_owner_tid_)) {
299      RAW_CHECK(lock_.IsHeld(), "Invariants violated");
300      recursion_count_++;
301      RAW_CHECK(recursion_count_ <= 5,
302                "recursive lock nesting unexpectedly deep");
303      return;
304    }
305  }
306  lock_.Lock();
307  {
308    SpinLockHolder l(&owner_lock_);
309    RAW_CHECK(recursion_count_ == 0,
310              "Last Unlock didn't reset recursion_count_");
311    if (libpthread_initialized)
312      lock_owner_tid_ = pthread_self();
313    recursion_count_ = 1;
314  }
315}
316
317void MemoryRegionMap::Unlock() {
318  SpinLockHolder l(&owner_lock_);
319  RAW_CHECK(recursion_count_ >  0, "unlock when not held");
320  RAW_CHECK(lock_.IsHeld(),
321            "unlock when not held, and recursion_count_ is wrong");
322  RAW_CHECK(current_thread_is(lock_owner_tid_), "unlock by non-holder");
323  recursion_count_--;
324  if (recursion_count_ == 0) {
325    lock_.Unlock();
326  }
327}
328
329bool MemoryRegionMap::LockIsHeld() {
330  SpinLockHolder l(&owner_lock_);
331  return lock_.IsHeld()  &&  current_thread_is(lock_owner_tid_);
332}
333
334const MemoryRegionMap::Region*
335MemoryRegionMap::DoFindRegionLocked(uintptr_t addr) {
336  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
337  if (regions_ != NULL) {
338    Region sample;
339    sample.SetRegionSetKey(addr);
340    RegionSet::iterator region = regions_->lower_bound(sample);
341    if (region != regions_->end()) {
342      RAW_CHECK(addr <= region->end_addr, "");
343      if (region->start_addr <= addr  &&  addr < region->end_addr) {
344        return &(*region);
345      }
346    }
347  }
348  return NULL;
349}
350
351bool MemoryRegionMap::FindRegion(uintptr_t addr, Region* result) {
352  Lock();
353  const Region* region = DoFindRegionLocked(addr);
354  if (region != NULL) *result = *region;  // create it as an independent copy
355  Unlock();
356  return region != NULL;
357}
358
359bool MemoryRegionMap::FindAndMarkStackRegion(uintptr_t stack_top,
360                                             Region* result) {
361  Lock();
362  const Region* region = DoFindRegionLocked(stack_top);
363  if (region != NULL) {
364    RAW_VLOG(10, "Stack at %p is inside region %p..%p",
365                reinterpret_cast<void*>(stack_top),
366                reinterpret_cast<void*>(region->start_addr),
367                reinterpret_cast<void*>(region->end_addr));
368    const_cast<Region*>(region)->set_is_stack();  // now we know
369      // cast is safe (set_is_stack does not change the set ordering key)
370    *result = *region;  // create *result as an independent copy
371  }
372  Unlock();
373  return region != NULL;
374}
375
376HeapProfileBucket* MemoryRegionMap::GetBucket(int depth,
377                                              const void* const key[]) {
378  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
379  // Make hash-value
380  uintptr_t hash = 0;
381  for (int i = 0; i < depth; i++) {
382    hash += reinterpret_cast<uintptr_t>(key[i]);
383    hash += hash << 10;
384    hash ^= hash >> 6;
385  }
386  hash += hash << 3;
387  hash ^= hash >> 11;
388
389  // Lookup stack trace in table
390  unsigned int hash_index = (static_cast<unsigned int>(hash)) % kHashTableSize;
391  for (HeapProfileBucket* bucket = bucket_table_[hash_index];
392       bucket != 0;
393       bucket = bucket->next) {
394    if ((bucket->hash == hash) && (bucket->depth == depth) &&
395        std::equal(key, key + depth, bucket->stack)) {
396      return bucket;
397    }
398  }
399
400  // Create new bucket
401  const size_t key_size = sizeof(key[0]) * depth;
402  HeapProfileBucket* bucket;
403  if (recursive_insert) {  // recursion: save in saved_buckets_
404    const void** key_copy = saved_buckets_keys_[saved_buckets_count_];
405    std::copy(key, key + depth, key_copy);
406    bucket = &saved_buckets_[saved_buckets_count_];
407    memset(bucket, 0, sizeof(*bucket));
408    ++saved_buckets_count_;
409    bucket->stack = key_copy;
410    bucket->next  = NULL;
411  } else {
412    recursive_insert = true;
413    const void** key_copy = static_cast<const void**>(
414        MyAllocator::Allocate(key_size));
415    recursive_insert = false;
416    std::copy(key, key + depth, key_copy);
417    recursive_insert = true;
418    bucket = static_cast<HeapProfileBucket*>(
419        MyAllocator::Allocate(sizeof(HeapProfileBucket)));
420    recursive_insert = false;
421    memset(bucket, 0, sizeof(*bucket));
422    bucket->stack = key_copy;
423    bucket->next  = bucket_table_[hash_index];
424  }
425  bucket->hash = hash;
426  bucket->depth = depth;
427  bucket_table_[hash_index] = bucket;
428  ++num_buckets_;
429  return bucket;
430}
431
432MemoryRegionMap::RegionIterator MemoryRegionMap::BeginRegionLocked() {
433  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
434  RAW_CHECK(regions_ != NULL, "");
435  return regions_->begin();
436}
437
438MemoryRegionMap::RegionIterator MemoryRegionMap::EndRegionLocked() {
439  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
440  RAW_CHECK(regions_ != NULL, "");
441  return regions_->end();
442}
443
444inline void MemoryRegionMap::DoInsertRegionLocked(const Region& region) {
445  RAW_VLOG(12, "Inserting region %p..%p from %p",
446              reinterpret_cast<void*>(region.start_addr),
447              reinterpret_cast<void*>(region.end_addr),
448              reinterpret_cast<void*>(region.caller()));
449  RegionSet::const_iterator i = regions_->lower_bound(region);
450  if (i != regions_->end() && i->start_addr <= region.start_addr) {
451    RAW_DCHECK(region.end_addr <= i->end_addr, "");  // lower_bound ensures this
452    return;  // 'region' is a subset of an already recorded region; do nothing
453    // We can be stricter and allow this only when *i has been created via
454    // an mmap with MAP_NORESERVE flag set.
455  }
456  if (DEBUG_MODE) {
457    RAW_CHECK(i == regions_->end()  ||  !region.Overlaps(*i),
458              "Wow, overlapping memory regions");
459    Region sample;
460    sample.SetRegionSetKey(region.start_addr);
461    i = regions_->lower_bound(sample);
462    RAW_CHECK(i == regions_->end()  ||  !region.Overlaps(*i),
463              "Wow, overlapping memory regions");
464  }
465  region.AssertIsConsistent();  // just making sure
466  // This inserts and allocates permanent storage for region
467  // and its call stack data: it's safe to do it now:
468  regions_->insert(region);
469  RAW_VLOG(12, "Inserted region %p..%p :",
470              reinterpret_cast<void*>(region.start_addr),
471              reinterpret_cast<void*>(region.end_addr));
472  if (VLOG_IS_ON(12))  LogAllLocked();
473}
474
475// These variables are local to MemoryRegionMap::InsertRegionLocked()
476// and MemoryRegionMap::HandleSavedRegionsLocked()
477// and are file-level to ensure that they are initialized at load time.
478
479// Number of unprocessed region inserts.
480static int saved_regions_count = 0;
481
482// Unprocessed inserts (must be big enough to hold all allocations that can
483// be caused by a InsertRegionLocked call).
484// Region has no constructor, so that c-tor execution does not interfere
485// with the any-time use of the static memory behind saved_regions.
486static MemoryRegionMap::Region saved_regions[20];
487
488inline void MemoryRegionMap::HandleSavedRegionsLocked(
489              void (*insert_func)(const Region& region)) {
490  while (saved_regions_count > 0) {
491    // Making a local-var copy of the region argument to insert_func
492    // including its stack (w/o doing any memory allocations) is important:
493    // in many cases the memory in saved_regions
494    // will get written-to during the (*insert_func)(r) call below.
495    Region r = saved_regions[--saved_regions_count];
496    (*insert_func)(r);
497  }
498}
499
500void MemoryRegionMap::RestoreSavedBucketsLocked() {
501  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
502  while (saved_buckets_count_ > 0) {
503    HeapProfileBucket bucket = saved_buckets_[--saved_buckets_count_];
504    unsigned int hash_index =
505        static_cast<unsigned int>(bucket.hash) % kHashTableSize;
506    bool is_found = false;
507    for (HeapProfileBucket* curr = bucket_table_[hash_index];
508         curr != 0;
509         curr = curr->next) {
510      if ((curr->hash == bucket.hash) && (curr->depth == bucket.depth) &&
511          std::equal(bucket.stack, bucket.stack + bucket.depth, curr->stack)) {
512        curr->allocs += bucket.allocs;
513        curr->alloc_size += bucket.alloc_size;
514        curr->frees += bucket.frees;
515        curr->free_size += bucket.free_size;
516        is_found = true;
517        break;
518      }
519    }
520    if (is_found) continue;
521
522    const size_t key_size = sizeof(bucket.stack[0]) * bucket.depth;
523    const void** key_copy = static_cast<const void**>(
524        MyAllocator::Allocate(key_size));
525    std::copy(bucket.stack, bucket.stack + bucket.depth, key_copy);
526    HeapProfileBucket* new_bucket = static_cast<HeapProfileBucket*>(
527        MyAllocator::Allocate(sizeof(HeapProfileBucket)));
528    memset(new_bucket, 0, sizeof(*new_bucket));
529    new_bucket->hash = bucket.hash;
530    new_bucket->depth = bucket.depth;
531    new_bucket->stack = key_copy;
532    new_bucket->next = bucket_table_[hash_index];
533    bucket_table_[hash_index] = new_bucket;
534    ++num_buckets_;
535  }
536}
537
538inline void MemoryRegionMap::InitRegionSetLocked() {
539  RAW_VLOG(12, "Initializing region set");
540  regions_ = regions_rep.region_set();
541  recursive_insert = true;
542  new(regions_) RegionSet();
543  HandleSavedRegionsLocked(&DoInsertRegionLocked);
544  recursive_insert = false;
545}
546
547inline void MemoryRegionMap::InsertRegionLocked(const Region& region) {
548  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
549  // We can be called recursively, because RegionSet constructor
550  // and DoInsertRegionLocked() (called below) can call the allocator.
551  // recursive_insert tells us if that's the case. When this happens,
552  // region insertion information is recorded in saved_regions[],
553  // and taken into account when the recursion unwinds.
554  // Do the insert:
555  if (recursive_insert) {  // recursion: save in saved_regions
556    RAW_VLOG(12, "Saving recursive insert of region %p..%p from %p",
557                reinterpret_cast<void*>(region.start_addr),
558                reinterpret_cast<void*>(region.end_addr),
559                reinterpret_cast<void*>(region.caller()));
560    RAW_CHECK(saved_regions_count < arraysize(saved_regions), "");
561    // Copy 'region' to saved_regions[saved_regions_count]
562    // together with the contents of its call_stack,
563    // then increment saved_regions_count.
564    saved_regions[saved_regions_count++] = region;
565  } else {  // not a recusrive call
566    if (regions_ == NULL)  // init regions_
567      InitRegionSetLocked();
568    recursive_insert = true;
569    // Do the actual insertion work to put new regions into regions_:
570    DoInsertRegionLocked(region);
571    HandleSavedRegionsLocked(&DoInsertRegionLocked);
572    recursive_insert = false;
573  }
574}
575
576// We strip out different number of stack frames in debug mode
577// because less inlining happens in that case
578#ifdef NDEBUG
579static const int kStripFrames = 1;
580#else
581static const int kStripFrames = 3;
582#endif
583
584void MemoryRegionMap::RecordRegionAddition(const void* start, size_t size) {
585  // Record start/end info about this memory acquisition call in a new region:
586  Region region;
587  region.Create(start, size);
588  // First get the call stack info into the local varible 'region':
589  const int depth =
590    max_stack_depth_ > 0
591    ? MallocHook::GetCallerStackTrace(const_cast<void**>(region.call_stack),
592                                      max_stack_depth_, kStripFrames + 1)
593    : 0;
594  region.set_call_stack_depth(depth);  // record stack info fully
595  RAW_VLOG(10, "New global region %p..%p from %p",
596              reinterpret_cast<void*>(region.start_addr),
597              reinterpret_cast<void*>(region.end_addr),
598              reinterpret_cast<void*>(region.caller()));
599  // Note: none of the above allocates memory.
600  Lock();  // recursively lock
601  map_size_ += size;
602  InsertRegionLocked(region);
603    // This will (eventually) allocate storage for and copy over the stack data
604    // from region.call_stack_data_ that is pointed by region.call_stack().
605  if (bucket_table_ != NULL) {
606    HeapProfileBucket* b = GetBucket(depth, region.call_stack);
607    ++b->allocs;
608    b->alloc_size += size;
609    if (!recursive_insert) {
610      recursive_insert = true;
611      RestoreSavedBucketsLocked();
612      recursive_insert = false;
613    }
614  }
615  Unlock();
616}
617
618void MemoryRegionMap::RecordRegionRemoval(const void* start, size_t size) {
619  Lock();
620  if (recursive_insert) {
621    // First remove the removed region from saved_regions, if it's
622    // there, to prevent overrunning saved_regions in recursive
623    // map/unmap call sequences, and also from later inserting regions
624    // which have already been unmapped.
625    uintptr_t start_addr = reinterpret_cast<uintptr_t>(start);
626    uintptr_t end_addr = start_addr + size;
627    int put_pos = 0;
628    int old_count = saved_regions_count;
629    for (int i = 0; i < old_count; ++i, ++put_pos) {
630      Region& r = saved_regions[i];
631      if (r.start_addr == start_addr && r.end_addr == end_addr) {
632        // An exact match, so it's safe to remove.
633        RecordRegionRemovalInBucket(r.call_stack_depth, r.call_stack, size);
634        --saved_regions_count;
635        --put_pos;
636        RAW_VLOG(10, ("Insta-Removing saved region %p..%p; "
637                     "now have %d saved regions"),
638                 reinterpret_cast<void*>(start_addr),
639                 reinterpret_cast<void*>(end_addr),
640                 saved_regions_count);
641      } else {
642        if (put_pos < i) {
643          saved_regions[put_pos] = saved_regions[i];
644        }
645      }
646    }
647  }
648  if (regions_ == NULL) {  // We must have just unset the hooks,
649                           // but this thread was already inside the hook.
650    Unlock();
651    return;
652  }
653  if (!recursive_insert) {
654    HandleSavedRegionsLocked(&InsertRegionLocked);
655  }
656    // first handle adding saved regions if any
657  uintptr_t start_addr = reinterpret_cast<uintptr_t>(start);
658  uintptr_t end_addr = start_addr + size;
659  // subtract start_addr, end_addr from all the regions
660  RAW_VLOG(10, "Removing global region %p..%p; have %" PRIuS " regions",
661              reinterpret_cast<void*>(start_addr),
662              reinterpret_cast<void*>(end_addr),
663              regions_->size());
664  Region sample;
665  sample.SetRegionSetKey(start_addr);
666  // Only iterate over the regions that might overlap start_addr..end_addr:
667  for (RegionSet::iterator region = regions_->lower_bound(sample);
668       region != regions_->end()  &&  region->start_addr < end_addr;
669       /*noop*/) {
670    RAW_VLOG(13, "Looking at region %p..%p",
671                reinterpret_cast<void*>(region->start_addr),
672                reinterpret_cast<void*>(region->end_addr));
673    if (start_addr <= region->start_addr  &&
674        region->end_addr <= end_addr) {  // full deletion
675      RAW_VLOG(12, "Deleting region %p..%p",
676                  reinterpret_cast<void*>(region->start_addr),
677                  reinterpret_cast<void*>(region->end_addr));
678      RecordRegionRemovalInBucket(region->call_stack_depth, region->call_stack,
679                                  region->end_addr - region->start_addr);
680      RegionSet::iterator d = region;
681      ++region;
682      regions_->erase(d);
683      continue;
684    } else if (region->start_addr < start_addr  &&
685               end_addr < region->end_addr) {  // cutting-out split
686      RAW_VLOG(12, "Splitting region %p..%p in two",
687                  reinterpret_cast<void*>(region->start_addr),
688                  reinterpret_cast<void*>(region->end_addr));
689      RecordRegionRemovalInBucket(region->call_stack_depth, region->call_stack,
690                                  end_addr - start_addr);
691      // Make another region for the start portion:
692      // The new region has to be the start portion because we can't
693      // just modify region->end_addr as it's the sorting key.
694      Region r = *region;
695      r.set_end_addr(start_addr);
696      InsertRegionLocked(r);
697      // cut *region from start:
698      const_cast<Region&>(*region).set_start_addr(end_addr);
699    } else if (end_addr > region->start_addr  &&
700               start_addr <= region->start_addr) {  // cut from start
701      RAW_VLOG(12, "Start-chopping region %p..%p",
702                  reinterpret_cast<void*>(region->start_addr),
703                  reinterpret_cast<void*>(region->end_addr));
704      RecordRegionRemovalInBucket(region->call_stack_depth, region->call_stack,
705                                  end_addr - region->start_addr);
706      const_cast<Region&>(*region).set_start_addr(end_addr);
707    } else if (start_addr > region->start_addr  &&
708               start_addr < region->end_addr) {  // cut from end
709      RAW_VLOG(12, "End-chopping region %p..%p",
710                  reinterpret_cast<void*>(region->start_addr),
711                  reinterpret_cast<void*>(region->end_addr));
712      RecordRegionRemovalInBucket(region->call_stack_depth, region->call_stack,
713                                  region->end_addr - start_addr);
714      // Can't just modify region->end_addr (it's the sorting key):
715      Region r = *region;
716      r.set_end_addr(start_addr);
717      RegionSet::iterator d = region;
718      ++region;
719      // It's safe to erase before inserting since r is independent of *d:
720      // r contains an own copy of the call stack:
721      regions_->erase(d);
722      InsertRegionLocked(r);
723      continue;
724    }
725    ++region;
726  }
727  RAW_VLOG(12, "Removed region %p..%p; have %" PRIuS " regions",
728              reinterpret_cast<void*>(start_addr),
729              reinterpret_cast<void*>(end_addr),
730              regions_->size());
731  if (VLOG_IS_ON(12))  LogAllLocked();
732  unmap_size_ += size;
733  Unlock();
734}
735
736void MemoryRegionMap::RecordRegionRemovalInBucket(int depth,
737                                                  const void* const stack[],
738                                                  size_t size) {
739  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
740  if (bucket_table_ == NULL) return;
741  HeapProfileBucket* b = GetBucket(depth, stack);
742  ++b->frees;
743  b->free_size += size;
744}
745
746void MemoryRegionMap::MmapHook(const void* result,
747                               const void* start, size_t size,
748                               int prot, int flags,
749                               int fd, off_t offset) {
750  // TODO(maxim): replace all 0x%"PRIxS" by %p when RAW_VLOG uses a safe
751  // snprintf reimplementation that does not malloc to pretty-print NULL
752  RAW_VLOG(10, "MMap = 0x%" PRIxPTR " of %" PRIuS " at %" PRIu64 " "
753              "prot %d flags %d fd %d offs %" PRId64,
754              reinterpret_cast<uintptr_t>(result), size,
755              reinterpret_cast<uint64>(start), prot, flags, fd,
756              static_cast<int64>(offset));
757  if (result != reinterpret_cast<void*>(MAP_FAILED)  &&  size != 0) {
758    RecordRegionAddition(result, size);
759  }
760}
761
762void MemoryRegionMap::MunmapHook(const void* ptr, size_t size) {
763  RAW_VLOG(10, "MUnmap of %p %" PRIuS, ptr, size);
764  if (size != 0) {
765    RecordRegionRemoval(ptr, size);
766  }
767}
768
769void MemoryRegionMap::MremapHook(const void* result,
770                                 const void* old_addr, size_t old_size,
771                                 size_t new_size, int flags,
772                                 const void* new_addr) {
773  RAW_VLOG(10, "MRemap = 0x%" PRIxPTR " of 0x%" PRIxPTR " %" PRIuS " "
774              "to %" PRIuS " flags %d new_addr=0x%" PRIxPTR,
775              (uintptr_t)result, (uintptr_t)old_addr,
776               old_size, new_size, flags,
777               flags & MREMAP_FIXED ? (uintptr_t)new_addr : 0);
778  if (result != reinterpret_cast<void*>(-1)) {
779    RecordRegionRemoval(old_addr, old_size);
780    RecordRegionAddition(result, new_size);
781  }
782}
783
784extern "C" void* __sbrk(ptrdiff_t increment);  // defined in libc
785
786void MemoryRegionMap::SbrkHook(const void* result, ptrdiff_t increment) {
787  RAW_VLOG(10, "Sbrk = 0x%" PRIxPTR " of %" PRIdS,
788           (uintptr_t)result, increment);
789  if (result != reinterpret_cast<void*>(-1)) {
790    if (increment > 0) {
791      void* new_end = sbrk(0);
792      RecordRegionAddition(result, reinterpret_cast<uintptr_t>(new_end) -
793                                   reinterpret_cast<uintptr_t>(result));
794    } else if (increment < 0) {
795      void* new_end = sbrk(0);
796      RecordRegionRemoval(new_end, reinterpret_cast<uintptr_t>(result) -
797                                   reinterpret_cast<uintptr_t>(new_end));
798    }
799  }
800}
801
802void MemoryRegionMap::LogAllLocked() {
803  RAW_CHECK(LockIsHeld(), "should be held (by this thread)");
804  RAW_LOG(INFO, "List of regions:");
805  uintptr_t previous = 0;
806  for (RegionSet::const_iterator r = regions_->begin();
807       r != regions_->end(); ++r) {
808    RAW_LOG(INFO, "Memory region 0x%" PRIxPTR "..0x%" PRIxPTR " "
809                  "from 0x%" PRIxPTR " stack=%d",
810                  r->start_addr, r->end_addr, r->caller(), r->is_stack);
811    RAW_CHECK(previous < r->end_addr, "wow, we messed up the set order");
812      // this must be caused by uncontrolled recursive operations on regions_
813    previous = r->end_addr;
814  }
815  RAW_LOG(INFO, "End of regions list");
816}
817