/art/runtime/ |
H A D | mem_map_test.cc | 39 const size_t page_size = static_cast<size_t>(kPageSize); 117 uintptr_t random_start = CreateStartPos(i * kPageSize); 143 kPageSize, 158 kPageSize, 175 kPageSize, 186 kPageSize, 197 kPageSize, 239 ptr -= kPageSize; // Now it's close to the top. 242 2 * kPageSize, // brings it over the top. 258 kPageSize, [all...] |
H A D | globals.h | 41 static constexpr int kPageSize = 4096; member in namespace:art 45 static constexpr size_t kLargeObjectAlignment = kPageSize;
|
H A D | image.cc | 51 CHECK_EQ(image_begin, RoundUp(image_begin, kPageSize)); 52 CHECK_EQ(oat_file_begin, RoundUp(oat_file_begin, kPageSize)); 53 CHECK_EQ(oat_data_begin, RoundUp(oat_data_begin, kPageSize)); 66 CHECK_ALIGNED(delta, kPageSize) << " patch delta must be page aligned"; 102 if (!IsAligned<kPageSize>(patch_delta_)) {
|
H A D | mem_map.cc | 95 // & ~(kPageSize - 1) =~0000000000000001111 118 constexpr uintptr_t mask = mask_ones & ~(kPageSize - 1); 267 size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); 334 for (uintptr_t ptr = next_mem_pos_; ptr < 4 * GB; ptr += kPageSize) { 343 CHECK_ALIGNED(ptr, kPageSize); 354 CHECK_ALIGNED(ptr, kPageSize); 370 ptr = LOW_MEM_START - kPageSize; 383 for (tail_ptr = ptr; tail_ptr < ptr + page_aligned_byte_count; tail_ptr += kPageSize) { 384 if (msync(reinterpret_cast<void*>(tail_ptr), kPageSize, 0) == 0) { 446 const size_t page_aligned_byte_count = RoundUp(byte_count, kPageSize); [all...] |
H A D | oat.cc | 100 CHECK(IsAligned<kPageSize>(image_file_location_oat_data_begin)); 131 if (!IsAligned<kPageSize>(executable_offset_)) { 134 if (!IsAligned<kPageSize>(image_patch_delta_)) { 153 if (!IsAligned<kPageSize>(executable_offset_)) { 156 if (!IsAligned<kPageSize>(image_patch_delta_)) { 190 DCHECK_ALIGNED(executable_offset_, kPageSize); 196 DCHECK_ALIGNED(executable_offset, kPageSize); 346 CHECK_ALIGNED(delta, kPageSize); 355 CHECK_ALIGNED(off, kPageSize);
|
H A D | monitor_pool.h | 163 static constexpr size_t kChunkCapacity = kPageSize / kAlignedMonitorSize; 166 static constexpr size_t kChunkSize = kPageSize;
|
H A D | indirect_reference_table.cc | 265 auto* release_start = AlignUp(reinterpret_cast<uint8_t*>(&table_[top_index]), kPageSize);
|
/art/runtime/gc/allocator/ |
H A D | rosalloc.cc | 48 size_t RosAlloc::dedicated_full_run_storage_[kPageSize / sizeof(size_t)] = { 0 }; 62 DCHECK_EQ(RoundUp(capacity, kPageSize), capacity); 63 DCHECK_EQ(RoundUp(max_capacity, kPageSize), max_capacity); 65 CHECK(IsAligned<kPageSize>(page_release_size_threshold_)); 81 size_t num_of_pages = footprint_ / kPageSize; 82 size_t max_num_of_pages = max_capacity_ / kPageSize; 85 RoundUp(max_num_of_pages, kPageSize), 97 DCHECK_EQ(capacity_ % kPageSize, static_cast<size_t>(0)); 119 const size_t req_byte_size = num_pages * kPageSize; 125 DCHECK_EQ(fpr_byte_size % kPageSize, static_cas [all...] |
H A D | dlmalloc.cc | 65 start = reinterpret_cast<void*>(art::RoundUp(reinterpret_cast<uintptr_t>(start), art::kPageSize)); 66 end = reinterpret_cast<void*>(art::RoundDown(reinterpret_cast<uintptr_t>(end), art::kPageSize));
|
H A D | rosalloc.h | 59 DCHECK_ALIGNED(byte_size, kPageSize); 64 DCHECK_EQ(byte_size % kPageSize, static_cast<size_t>(0)); 105 DCHECK_EQ(byte_size % kPageSize, static_cast<size_t>(0)); 196 return reinterpret_cast<uint8_t*>(this) + kPageSize * numOfPages[size_bracket_idx_]; 379 DCHECK_EQ(byte_offset % static_cast<size_t>(kPageSize), static_cast<size_t>(0)); 380 return byte_offset / kPageSize; 385 return (reinterpret_cast<uintptr_t>(addr) - reinterpret_cast<uintptr_t>(base_)) / kPageSize; 596 return RoundUp(bytes, kPageSize); 635 DCHECK_LT(idx, capacity_ / kPageSize);
|
/art/runtime/gc/space/ |
H A D | malloc_space.cc | 88 *growth_limit = RoundUp(*growth_limit, kPageSize); 89 *capacity = RoundUp(*capacity, kPageSize); 123 growth_limit = RoundUp(growth_limit, kPageSize); 166 SetEnd(reinterpret_cast<uint8_t*>(RoundUp(reinterpret_cast<uintptr_t>(End()), kPageSize))); 169 DCHECK(IsAligned<kPageSize>(begin_)); 170 DCHECK(IsAligned<kPageSize>(End())); 171 size_t size = RoundUp(Size(), kPageSize); 183 SetGrowthLimit(RoundUp(size, kPageSize));
|
H A D | image_space.cc | 57 CHECK_ALIGNED(min_delta, kPageSize); 58 CHECK_ALIGNED(max_delta, kPageSize); 66 r = RoundUp(r, kPageSize); 68 r = RoundDown(r, kPageSize); 72 CHECK_ALIGNED(r, kPageSize);
|
H A D | large_object_space.h | 161 static constexpr size_t kAlignment = kPageSize;
|
/art/runtime/gc/collector/ |
H A D | semi_space-inl.h | 37 CHECK(IsAligned<kPageSize>(obj));
|
H A D | semi_space.cc | 442 if (LIKELY(size <= static_cast<size_t>(kPageSize))) { 460 size_t page_remain = AlignUp(byte_dest, kPageSize) - byte_dest; 465 DCHECK_ALIGNED(reinterpret_cast<uintptr_t>(byte_dest), kPageSize); local 468 while (byte_src + kPageSize < limit) { 472 for (size_t i = 0; i < kPageSize / sizeof(*word_src); ++i) { 481 saved_bytes += kPageSize; 483 byte_src += kPageSize; 484 byte_dest += kPageSize;
|
/art/runtime/gc/accounting/ |
H A D | card_table.cc | 120 uint8_t* round_start = AlignUp(start_card, kPageSize); 121 uint8_t* round_end = AlignDown(end_card, kPageSize);
|
H A D | bitmap.cc | 45 RoundUp(num_bits, kBitsPerBitmapWord) / kBitsPerBitmapWord * sizeof(uintptr_t), kPageSize);
|
H A D | space_bitmap_test.cc | 165 RunTest<kPageSize>();
|
/art/imgdiag/ |
H A D | imgdiag.cc | 178 const uint8_t* image_begin = AlignDown(image_begin_unaligned, kPageSize); 179 const uint8_t* image_end = AlignUp(image_end_unaligned, kPageSize); 246 for (uintptr_t begin = boot_map.start; begin != boot_map.end; begin += kPageSize) { 255 if (memcmp(local_ptr, remote_ptr, kPageSize) != 0) { 259 for (size_t i = 0; i < kPageSize / sizeof(uint32_t); ++i) { 281 virtual_page_idx = reinterpret_cast<uintptr_t>(local_ptr) / kPageSize; 284 page_idx = (offset + page_off_begin) / kPageSize; 291 size_t remote_virtual_page_idx = begin / kPageSize; 380 current_page_idx = object_address / kPageSize + page_off; 388 } while ((current_page_idx * kPageSize) < [all...] |
/art/runtime/jit/ |
H A D | jit_code_cache.cc | 48 uint8_t* divider = mem_map->Begin() + RoundUp(mem_map->Size() / 4, kPageSize);
|
/art/compiler/utils/ |
H A D | swap_space.cc | 135 size_t next_part = std::max(RoundUp(min_size, kPageSize), RoundUp(kMininumMapSize, kPageSize));
|
/art/runtime/arch/x86/ |
H A D | thread_x86.cc | 48 const size_t limit = kPageSize;
|
/art/compiler/ |
H A D | common_compiler_test.cc | 113 uintptr_t base = RoundDown(data, kPageSize); 114 uintptr_t limit = RoundUp(data + code_length, kPageSize);
|
H A D | elf_builder.h | 122 link, 0, kPageSize, sizeof(Elf_Dyn)) {} 249 : Section(name, SHT_NOBITS, SHF_ALLOC, nullptr, 0, kPageSize, 0), 502 nullptr, 0, kPageSize, 0, rodata_size, rodata_writer), 504 nullptr, 0, kPageSize, 0, text_size, text_writer),
|
H A D | image_writer.cc | 211 CHECK_ALIGNED(bitmap_section.Offset(), kPageSize); 522 kPageSize); 534 "image bitmap", image_->Begin(), RoundUp(image_end_, kPageSize))); 999 *bitmap_section = ImageSection(RoundUp(cur_pos, kPageSize), RoundUp(bitmap_bytes, kPageSize)); 1010 CHECK_EQ(AlignUp(image_begin_ + image_end, kPageSize), oat_file_begin) << 1506 bin_slot_sizes_[kBinArtMethodClean] + intern_table_bytes_, kPageSize);
|