/art/disassembler/ |
H A D | disassembler_arm.cc | 235 int count = (rhs.first.size != 0 ? ((rhs.imm8 + 1u) >> 1) : rhs.imm8); local 236 if (count > 1) { 237 os << "-" << FpRegister(rhs.first, count - 1); 1831 size_t count = 3 - CTZ(mask); local 1832 it_conditions_.resize(count + 2); // Plus the implicit 't', plus the "" for the IT itself. 1833 for (size_t i = 0; i < count; ++i) { 1844 it_conditions_[count] = kConditionCodeNames[first_cond]; // The implicit 't'. 1846 it_conditions_[count + 1] = ""; // No condition code for the IT itself...
|
/art/imgdiag/ |
H A D | imgdiag.cc | 290 // Independently count the # of dirty pages on the remote side 327 std::map<mirror::Class*, int /*count*/> dirty_object_class_map; 332 std::map<mirror::Class*, int /*count*/> clean_object_class_map; 336 std::map<off_t /* field offset */, int /* count */> art_method_field_dirty_count; 339 std::map<off_t /* field offset */, int /* count */> class_field_dirty_count; 491 // vector of pairs (int count, Class*) 495 os << "\n" << " Dirty object count by class:\n"; 520 os << " dirty byte +offset:count list = "; 524 int count = pair.first; local 526 os << "+" << offset << ":" << count << ", "; local 578 int count = pair.first; local 580 os << "+" << offset << ":" << count << ", "; local [all...] |
/art/oatdump/ |
H A D | oatdump.cc | 1656 for (size_t i = 0, count = klass->NumInstanceFields(); i < count; i++) { 1880 if (already_seen_.count(oat_data) == 0) { 1944 SizeAndCount(size_t bytes_in, size_t count_in) : bytes(bytes_in), count(count_in) {} 1946 size_t count; member in struct:art::ImageDumper::Stats::SizeAndCount 1955 it->second.count += 1; 2111 static_cast<double>(sizes_and_count.second.count); 2116 sizes_and_count.second.count, average, percent);
|
/art/patchoat/ |
H A D | patchoat.cc | 445 void VisitRoots(mirror::Object*** roots, size_t count, const RootInfo& info ATTRIBUTE_UNUSED) 447 for (size_t i = 0; i < count; ++i) { 452 void VisitRoots(mirror::CompressedReference<mirror::Object>** roots, size_t count, 455 for (size_t i = 0; i < count; ++i) { 478 for (size_t i = 0, count = dex_caches->GetLength(); i < count; ++i) {
|
/art/runtime/arch/arm/ |
H A D | quick_entrypoints_arm.S | 476 orr r2, r1, r2 @ r2 holds thread id with count of 0 with preserved read barrier bits 481 .Lnot_unlocked: @ r1: original lock word, r2: thread_id with count of 0 and zero read barrier bits 488 mov r3, r1 @ copy the lock word to check count overflow. 490 add r2, r3, #LOCK_WORD_THIN_LOCK_COUNT_ONE @ increment count in lock word placing in r2 to check overflow 492 cbnz r3, .Lslow_lock @ if we overflow the count go slow path 493 add r2, r1, #LOCK_WORD_THIN_LOCK_COUNT_ONE @ increment count for real 545 sub r1, r1, #LOCK_WORD_THIN_LOCK_COUNT_ONE @ decrement count 1138 * r2: shift count 1160 * r2: shift count 1182 * r2: shift count [all...] |
/art/runtime/arch/ |
H A D | memcmp16.cc | 22 int32_t memcmp16_generic_static(const uint16_t* s0, const uint16_t* s1, size_t count); 23 int32_t memcmp16_generic_static(const uint16_t* s0, const uint16_t* s1, size_t count) { argument 24 for (size_t i = 0; i < count; i++) { 36 int32_t MemCmp16Testing(const uint16_t* s0, const uint16_t* s1, size_t count) { argument 37 return MemCmp16(s0, s1, count);
|
H A D | memcmp16.h | 35 extern "C" uint32_t __memcmp16(const uint16_t* s0, const uint16_t* s1, size_t count); 41 static inline int32_t MemCmp16(const uint16_t* s0, const uint16_t* s1, size_t count) { argument 42 for (size_t i = 0; i < count; i++) { 50 extern "C" int32_t memcmp16_generic_static(const uint16_t* s0, const uint16_t* s1, size_t count); 60 int32_t MemCmp16Testing(const uint16_t* s0, const uint16_t* s1, size_t count);
|
H A D | memcmp16_test.cc | 37 int32_t memcmp16_compare(const uint16_t* s0, const uint16_t* s1, size_t count) { argument 38 for (size_t i = 0; i < count; i++) {
|
H A D | stub_test.cc | 870 EXPECT_EQ(lock_after.ThinLockCount(), 0U); // Thin lock starts count at zero 875 // Check we're at lock count i 1038 // Unlock the remaining count times and then check it's unlocked. Then deallocate. 1042 size_t count = counts[index]; variable 1043 while (count > 0) { 1046 count--;
|
/art/runtime/ |
H A D | art_field.cc | 53 for (size_t i = 0, count = klass->NumInstanceFields(); i < count; ++i) { 66 for (size_t i = 0, count = klass->NumStaticFields(); i < count; ++i) {
|
H A D | art_method.cc | 543 auto count = proto_params != nullptr ? proto_params->Size() : 0u; local 545 if (param_len != count) { 549 for (size_t i = 0; i < count; ++i) {
|
H A D | barrier.cc | 25 Barrier::Barrier(int count) argument 26 : count_(count), 40 void Barrier::Init(Thread* self, int count) { argument 42 SetCountLocked(self, count); 49 // Increment the count. If it becomes zero after the increment 53 // Pass function is called by the last thread, the count will 82 void Barrier::SetCountLocked(Thread* self, int count) { argument 83 count_ = count; 84 if (count == 0) { 90 CHECK_EQ(count_, 0) << "Attempted to destroy barrier with non zero count"; [all...] |
H A D | barrier.h | 21 // latch API, it is possible to initialize the latch to a count of zero, repeatedly call 22 // Pass() or Wait(), and only then set the count using the Increment() method. Threads at 23 // a Wait() are only awoken if the count reaches zero AFTER the decrement is applied. 25 // decrementing the count, and thus nobody can spuriosly wake up on the initial zero. 38 explicit Barrier(int count); 41 // Pass through the barrier, decrement the count but do not block. 44 // Wait on the barrier, decrement the count. 53 // Increment the count by delta, wait on condition if count is non zero. 56 // Increment the count b [all...] |
H A D | barrier_test.cc | 89 CheckPassTask(Barrier* barrier, AtomicInteger* count, size_t subtasks) argument 91 count_(count), 116 AtomicInteger count(0); 120 thread_pool.AddTask(self, new CheckPassTask(&barrier, &count, num_sub_tasks)); 127 EXPECT_EQ(count.LoadRelaxed(), expected_total_tasks);
|
H A D | class_linker.cc | 4004 for (size_t i = 0, count = new_class->NumInstanceFields(); i < count; i++) { 4012 for (size_t i = 0, count = new_class->NumStaticFields(); i < count; i++) { 4186 *error_msg = StringPrintf("Virtual method count off: %zu vs %zu\n%s", dex_virtual_methods1, 4192 *error_msg = StringPrintf("Direct method count off: %zu vs %zu\n%s", dex_direct_methods1, 4198 *error_msg = StringPrintf("Static field count off: %zu vs %zu\n%s", dex_static_fields1, 4204 *error_msg = StringPrintf("Instance field count off: %zu vs %zu\n%s", dex_instance_fields1, 4391 size_t count = klass->NumVirtualMethods(); local 4392 if (!IsUint<16>(count)) { [all...] |
H A D | debugger.cc | 161 void SetByteCount(size_t count) { argument 162 byte_count_ = count; 1231 JDWP::JdwpError Dbg::OutputArray(JDWP::ObjectId array_id, int offset, int count, JDWP::ExpandBuf* pReply) { argument 1238 if (offset < 0 || count < 0 || offset > a->GetLength() || a->GetLength() - offset < count) { 1239 LOG(WARNING) << __FUNCTION__ << " access out of bounds: offset=" << offset << "; count=" << count; local 1244 expandBufAdd4BE(pReply, count); 1248 uint8_t* dst = expandBufAddSpace(pReply, count * width); 1251 for (int i = 0; i < count; 1289 SetArrayElements(JDWP::ObjectId array_id, int offset, int count, JDWP::Request* request) argument 1298 LOG(WARNING) << __FUNCTION__ << " access out of bounds: offset=" << offset << "; count=" << count; local 4964 uint16_t count = capped_count; local 4966 LOG(INFO) << "Tracked allocations, (head=" << alloc_record_head_ << " count=" << count << ")"; local 5097 uint16_t count = capped_count; local [all...] |
H A D | debugger.h | 312 static JDWP::JdwpError OutputArray(JDWP::ObjectId array_id, int offset, int count, 315 static JDWP::JdwpError SetArrayElements(JDWP::ObjectId array_id, int offset, int count,
|
H A D | dex_file.h | 255 static constexpr size_t GetListSize(size_t count) { argument 256 return GetHeaderSize() + sizeof(TypeItem) * count;
|
H A D | dex_file_verifier.cc | 173 bool DexFileVerifier::CheckListSize(const void* start, size_t count, size_t elem_size, argument 185 if (max_count < count) { 188 count, elem_size); 192 const uint8_t* range_end = range_start + count * elem_size; 205 // Check that the list is available. The first 4B are the count. 210 uint32_t count = *reinterpret_cast<const uint32_t*>(*ptr); local 211 if (count > 0) { 212 if (!CheckListSize(*ptr + 4, count, element_size, label)) { 217 *ptr += 4 + count * element_size; 297 uint32_t count local 1275 CheckIntraIdSection(size_t offset, uint32_t count, uint16_t type) argument 1323 CheckIntraDataSection(size_t offset, uint32_t count, uint16_t type) argument 1350 uint32_t count = map->size_; local 1824 uint32_t count = list->size_; local 1841 uint32_t count = set->size_; local 1969 CheckInterSectionIterate(size_t offset, uint32_t count, uint16_t type) argument 2065 uint32_t count = map->size_; local [all...] |
/art/runtime/base/ |
H A D | bit_vector.cc | 233 uint32_t count = 0; local 235 count += POPCOUNT(storage_[word]); 237 return count; 323 uint32_t count = 0u; local 325 count += POPCOUNT(storage[word]); 328 count += POPCOUNT(storage[word_end] & ~(0xffffffffu << partial_word_bits)); 330 return count;
|
H A D | hash_set_test.cc | 82 static constexpr size_t count = 1000; local 84 for (size_t i = 0; i < count; ++i) { 94 for (size_t i = 1; i < count; i += 2) { 101 for (size_t i = 1; i < count; i += 2) { 105 for (size_t i = 0; i < count; i += 2) { 115 static constexpr size_t count = 1000; local 117 for (size_t i = 0; i < count; ++i) { 127 for (size_t i = 0; i < count; ++i) { 137 for (size_t i = 0; i < count; ++i) { 145 static constexpr size_t count local [all...] |
H A D | hex_dump.cc | 71 size_t count = std::min(byte_count, 16 - gap); local 72 // CHECK_NE(count, 0U); 73 // CHECK_LE(count + gap, 16U); 82 for (i = gap ; i < count + gap; i++) { 104 byte_count -= count; 105 offset += count;
|
H A D | mutex.cc | 238 ++log[slot].count; 247 log[new_slot].count.StoreRelaxed(1); 269 uint32_t count = log[i].count.LoadRelaxed(); local 270 if (count > 0) { 273 most_common_blocked.Overwrite(blocked_tid, it->second + count); 275 most_common_blocked.Put(blocked_tid, count); 279 most_common_blocker.Overwrite(owner_tid, it->second + count); 281 most_common_blocker.Put(owner_tid, count); 393 CHECK(recursion_count_ == 1 || recursive_) << "Unexpected recursion count o [all...] |
H A D | mutex.h | 168 AtomicInteger count; member in struct:art::BaseMutex::ContentionLogEntry
|
/art/runtime/base/unix_file/ |
H A D | fd_file.cc | 181 static ssize_t ReadIgnoreOffset(int fd, void *buf, size_t count, off_t offset) { argument 183 return read(fd, buf, count);
|