/art/runtime/jdwp/ |
H A D | jdwp_request.cc | 90 FieldId id = Read8BE(); local 91 VLOG(jdwp) << " field id " << DescribeField(id); 92 return id; 96 MethodId id = Read8BE(); local 97 VLOG(jdwp) << " method id " << DescribeMethod(id); 98 return id; 102 ObjectId id = Read8BE(); local 103 VLOG(jdwp) << StringPrintf(" %s id 124 RefTypeId id = Read8BE(); local 130 FrameId id = Read8BE(); local [all...] |
H A D | object_registry.h | 47 // The corresponding id, so we only need one map lookup in Add. 48 JDWP::ObjectId id; member in struct:art::ObjectRegistryEntry 86 template<typename T> T Get(JDWP::ObjectId id, JDWP::JdwpError* error) 88 if (id == 0) { 92 return down_cast<T>(InternalGet(id, error)); 97 void DisableCollection(JDWP::ObjectId id) 100 void EnableCollection(JDWP::ObjectId id) 103 bool IsCollected(JDWP::ObjectId id) 106 void DisposeObject(JDWP::ObjectId id, uint32_t reference_count) 111 jobject GetJObject(JDWP::ObjectId id) SHARED_LOCKS_REQUIRE [all...] |
H A D | object_registry.cc | 30 << ",id=" << rhs.id << "]"; 100 entry->id = 0; 112 entry->id = next_id_++; 114 id_to_entry_.Put(entry->id, entry); 118 return entry->id; 170 mirror::Object* ObjectRegistry::InternalGet(JDWP::ObjectId id, JDWP::JdwpError* error) { argument 173 auto it = id_to_entry_.find(id); 183 jobject ObjectRegistry::GetJObject(JDWP::ObjectId id) { argument 184 if (id 195 DisableCollection(JDWP::ObjectId id) argument 203 EnableCollection(JDWP::ObjectId id) argument 233 IsCollected(JDWP::ObjectId id) argument 247 DisposeObject(JDWP::ObjectId id, uint32_t reference_count) argument [all...] |
H A D | jdwp.h | 70 static inline void expandBufAddFieldId(ExpandBuf* pReply, FieldId id) { expandBufAdd8BE(pReply, id); } argument 71 static inline void expandBufAddMethodId(ExpandBuf* pReply, MethodId id) { expandBufAdd8BE(pReply, id); } argument 72 static inline void expandBufAddObjectId(ExpandBuf* pReply, ObjectId id) { expandBufAdd8BE(pReply, id); } argument 73 static inline void expandBufAddRefTypeId(ExpandBuf* pReply, RefTypeId id) { expandBufAdd8BE(pReply, id); } argument 74 static inline void expandBufAddFrameId(ExpandBuf* pReply, FrameId id) { expandBufAdd8BE(pReply, id); } argument [all...] |
/art/compiler/optimizing/ |
H A D | side_effects_analysis.cc | 84 int id = info->GetHeader()->GetBlockId(); local 85 loop_effects_.Put(id, loop_effects_.Get(id).Union(effects));
|
H A D | dead_code_elimination.cc | 77 int id = block->GetBlockId(); local 78 if (live_blocks.IsBitSet(id)) { 79 if (affected_loops.IsBitSet(id)) {
|
H A D | graph_checker.cc | 138 AddError(StringPrintf("Instruction id %d is duplicate in graph.", 314 int id = loop_header->GetBlockId(); local 322 id)); 331 id, 338 id)); 345 id)); 357 id)); 364 id, 375 id, 380 id)); [all...] |
H A D | ssa_test.cc | 66 int id = 0; local 70 it.Current()->SetId(id++); 73 it.Current()->SetId(id++);
|
H A D | locations.h | 430 bool ContainsCoreRegister(uint32_t id) { 431 return Contains(core_registers_, id); 434 bool ContainsFloatingPointRegister(uint32_t id) { 435 return Contains(floating_point_registers_, id);
|
H A D | code_generator.cc | 756 int id = location.reg(); local 757 if (slow_path != nullptr && slow_path->IsCoreRegisterSaved(id)) { 758 uint32_t offset = slow_path->GetStackOffsetOfCoreRegister(id); 766 stack_map_stream_.AddDexRegisterEntry(i, DexRegisterLocation::Kind::kInRegister, id); 768 stack_map_stream_.AddDexRegisterEntry(++i, DexRegisterLocation::Kind::kInRegister, id); 776 int id = location.reg(); local 777 if (slow_path != nullptr && slow_path->IsFpuRegisterSaved(id)) { 778 uint32_t offset = slow_path->GetStackOffsetOfFpuRegister(id); 786 stack_map_stream_.AddDexRegisterEntry(i, DexRegisterLocation::Kind::kInFpuRegister, id); 789 ++i, DexRegisterLocation::Kind::kInFpuRegister, id); [all...] |
/art/compiler/dex/ |
H A D | global_value_numbering.cc | 71 work_lvn_.reset(new (allocator) LocalValueNumbering(this, bb->id, allocator)); 93 loop_head_idx != mir_graph_->GetTopologicalSortOrderIndexes()[bb->id]; 127 DCHECK_EQ(bb->id, work_lvn_->Id()); 133 change = (lvns_[bb->id] == nullptr) || !lvns_[bb->id]->Equals(*work_lvn_); 138 std::unique_ptr<const LocalValueNumbering> old_lvn(lvns_[bb->id]); 139 lvns_[bb->id] = work_lvn_.release(); 142 DCHECK(lvns_[bb->id] != nullptr); 143 DCHECK(lvns_[bb->id]->Equals(*work_lvn_));
|
H A D | mir_graph.cc | 224 bb_taken->ErasePredecessor(orig_block->id); 225 bb_taken->predecessors.push_back(bottom_block->id); 230 orig_block->fall_through = bottom_block->id; 231 bottom_block->predecessors.push_back(orig_block->id); 234 bb_fall_through->ErasePredecessor(orig_block->id); 235 bb_fall_through->predecessors.push_back(bottom_block->id); 247 bb->ErasePredecessor(orig_block->id); 248 bb->predecessors.push_back(bottom_block->id); 273 p->bb = bottom_block->id; 277 p->bb = bottom_block->id; 2474 BasicBlockId id = static_cast<BasicBlockId>(block_list_.size()); local [all...] |
H A D | ssa_transformation.cc | 63 /* Enqueue the pre_order block id */ 64 if (block->id != NullBasicBlockId) { 65 dfs_order_.push_back(block->id); 84 if (curr->id != NullBasicBlockId) { 85 dfs_post_order_.push_back(curr->id); 130 temp_.ssa.def_block_matrix[idx]->SetBit(bb->id); 167 temp_.ssa.def_block_matrix[in_reg]->SetBit(GetEntryBlock()->id); 197 if (curr_bb->id != NullBasicBlockId) { 198 dom_post_order_traversal_.push_back(curr_bb->id); 211 if (succ_bb->i_dom != dom_bb->id [all...] |
H A D | post_opt_passes.cc | 74 child->predecessors.push_back(bb->id);
|
H A D | mir_optimization.cc | 484 local_valnum.reset(new (allocator.get()) LocalValueNumbering(global_valnum.get(), bb->id, 530 successor_to_unlink->ErasePredecessor(bb->id); 703 // to update all Phi inputs correctly with UpdatePredecessor(ft->id, bb->id) 706 ft_ft->UpdatePredecessor(ft->id, bb->id); 708 ft->ErasePredecessor(bb->id); 710 bb->fall_through = ft_ft->id; 849 bb_next->ErasePredecessor(bb->id); 856 DCHECK_EQ(bb_taken->predecessors[0], bb->id); [all...] |
H A D | type_inference.cc | 181 split_s_reg_data.ending_mod_s_reg[def_bb->id] = s_reg; 182 split_s_reg_data.def_phi_blocks_->SetBit(def_bb->id); 233 BasicBlockId bb_id = bb->id; 241 int32_t mod_s_reg = data.ending_mod_s_reg[pred_bb->id]; 281 entry.second.current_mod_s_reg = entry.second.starting_mod_s_reg[bb->id]; 288 DCHECK_EQ(entry.second.current_mod_s_reg, entry.second.starting_mod_s_reg[bb->id]); 289 if (entry.second.def_phi_blocks_->IsBitSet(bb->id)) { 358 DCHECK_LT(bb->id, indexes.size()); 359 size_t best_idx = indexes[bb->id]; 369 DCHECK_LT(best_idx, indexes[bb->id]); [all...] |
/art/runtime/ |
H A D | monitor_pool.cc | 74 // Eagerly compute id. 106 // Pull out the id which was preinitialized. 107 MonitorId id = mon_uninitialized->monitor_id_; local 110 Monitor* monitor = new(mon_uninitialized) Monitor(self, owner, obj, hash_code, id); 119 // Keep the monitor id. Don't trust it's not cleared. 120 MonitorId id = monitor->monitor_id_; local 130 // Rewrite monitor id. 131 monitor->monitor_id_ = id;
|
H A D | monitor_pool.h | 34 // monitor id loses the alignment bits of the Monitor*. 149 static size_t MonitorIdToOffset(MonitorId id) { argument 150 return id << 3; 164 // Chunk size that is referenced in the id. We can collapse this to the actually used storage
|
H A D | thread_list.h | 83 // Suspend a thread using its thread id, typically used by lock/monitor inflation. Returns the 84 // thread on success else null. The thread id is used to identify the thread to avoid races with 92 // Find an already suspended thread (or self) by its id. 154 void ReleaseThreadId(Thread* self, uint32_t id) LOCKS_EXCLUDED(Locks::allocated_thread_ids_lock_);
|
H A D | debugger.h | 279 static std::string GetClassName(JDWP::RefTypeId id) 283 static JDWP::JdwpError GetClassObject(JDWP::RefTypeId id, JDWP::ObjectId* class_object_id) 285 static JDWP::JdwpError GetSuperclass(JDWP::RefTypeId id, JDWP::RefTypeId* superclass_id) 287 static JDWP::JdwpError GetClassLoader(JDWP::RefTypeId id, JDWP::ExpandBuf* pReply) 289 static JDWP::JdwpError GetModifiers(JDWP::RefTypeId id, JDWP::ExpandBuf* pReply) 400 static void OutputVariableTable(JDWP::RefTypeId ref_type_id, JDWP::MethodId id, bool with_generic,
|
/art/runtime/verifier/ |
H A D | reg_type_cache-inl.h | 30 inline const art::verifier::RegType& RegTypeCache::GetFromId(uint16_t id) const { 31 DCHECK_LT(id, entries_.size()); 32 const RegType* result = entries_[id];
|
/art/test/004-ThreadStress/src/ |
H A D | Main.java | 428 int id = threadStress.id; 429 System.out.println("Starting worker for " + id); 431 Thread thread = new Thread(ts, "Worker thread " + id); 437 System.out.println("Thread exited for " + id + " with " 472 private final int id; field in class:Main 476 private Main(Object lock, int id, Operation[] operations) { argument 478 this.id = id; 485 System.out.println("Starting ThreadStress " + id); [all...] |
/art/runtime/gc/space/ |
H A D | large_object_space_test.cc | 118 AllocRaceTask(size_t id, size_t iterations, size_t size, LargeObjectSpace* los) : argument 119 id_(id), iterations_(iterations), size_(size), los_(los) {} 127 NanoSleep((id_ + 3) * 1000); // (3+id) mu s
|
/art/runtime/native/ |
H A D | dalvik_system_VMDebug.cc | 253 static void VMDebug_infopoint(JNIEnv*, jclass, jint id) { argument 254 LOG(INFO) << "VMDebug infopoint " << id << " hit"; 391 static bool SetRuntimeStatValue(JNIEnv* env, jobjectArray result, VMDebugRuntimeStatId id, argument 397 env->SetObjectArrayElement(result, static_cast<jint>(id), jvalue.get());
|
/art/runtime/hprof/ |
H A D | hprof.cc | 536 const size_t id = p.second; local 544 __ AddU4(id); 592 HprofStringId id = next_string_id_++; local 593 strings_.Put(string, id); 594 return id;
|