/art/runtime/ |
H A D | mapping_table.h | 25 // A utility for processing the raw uleb128 encoded mapping table created by the quick compiler. 32 const uint8_t* table = encoded_table_; variable 33 if (table == nullptr) { 36 return DecodeUnsignedLeb128(&table); 41 const uint8_t* table = encoded_table_; variable 42 if (table == nullptr) { 45 uint32_t total_size = DecodeUnsignedLeb128(&table); 46 uint32_t pc_to_dex_size = DecodeUnsignedLeb128(&table); 52 const uint8_t* table = encoded_table_; local 53 if (table ! 68 DexToPcIterator(const MappingTable* table, uint32_t element) argument 124 const uint8_t* table = encoded_table_; variable 128 DecodeUnsignedLeb128(&table); // Total_size, unused. variable 135 const uint8_t* table = encoded_table_; local 145 PcToDexIterator(const MappingTable* table, uint32_t element) argument [all...] |
H A D | class_table_test.cc | 84 ClassTable table; local 85 EXPECT_EQ(table.NumZygoteClasses(class_loader.Get()), 0u); 86 EXPECT_EQ(table.NumNonZygoteClasses(class_loader.Get()), 0u); 88 // Add h_X to the class table. 89 table.Insert(h_X.Get()); 90 EXPECT_EQ(table.LookupByDescriptor(h_X.Get()), h_X.Get()); 91 EXPECT_EQ(table.Lookup(descriptor_x, ComputeModifiedUtf8Hash(descriptor_x)), h_X.Get()); 92 EXPECT_EQ(table.Lookup("NOT_THERE", ComputeModifiedUtf8Hash("NOT_THERE")), nullptr); 93 EXPECT_EQ(table.NumZygoteClasses(class_loader.Get()), 0u); 94 EXPECT_EQ(table [all...] |
H A D | type_lookup_table_test.cc | 36 std::unique_ptr<TypeLookupTable> table(TypeLookupTable::Create(*dex_file)); 37 ASSERT_NE(nullptr, table.get()); 38 ASSERT_NE(nullptr, table->RawData()); 39 ASSERT_EQ(32U, table->RawDataLength()); 45 std::unique_ptr<TypeLookupTable> table(TypeLookupTable::Create(*dex_file)); 46 ASSERT_NE(nullptr, table.get()); 50 uint32_t class_def_idx = table->Lookup(descriptor, hash);
|
H A D | bytecode_utils.h | 63 const uint16_t* table = reinterpret_cast<const uint16_t*>(&instruction) + table_offset; local 64 DCHECK_EQ(table[0], sparse_ ? static_cast<uint16_t>(Instruction::kSparseSwitchSignature) 66 num_entries_ = table[1]; 67 values_ = reinterpret_cast<const int32_t*>(&table[2]); 76 // In a sparse table, we have num_entries_ keys and num_entries_ values, in that order. 79 // In a packed table, we have the starting key and num_entries_ values. 96 // Index of the first value in the table. 99 // In a sparse table, we have num_entries_ keys and num_entries_ values, in that order. 102 // In a packed table, we have the starting key and num_entries_ values. 117 // Whether this is a sparse-switch table (o 135 DexSwitchTableIterator(const DexSwitchTable& table) argument [all...] |
H A D | image-inl.h | 74 auto* table = reinterpret_cast<ImtConflictTable*>(base + section.Offset() + pos); local 75 table->Visit([&visitor](const std::pair<ArtMethod*, ArtMethod*>& methods) { 78 pos += table->ComputeSize(pointer_size);
|
H A D | intern_table.cc | 57 os << "Intern table: " << StrongSize() << " strong; " << WeakSize() << " weak\n"; 86 // Note: we deliberately don't visit the weak_interns_ table and the immutable image roots. 229 // Check the strong table for a match. 251 // There is no match in the strong table, check the weak table. 255 // A match was found in the weak table. Promote to the strong table. 261 // No match in the strong table or the weak table. Insert into the strong / weak table [all...] |
H A D | indirect_reference_table.cc | 75 table_mem_map_.reset(MemMap::MapAnonymous("indirect ref table", nullptr, table_bytes, 78 *error_msg = "Unable to map memory for indirect ref table"; 130 // are usually near the end of the table (see the header, TODO: verify this assumption). To avoid 160 static size_t CountNullEntries(const IrtEntry* table, size_t from, size_t to) { argument 163 if (table[index].GetReference()->IsNull()) { 192 static inline void CheckHoleCount(IrtEntry* table, argument 197 size_t count = CountNullEntries(table, prev_state.top_index, cur_state.top_index); 207 std::unique_ptr<MemMap> new_map(MemMap::MapAnonymous("indirect ref table", 243 LOG(FATAL) << "JNI ERROR (app bug): " << kind_ << " table overflow " 252 LOG(FATAL) << "JNI ERROR (app bug): " << kind_ << " table overflo [all...] |
H A D | imt_conflict_table.h | 30 // Table to resolve IMT conflicts at runtime. The table is attached to 32 // The table contains a list of pairs of { interface_method, implementation_method } 43 // Build a new table copying `other` and adding the new entry formed of 149 // Compute the number of entries in this table. 158 // Compute the size in bytes taken by this table. 164 // Compute the size in bytes needed for copying the given `table` and add 166 static size_t ComputeSizeWithOneMoreEntry(ImtConflictTable* table, PointerSize pointer_size) { argument 167 return table->ComputeSize(pointer_size) + EntrySize(pointer_size);
|
H A D | indirect_reference_table.h | 45 // Maintain a table of indirect references. Used for local/global JNI references. 47 // The table contains object references, where the strong (local/global) references are part of the 54 // operate on segments of the table, where segments are pushed and popped as if on a stack. For 57 // table must be made in the current segment even if space is available in an earlier area. 62 // The GC must be able to scan the entire table quickly. 71 // - scanning the entire table straight through 73 // If there's more than one segment, we don't guarantee that the table will fill completely before 82 // We need a (potentially) large table index and a 2-bit reference type (global, local, weak 84 // serial number in the extra bits, and keep a copy of the serial number in the table. This requires 94 kHandleScopeOrInvalid = 0, // <<stack indirect reference table o [all...] |
H A D | intern_table_test.cc | 67 // Check if table indexes match on 64 and 32 bit machines. 69 // Otherwise cross compilation can cause a table to be filled on host using one indexing algorithm 71 // Thus the table may provide wrong data. 80 for (InternTable::Table::UnorderedSet& table : t.strong_interns_.tables_) { 82 ASSERT_TRUE(IsUint<32>(table.hashfn_(str)));
|
/art/tools/ahat/src/ |
H A D | style.css | 32 table td { 38 table th {
|
H A D | Doc.java | 51 * Start a table with the given columns. 55 * This should be followed by calls to the 'row' method to fill in the table 56 * contents and the 'end' method to end the table. 58 void table(Column... columns); method in interface:Doc 61 * Start a table with the following heading structure: 68 * This should be followed by calls to the 'row' method to fill in the table 69 * contents and the 'end' method to end the table. 71 void table(DocString description, List<Column> subcols, List<Column> cols); method in interface:Doc 74 * Add a row to the currently active table. 76 * currently active table [all...] |
H A D | SitePrinter.java | 35 HeapTable.TableConfig<Site> table = new HeapTable.TableConfig<Site>() { 61 HeapTable.render(doc, query, id, table, snapshot, path);
|
H A D | OverviewHandler.java | 75 HeapTable.TableConfig<TableElem> table = new HeapTable.TableConfig<TableElem>() { 88 HeapTable.render(doc, query, OVERVIEW_ID, table, mSnapshot, dummy);
|
H A D | SiteHandler.java | 57 HeapTable.TableConfig<Site> table = new HeapTable.TableConfig<Site>() { 79 HeapTable.render(doc, query, SITES_CALLED_ID, table, mSnapshot, children); 84 doc.table(
|
H A D | ObjectHandler.java | 129 doc.table( 154 doc.table( 210 doc.table(new Column("Object")); 222 doc.table(new Column("Object")); 257 HeapTable.TableConfig<PathElement> table = new HeapTable.TableConfig<PathElement>() { 292 HeapTable.render(doc, query, DOMINATOR_PATH_ID, table, mSnapshot, path);
|
H A D | HtmlDoc.java | 81 public void table(Column... columns) { method in class:HtmlDoc 87 ps.println("<table>"); 102 public void table(DocString description, List<Column> subcols, List<Column> cols) { method in class:HtmlDoc 118 ps.println("<table>"); 147 throw new IllegalStateException("table method must be called before row"); 171 ps.println("<table>"); 181 ps.println("</table>");
|
H A D | ObjectsHandler.java | 57 doc.table(
|
/art/compiler/dex/ |
H A D | dex_to_dex_decompiler_test.cc | 101 ArrayRef<const uint8_t> table; local 103 table = compiled_method->GetVmapTable(); 106 *it.GetMethodCodeItem(), table, /* decompile_return_instruction */ true); 113 ArrayRef<const uint8_t> table; local 115 table = compiled_method->GetVmapTable(); 118 *it.GetMethodCodeItem(), table, /* decompile_return_instruction */ true);
|
/art/runtime/gc/accounting/ |
H A D | mod_union_table_test.cc | 76 // inside the card cache mod-union table and causes the check 77 // ASSERT_FALSE(table->ContainsCardFor(reinterpret_cast<uintptr_t>(obj3))); 117 // A mod union table that only holds references to a specified target space. 153 name << "Mod union table: " << type; 194 std::unique_ptr<ModUnionTable> table(ModUnionTableFactory::Create( 196 ASSERT_TRUE(table.get() != nullptr); 218 table->ProcessCards(); 221 table->UpdateAndMarkReferences(&collector_before); 227 // obj1, obj2 cards should still be in mod union table since they have references to other 229 ASSERT_TRUE(table [all...] |
/art/compiler/driver/ |
H A D | compiled_method_storage.h | 58 const LengthPrefixedArray<uint8_t>* DeduplicateVMapTable(const ArrayRef<const uint8_t>& table); 59 void ReleaseVMapTable(const LengthPrefixedArray<uint8_t>* table);
|
H A D | compiled_method_storage.cc | 177 dedupe_vmap_table_("dedupe vmap table", 196 os << "\nVmap table dedupe: " << dedupe_vmap_table_.DumpStats(self); 220 const ArrayRef<const uint8_t>& table) { 221 return AllocateOrDeduplicateArray(table, &dedupe_vmap_table_); 224 void CompiledMethodStorage::ReleaseVMapTable(const LengthPrefixedArray<uint8_t>* table) { argument 225 ReleaseArrayIfNotDeduplicated(table); 219 DeduplicateVMapTable( const ArrayRef<const uint8_t>& table) argument
|
/art/compiler/optimizing/ |
H A D | block_builder.cc | 86 DexSwitchTable table(instruction, dex_pc); 87 for (DexSwitchTableIterator s_it(table); !s_it.Done(); s_it.Advance()) { 92 if (table.ShouldBuildDecisionTree() && !s_it.IsLast()) { 160 DexSwitchTable table(instruction, dex_pc); 161 for (DexSwitchTableIterator s_it(table); !s_it.Done(); s_it.Advance()) { 165 if (table.ShouldBuildDecisionTree() && !s_it.IsLast()) {
|
/art/runtime/gc/collector/ |
H A D | mark_compact.cc | 175 // Clear the whole card table since we cannot get any additional dirty cards during the 214 accounting::ModUnionTable* table = heap_->FindModUnionTableFromSpace(space); local 215 if (table != nullptr) { 220 table->UpdateAndMarkReferences(this); 350 accounting::ModUnionTable* table = heap_->FindModUnionTableFromSpace(space); local 351 if (table != nullptr) { 357 table->UpdateAndMarkReferences(this); 359 // No mod union table, so we need to scan the space using bitmap visit.
|
/art/runtime/arch/x86/ |
H A D | thread_x86.cc | 29 #include <architecture/i386/table.h>
|