/art/runtime/jdwp/ |
H A D | jdwp_expand_buf.cc | 38 uint8_t* storage; member in struct:art::JDWP::ExpandBuf 46 * Allocate a JdwpBuf and some initial storage. 50 newBuf->storage = reinterpret_cast<uint8_t*>(malloc(kInitialStorage)); 57 * Free a JdwpBuf and associated storage. 64 free(pBuf->storage); 72 return pBuf->storage; 95 uint8_t* newPtr = reinterpret_cast<uint8_t*>(realloc(pBuf->storage, pBuf->maxLen)); 100 pBuf->storage = newPtr; 110 gapStart = pBuf->storage + pBuf->curLen; 122 *(pBuf->storage [all...] |
/art/runtime/base/ |
H A D | arena_bit_vector.h | 38 void* storage = arena->template Alloc<ArenaBitVector>(kind); local 39 return new (storage) ArenaBitVector(arena, start_bits, expandable, kind);
|
H A D | arena_bit_vector.cc | 56 void* storage = arena->template Alloc<ArenaBitVectorAllocator>(kind); local 57 return new (storage) ArenaBitVectorAllocator(arena, kind);
|
H A D | bit_vector.h | 135 uint32_t* storage); 251 // Is bit set in storage. (No range check.) 252 static bool IsBitSet(const uint32_t* storage, uint32_t idx) { argument 253 return (storage[WordIndex(idx)] & BitMask(idx)) != 0; 256 // Number of bits set in range [0, end) in storage. (No range check.) 257 static uint32_t NumSetBits(const uint32_t* storage, uint32_t end); 286 // The index of the word within storage. 299 uint32_t* storage_; // The storage for the bit vector.
|
H A D | bit_vector.cc | 30 uint32_t* storage) 31 : storage_(storage), 87 // ie. we are comparing all storage cells that could have difference, if both vectors have cells above our_highest_index, 151 // Is the storage size smaller than src's? 157 // Paranoid: storage size should be big enough to hold this bit now. 185 // Is the storage size smaller than src's? 189 // Paranoid: storage size should be big enough to hold this bit now. 256 // We can set every storage element with -1. 281 // Return highest bit set in value plus bits from previous storage indexes. 303 // Now set until highest bit's storage 27 BitVector(bool expandable, Allocator* allocator, uint32_t storage_size, uint32_t* storage) argument 315 NumSetBits(const uint32_t* storage, uint32_t end) argument [all...] |
/art/runtime/ |
H A D | type_lookup_table.cc | 61 TypeLookupTable* TypeLookupTable::Create(const DexFile& dex_file, uint8_t* storage) { argument 64 ? new TypeLookupTable(dex_file, storage) 72 TypeLookupTable::TypeLookupTable(const DexFile& dex_file, uint8_t* storage) argument 75 entries_(storage != nullptr ? reinterpret_cast<Entry*>(storage) : new Entry[mask_ + 1]), 76 owns_entries_(storage == nullptr) { 78 DCHECK_ALIGNED(storage, alignof(Entry));
|
H A D | type_lookup_table.h | 63 static TypeLookupTable* Create(const DexFile& dex_file, uint8_t* storage = nullptr); 119 explicit TypeLookupTable(const DexFile& dex_file, uint8_t* storage);
|
H A D | handle_scope.h | 91 static HandleScope* Create(void* storage, HandleScope* link, uint32_t num_references) 93 return new (storage) HandleScope(link, num_references); 97 // Return backing storage used for references. 144 // Scoped handle storage of a fixed size that is usually stack allocated. 172 // Reference storage needs to be first as expected by the HandleScope layout.
|
H A D | dex_file.h | 1176 void CreateTypeLookupTable(uint8_t* storage = nullptr) const;
|
H A D | dex_file.cc | 731 void DexFile::CreateTypeLookupTable(uint8_t* storage) const { 732 lookup_table_.reset(TypeLookupTable::Create(*this, storage));
|
/art/compiler/ |
H A D | compiled_method.cc | 155 CompiledMethodStorage* storage = GetCompilerDriver()->GetCompiledMethodStorage(); local 156 storage->ReleaseLinkerPatches(patches_); 157 storage->ReleaseCFIInfo(cfi_info_); 158 storage->ReleaseVMapTable(vmap_table_); 159 storage->ReleaseSrcMappingTable(src_mapping_table_);
|
/art/compiler/utils/ |
H A D | intrusive_forward_list_test.cc | 77 std::vector<IFLTestValue> storage(ref.begin(), ref.end()); 78 IntrusiveForwardList<IFLTestValue> ifl(storage.begin(), storage.end()); 180 std::vector<IFLTestValue> storage(ref.begin(), ref.end()); 181 IntrusiveForwardList<IFLTestValue> ifl(storage.begin(), storage.end()); 235 std::vector<IFLTestValue> storage(ref.begin(), ref.end()); 236 IntrusiveForwardList<IFLTestValue> ifl(storage.begin(), storage.end()); 295 std::vector<IFLTestValue> storage(ref [all...] |
/art/runtime/lambda/ |
H A D | closure_builder.cc | 50 ShortyFieldTypeTraits::MaxType storage = 0; local 52 static_assert(sizeof(storage) >= sizeof(compressed_reference), 54 memcpy(&storage, &compressed_reference, sizeof(compressed_reference)); 56 values_.push_back(storage);
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space.cc | 139 // Returns the start of the storage. 145 uint8_t* storage = reinterpret_cast<uint8_t*>( local 147 if (LIKELY(storage != nullptr)) { 148 BlockHeader* header = reinterpret_cast<BlockHeader*>(storage); 150 storage += sizeof(BlockHeader); 153 return storage;
|
/art/compiler/driver/ |
H A D | compiled_method_storage_test.cc | 49 CompiledMethodStorage* storage = driver.GetCompiledMethodStorage(); local 51 ASSERT_TRUE(storage->DedupeEnabled()); // The default.
|
H A D | compiled_method_storage.cc | 37 void* storage = allocator.allocate(LengthPrefixedArray<T>::ComputeSize(array.size())); local 38 LengthPrefixedArray<T>* array_copy = new(storage) LengthPrefixedArray<T>(array.size());
|
/art/runtime/gc/ |
H A D | allocation_record.cc | 35 const char* AllocRecord::GetClassDescriptor(std::string* storage) const { 37 return klass_.IsNull() ? "null" : klass_.Read()->GetDescriptor(storage);
|
H A D | allocation_record.h | 181 const char* GetClassDescriptor(std::string* storage) const
|
/art/runtime/mirror/ |
H A D | class.cc | 812 const char* Class::GetDescriptor(std::string* storage) { argument 816 return GetArrayDescriptor(storage); 818 *storage = Runtime::Current()->GetClassLinker()->GetDescriptorForProxy(this); 819 return storage->c_str(); 827 const char* Class::GetArrayDescriptor(std::string* storage) { argument 830 *storage = "["; 831 *storage += elem_desc; 832 return storage->c_str();
|
H A D | class.h | 1163 // always create one the storage argument is populated and its internal c_str() returned. We do 1165 const char* GetDescriptor(std::string* storage) SHARED_REQUIRES(Locks::mutator_lock_); 1167 const char* GetArrayDescriptor(std::string* storage) SHARED_REQUIRES(Locks::mutator_lock_); 1418 // Total size of the Class instance; used when allocating storage on gc heap. 1439 // Total object size; used when allocating storage on gc heap.
|
/art/compiler/debug/ |
H A D | elf_debug_info_writer.h | 474 void* storage = allocator->Alloc(Thread::Current(), sizeof(LengthPrefixedArray<ArtMethod>)); local 475 methods_ptr = new (storage) LengthPrefixedArray<ArtMethod>(0);
|
/art/runtime/arch/mips64/ |
H A D | quick_entrypoints_mips64.S | 144 ld $t8, 56($sp) # Restore gp back to it's temp storage. 168 ld $t8, 56($sp) # Restore gp back to it's temp storage. 272 ld $t8, 184($sp) # Restore gp back to it's temp storage. 440 ld $t8, 80($sp) # Restore gp back to it's temp storage. 1631 * Entry from managed code when uninitialized static storage, this stub will run the class 1632 * initializer and deliver the exception on error. On success the static storage base is
|