/art/runtime/base/unix_file/ |
H A D | random_access_file_utils_test.cc | 27 StringFile src; local 31 src.Assign(content); 32 ASSERT_EQ(src.ToStringPiece(), content); 35 ASSERT_TRUE(CopyFile(src, &dst)); 36 ASSERT_EQ(src.ToStringPiece(), dst.ToStringPiece()); 40 FdFile src(-1, false); 42 ASSERT_FALSE(CopyFile(src, &dst)); 46 StringFile src; local 51 src.Assign("hello"); 53 ASSERT_FALSE(CopyFile(src, [all...] |
H A D | random_access_file_utils.cc | 23 bool CopyFile(const RandomAccessFile& src, RandomAccessFile* dst) { argument 24 // We don't call src->GetLength because some files (those in /proc, say) 29 while ((n = src.Read(&buf[0], buf.size(), offset)) > 0) {
|
H A D | mapped_file_test.cc | 40 StringFile src; local 41 src.Assign(kContent); 43 ASSERT_TRUE(CopyFile(src, &dst));
|
/art/runtime/ |
H A D | reflection-inl.h | 32 const JValue& src, JValue* dst) { 35 dst->SetJ(src.GetJ()); 46 dst->SetS(src.GetI()); 53 dst->SetI(src.GetI()); 60 dst->SetJ(src.GetI()); 67 dst->SetF(src.GetI()); 70 dst->SetF(src.GetJ()); 77 dst->SetD(src.GetI()); 80 dst->SetD(src.GetJ()); 83 dst->SetD(src 30 ConvertPrimitiveValue(const ThrowLocation* throw_location, bool unbox_for_result, Primitive::Type srcType, Primitive::Type dstType, const JValue& src, JValue* dst) argument [all...] |
H A D | dex_file_verifier_test.cc | 58 static inline byte* DecodeBase64(const char* src, size_t* dst_size) { argument 62 for (size_t i = 0; src[i] != '\0'; ++i) { 63 byte c = kBase64Map[src[i] & 0xFF];
|
H A D | dex_file_test.cc | 63 static inline byte* DecodeBase64(const char* src, size_t* dst_size) { argument 67 for (size_t i = 0; src[i] != '\0'; ++i) { 68 byte c = kBase64Map[src[i] & 0xFF];
|
/art/compiler/ |
H A D | buffered_output_stream.cc | 37 const uint8_t* src = reinterpret_cast<const uint8_t*>(buffer); local 38 memcpy(&buffer_[used_], src, byte_count);
|
/art/runtime/base/ |
H A D | bit_vector.h | 160 void Copy(const BitVector* src); 162 bool Union(const BitVector* src); 167 void Subtract(const BitVector* src); 169 bool Equal(const BitVector* src) { argument 170 return (storage_size_ == src->GetStorageSize()) && 171 (expandable_ == src->IsExpandable()) && 172 (memcmp(storage_, src->GetRawStorage(), storage_size_ * sizeof(uint32_t)) == 0); 179 bool SameBitsSet(const BitVector *src);
|
H A D | bit_vector.cc | 111 bool BitVector::SameBitsSet(const BitVector *src) { argument 113 int src_highest = src->GetHighestBitSet(); 134 return (memcmp(storage_, src->GetRawStorage(), our_highest_index * kWordBytes) == 0); 138 void BitVector::Intersect(const BitVector* src) { argument 139 uint32_t src_storage_size = src->storage_size_; 146 storage_[idx] &= src->GetRawStorageWord(idx); 150 // - Either src was larger than us: we don't care, all upper bits would thus be 0. 151 // - Either we are larger than src: we don't care, all upper bits would have been 0 too. 161 bool BitVector::Union(const BitVector* src) { argument 163 int highest_bit = src 244 Subtract(const BitVector *src) argument 353 Copy(const BitVector *src) argument [all...] |
/art/runtime/verifier/ |
H A D | register_line.h | 106 void CopyFromLine(const RegisterLine* src) { argument 107 DCHECK_EQ(num_regs_, src->num_regs_); 108 memcpy(&line_, &src->line_, num_regs_ * sizeof(uint16_t)); 109 monitors_ = src->monitors_; 110 reg_to_lock_depths_ = src->reg_to_lock_depths_; 295 void CopyRegToLockDepth(size_t dst, size_t src) { 296 auto it = reg_to_lock_depths_.find(src);
|
H A D | reg_type.cc | 757 bool RegType::IsAssignableFrom(RegType& src) { argument 758 return AssignableFrom(*this, src, false); 761 bool RegType::IsStrictlyAssignableFrom(RegType& src) { argument 762 return AssignableFrom(*this, src, true);
|
/art/compiler/dex/quick/ |
H A D | mir_to_lir-inl.h | 99 inline LIR* Mir2Lir::NewLIR2NoDest(int opcode, int src, int info) { argument 104 LIR* insn = RawLIR(current_dalvik_offset_, opcode, src, info);
|
H A D | local_optimizations.cc | 70 void Mir2Lir::ConvertMemOpIntoMove(LIR* orig_lir, RegStorage dest, RegStorage src) { argument 73 move_lir = OpRegCopyNoInsert(dest, src); 78 * will need to be re-checked (eg the new dest clobbers the src used in
|
/art/compiler/utils/ |
H A D | assembler.cc | 157 void Assembler::LoadFromThread32(ManagedRegister dest, ThreadOffset<4> src, size_t size) { argument 161 void Assembler::LoadFromThread64(ManagedRegister dest, ThreadOffset<8> src, size_t size) { argument
|
/art/runtime/mirror/ |
H A D | array-inl.h | 244 inline void PrimitiveArray<T>::Memmove(int32_t dst_pos, PrimitiveArray<T>* src, int32_t src_pos, argument 252 DCHECK(src != nullptr); 255 DCHECK_LT(src_pos, src->GetLength()); 256 DCHECK_LE(src_pos, src->GetLength() - count); 260 if (LIKELY(src != this)) { 262 Memcpy(dst_pos, src, src_pos, count); 266 const void* src_raw = src->GetRawData(sizeof(T), src_pos); 304 inline void PrimitiveArray<T>::Memcpy(int32_t dst_pos, PrimitiveArray<T>* src, int32_t src_pos, argument 312 DCHECK(src != nullptr); 315 DCHECK_LT(src_pos, src [all...] |
H A D | object.cc | 70 Object* Object::CopyObject(Thread* self, mirror::Object* dest, mirror::Object* src, argument 74 byte* src_bytes = reinterpret_cast<byte*>(src); 83 src->VisitReferences<true>(visitor, visitor); 87 Class* c = src->GetClass();
|
H A D | object_array-inl.h | 118 inline void ObjectArray<T>::AssignableMemmove(int32_t dst_pos, ObjectArray<T>* src, argument 123 src->GetWithoutChecks(src_pos + i); 129 IntArray* srcAsIntArray = reinterpret_cast<IntArray*>(src); 132 const bool copy_forward = (src != this) || (dst_pos < src_pos) || (dst_pos - src_pos >= count); 137 Object* obj = src->GetWithoutChecks(src_pos + i); 144 Object* obj = src->GetWithoutChecks(src_pos + i); 161 inline void ObjectArray<T>::AssignableMemcpy(int32_t dst_pos, ObjectArray<T>* src, argument 166 src->GetWithoutChecks(src_pos + i); 172 IntArray* srcAsIntArray = reinterpret_cast<IntArray*>(src); 177 T* obj = src 193 AssignableCheckingMemcpy(int32_t dst_pos, ObjectArray<T>* src, int32_t src_pos, int32_t count, bool throw_exception) argument [all...] |
H A D | class-inl.h | 208 // Determine whether "this" is assignable from "src", where both of these 227 inline bool Class::IsArrayAssignableFromArray(Class* src) { argument 229 DCHECK(src->IsArrayClass()) << PrettyClass(src); 230 return GetComponentType()->IsAssignableFrom(src->GetComponentType()); 233 inline bool Class::IsAssignableFromArray(Class* src) { argument 235 DCHECK(src->IsArrayClass()) << PrettyClass(src); 238 // src's super should be java_lang_Object, since it is an array. 239 Class* java_lang_Object = src [all...] |
/art/compiler/dex/quick/x86/ |
H A D | utility_x86.cc | 305 int src = r_src.IsPair() ? r_src.GetLowReg() : r_src.GetReg(); local 353 return NewLIR3(opcode, r_base.GetReg(), offset, src);
|
/art/compiler/utils/arm64/ |
H A D | assembler_arm64.cc | 123 Arm64ManagedRegister src = m_src.AsArm64(); local 124 if (src.IsNoRegister()) { 126 } else if (src.IsWRegister()) { 128 StoreWToOffset(kStoreWord, src.AsWRegister(), SP, offs.Int32Value()); 129 } else if (src.IsCoreRegister()) { 131 StoreToOffset(src.AsCoreRegister(), SP, offs.Int32Value()); 132 } else if (src.IsSRegister()) { 133 StoreSToOffset(src.AsSRegister(), SP, offs.Int32Value()); 135 CHECK(src.IsDRegister()) << src; 141 Arm64ManagedRegister src = m_src.AsArm64(); local 148 Arm64ManagedRegister src = m_src.AsArm64(); local 279 Load(ManagedRegister m_dst, FrameOffset src, size_t size) argument 283 LoadFromThread64(ManagedRegister m_dst, ThreadOffset<8> src, size_t size) argument 321 Arm64ManagedRegister src = m_src.AsArm64(); local 366 CopyRef(FrameOffset dest, FrameOffset src, ManagedRegister m_scratch) argument 376 Copy(FrameOffset dest, FrameOffset src, ManagedRegister m_scratch, size_t size) argument 411 Copy(ManagedRegister m_dest_base, Offset dest_offs, FrameOffset src, ManagedRegister m_scratch, size_t size) argument 439 Arm64ManagedRegister src = m_src.AsArm64(); local [all...] |
/art/runtime/gc/collector/ |
H A D | semi_space.cc | 430 static inline size_t CopyAvoidingDirtyingPages(void* dest, const void* src, size_t size) { argument 436 memcpy(dest, src, size); 447 const byte* byte_src = reinterpret_cast<const byte*>(src); 451 memcpy(dest, src, page_remain);
|
/art/runtime/interpreter/ |
H A D | interpreter_common.cc | 884 ObjectArray<Object>* src = shadow_frame->GetVRegReference(arg_offset)->AsObjectArray<Object>(); local 887 dst->Set(dstPos + i, src->Get(srcPos + i)); 890 CharArray* src = shadow_frame->GetVRegReference(arg_offset)->AsCharArray(); local 893 dst->Set(dstPos + i, src->Get(srcPos + i)); 896 IntArray* src = shadow_frame->GetVRegReference(arg_offset)->AsIntArray(); local 899 dst->Set(dstPos + i, src->Get(srcPos + i));
|
/art/compiler/optimizing/ |
H A D | code_generator_x86.cc | 1491 void ParallelMoveResolverX86::MoveMemoryToMemory(int dst, int src) { argument 1495 __ movl(static_cast<Register>(ensure_scratch.GetRegister()), Address(ESP, src + stack_offset));
|
/art/compiler/sea_ir/ir/ |
H A D | sea.cc | 56 void SeaGraph::AddEdge(Region* src, Region* dst) const { argument 57 src->AddSuccessor(dst); 58 dst->AddPredecessor(src);
|
/art/compiler/dex/ |
H A D | local_value_numbering.cc | 309 const ScopedArenaSafeMap<K, AliasingValues>& src) { 312 for (const auto& entry : src) { 504 void LocalValueNumbering::CopyLiveSregValues(SregValueMap* dest, const SregValueMap& src) { argument 508 for (const auto& entry : src) { 308 CopyAliasingValuesMap(ScopedArenaSafeMap<K, AliasingValues>* dest, const ScopedArenaSafeMap<K, AliasingValues>& src) argument
|