/art/runtime/native/ |
H A D | libcore_util_CharsetUtils.cc | 110 static void CharsetUtils_asciiBytesToChars(JNIEnv* env, jclass, jbyteArray javaBytes, jint offset, argument 121 const jbyte* src = &bytes[offset]; 131 jint offset, jint length, jcharArray javaChars) { 141 const jbyte* src = &bytes[offset]; 153 static jbyteArray charsToBytes(JNIEnv* env, jstring java_string, jint offset, jint length, argument 168 const jchar* src = &(string->GetValue()[offset]); 181 static jbyteArray CharsetUtils_toAsciiBytes(JNIEnv* env, jclass, jstring java_string, jint offset, argument 183 return charsToBytes(env, java_string, offset, length, 0x7f); 187 jint offset, jint length) { 188 return charsToBytes(env, java_string, offset, lengt 130 CharsetUtils_isoLatin1BytesToChars(JNIEnv* env, jclass, jbyteArray javaBytes, jint offset, jint length, jcharArray javaChars) argument 186 CharsetUtils_toIsoLatin1Bytes(JNIEnv* env, jclass, jstring java_string, jint offset, jint length) argument 191 CharsetUtils_toUtf8Bytes(JNIEnv* env, jclass, jstring java_string, jint offset, jint length) argument [all...] |
H A D | java_lang_reflect_Field.cc | 65 MemberOffset offset(f->GetOffset()); 69 value->SetZ(is_volatile ? o->GetFieldBooleanVolatile(offset) : o->GetFieldBoolean(offset)); 72 value->SetB(is_volatile ? o->GetFieldByteVolatile(offset) : o->GetFieldByte(offset)); 75 value->SetC(is_volatile ? o->GetFieldCharVolatile(offset) : o->GetFieldChar(offset)); 79 value->SetI(is_volatile ? o->GetField32Volatile(offset) : o->GetField32(offset)); 83 value->SetJ(is_volatile ? o->GetField64Volatile(offset) [all...] |
/art/compiler/linker/ |
H A D | buffered_output_stream.cc | 63 off_t BufferedOutputStream::Seek(off_t offset, Whence whence) { argument 67 return out_->Seek(offset, whence);
|
H A D | file_output_stream.h | 34 off_t Seek(off_t offset, Whence whence) OVERRIDE;
|
H A D | relative_patcher.h | 43 * Find the offset of the target method of a relative call if known. 47 * has an offset assigned and, if so, what's that offset. If the offset has not yet been 53 * if found, the second element specifies the offset. 85 // Reserve space for thunks if needed before a method, return adjusted offset. 86 virtual uint32_t ReserveSpace(uint32_t offset, 90 // Reserve space for thunks if needed after the last method, return adjusted offset. 94 virtual uint32_t ReserveSpaceEnd(uint32_t offset) = 0; 96 // Write relative call thunks if needed, return adjusted offset [all...] |
H A D | buffered_output_stream.h | 36 off_t Seek(off_t offset, Whence whence) OVERRIDE;
|
/art/disassembler/ |
H A D | disassembler.cc | 53 size_t offset = begin - disassembler_options_->base_address_; local 54 return StringPrintf("0x%08zx", offset);
|
H A D | disassembler_mips.cc | 435 case 'B': // Branch offset. 437 int32_t offset = static_cast<int16_t>(instruction & 0xffff); local 438 offset <<= 2; 439 offset += 4; // Delay slot. 440 args << FormatInstructionPointer(instr_ptr + offset) 441 << StringPrintf(" ; %+d", offset); 444 case 'b': // 21-bit branch offset. 446 int32_t offset = (instruction & 0x1fffff) - ((instruction & 0x100000) << 1); local 447 offset <<= 2; 448 offset 492 int32_t offset = static_cast<int16_t>(instruction) >> 7; local 498 int32_t offset = static_cast<int16_t>(instruction & 0xffff); local 512 int32_t offset = (instruction & 0x3ffffff) - ((instruction & 0x2000000) << 1); local 521 int32_t offset = (instruction & 0x7ffff) - ((instruction & 0x40000) << 1); local 554 uint32_t offset = (last_instr_ << 16) | (instruction & 0xFFFF); local [all...] |
/art/runtime/gc/accounting/ |
H A D | space_bitmap_test.cc | 127 size_t offset = RoundDown(r.next() % heap_capacity, kAlignment); variable 131 space_bitmap->Set(reinterpret_cast<mirror::Object*>(heap_begin + offset)); 133 space_bitmap->Clear(reinterpret_cast<mirror::Object*>(heap_begin + offset)); 141 size_t offset = RoundDown(r.next() % heap_capacity, kAlignment); variable 142 size_t remain = heap_capacity - offset; 143 size_t end = offset + RoundDown(r.next() % (remain + 1), kAlignment); 145 space_bitmap->VisitMarkedRange(reinterpret_cast<uintptr_t>(heap_begin) + offset, 149 for (uintptr_t k = offset; k < end; k += kAlignment) {
|
H A D | card_table.cc | 78 // kCardDirty, compute a offset value to make this the case 79 size_t offset = 0; local 85 offset = delta + (delta < 0 ? 0x100 : 0); 86 biased_begin += offset; 89 return new CardTable(mem_map.release(), biased_begin, offset); 92 CardTable::CardTable(MemMap* mem_map, uint8_t* biased_begin, size_t offset) argument 93 : mem_map_(mem_map), biased_begin_(biased_begin), offset_(offset) {
|
H A D | space_bitmap-inl.h | 36 const uintptr_t offset = addr - heap_begin_; local 37 const size_t index = OffsetToIndex(offset); 38 const uintptr_t mask = OffsetToMask(offset); 60 const uintptr_t offset = addr - heap_begin_; local 61 return (bitmap_begin_[OffsetToIndex(offset)] & OffsetToMask(offset)) != 0; 163 const uintptr_t offset = addr - heap_begin_; local 164 const size_t index = OffsetToIndex(offset); 165 const uintptr_t mask = OffsetToMask(offset);
|
/art/runtime/interpreter/mterp/mips/ |
H A D | op_goto.S | 2 * Unconditional branch, 8-bit offset. 4 * The branch distance is a signed code-unit offset, which we need to 5 * double to get a byte offset. 15 JAL(MterpProfileBranch) # (self, shadow_frame, offset) 16 bnez v0, MterpOnStackReplacement # Note: offset must be in rINST 17 addu a2, rINST, rINST # a2 <- byte offset 29 addu a2, rINST, rINST # a2 <- byte offset
|
H A D | op_goto_32.S | 2 * Unconditional branch, 32-bit offset. 4 * The branch distance is a signed code-unit offset, which we need to 5 * double to get a byte offset. 20 JAL(MterpProfileBranch) # (self, shadow_frame, offset) 21 bnez v0, MterpOnStackReplacement # Note: offset must be in rINST 22 addu a1, rINST, rINST # a1 <- byte offset 35 addu a1, rINST, rINST # a1 <- byte offset
|
H A D | op_iput_object_quick.S | 2 # op vA, vB, offset /* CCCC */
|
H A D | op_iput_quick.S | 3 # op vA, vB, offset /* CCCC */ 6 FETCH(a1, 1) # a1 <- field byte offset
|
H A D | op_packed_switch.S | 20 JAL($func) # a0 <- code-unit branch offset 26 JAL(MterpProfileBranch) # (self, shadow_frame, offset) 27 bnez v0, MterpOnStackReplacement # Note: offset must be in rINST 28 addu a1, rINST, rINST # a1 <- byte offset 41 JAL($func) # a0 <- code-unit branch offset 43 addu a1, rINST, rINST # a1 <- byte offset
|
/art/tools/dexfuzz/src/dexfuzz/rawdex/ |
H A D | DexRandomAccessFile.java | 299 long offset = getFilePointer(); 301 if ((offset & mask) != 0) { 302 int extra = alignment - (int) (offset & mask); 303 seek(offset + extra); 312 long offset = getFilePointer(); 314 if ((offset & mask) != 0) { 315 offset &= (~mask); 316 seek(offset);
|
/art/compiler/linker/arm/ |
H A D | relative_patcher_thumb2.h | 41 void SetInsn32(std::vector<uint8_t>* code, uint32_t offset, uint32_t value); 42 static uint32_t GetInsn32(ArrayRef<const uint8_t> code, uint32_t offset); 45 static uint32_t GetInsn32(Vector* code, uint32_t offset);
|
/art/runtime/mirror/ |
H A D | string-inl.h | 60 SetStringCountAndBytesVisitor(int32_t count, Handle<ByteArray> src_array, int32_t offset, argument 62 : count_(count), src_array_(src_array), offset_(offset), high_byte_(high_byte) { 88 int32_t offset) : 89 count_(count), src_array_(src_array), offset_(offset) { 111 int32_t offset) : 112 count_(count), src_string_(src_string), offset_(offset) { 192 Handle<ByteArray> array, int32_t offset, 194 SetStringCountAndBytesVisitor visitor(byte_length, array, offset, high_byte << 8); 201 Handle<CharArray> array, int32_t offset, 205 SetStringCountAndValueVisitorFromCharArray visitor(count, array, offset); 87 SetStringCountAndValueVisitorFromCharArray(int32_t count, Handle<CharArray> src_array, int32_t offset) argument 110 SetStringCountAndValueVisitorFromString(int32_t count, Handle<String> src_string, int32_t offset) argument 191 AllocFromByteArray(Thread* self, int32_t byte_length, Handle<ByteArray> array, int32_t offset, int32_t high_byte, gc::AllocatorType allocator_type) argument 200 AllocFromCharArray(Thread* self, int32_t count, Handle<CharArray> array, int32_t offset, gc::AllocatorType allocator_type) argument 211 AllocFromString(Thread* self, int32_t string_length, Handle<String> string, int32_t offset, gc::AllocatorType allocator_type) argument [all...] |
H A D | object.cc | 49 void operator()(Object* obj, MemberOffset offset, bool /* is_static */) const 52 Object* ref = obj->GetFieldObject<Object>(offset); 55 dest_obj_->SetFieldObjectWithoutWriteBarrier<false, false>(offset, ref); 81 size_t offset = sizeof(Object); local 82 memcpy(dst_bytes + offset, src_bytes + offset, num_bytes - offset); 243 << " of type " << PrettyDescriptor(c) << " at offset " << field_offset; 247 ArtField* Object::FindFieldByOffset(MemberOffset offset) { argument 248 return IsClass() ? ArtField::FindStaticFieldWithOffset(AsClass(), offset [all...] |
/art/runtime/ |
H A D | oat.h | 72 void SetInterpreterToInterpreterBridgeOffset(uint32_t offset); 75 void SetInterpreterToCompiledCodeBridgeOffset(uint32_t offset); 79 void SetJniDlsymLookupOffset(uint32_t offset); 83 void SetQuickGenericJniTrampolineOffset(uint32_t offset); 86 void SetQuickResolutionTrampolineOffset(uint32_t offset); 89 void SetQuickImtConflictTrampolineOffset(uint32_t offset); 92 void SetQuickToInterpreterBridgeOffset(uint32_t offset);
|
/art/test/004-checker-UnsafeTest18/src/ |
H A D | Main.java | 53 private static int set32(Object o, long offset, int newValue) { argument 54 return unsafe.getAndSetInt(o, offset, newValue); 60 private static long set64(Object o, long offset, long newValue) { argument 61 return unsafe.getAndSetLong(o, offset, newValue); 67 private static Object setObj(Object o, long offset, Object newValue) { argument 68 return unsafe.getAndSetObject(o, offset, newValue); 78 private static int add32(Object o, long offset, int delta) { argument 79 return unsafe.getAndAddInt(o, offset, delta); 85 private static long add64(Object o, long offset, long delta) { argument 86 return unsafe.getAndAddLong(o, offset, delt [all...] |
/art/compiler/utils/arm64/ |
H A D | assembler_arm64.h | 108 void SpillRegisters(vixl::CPURegList registers, int offset); 109 void UnspillRegisters(vixl::CPURegList registers, int offset); 195 // Call to address held at [base+offset]. 196 void Call(ManagedRegister base, Offset offset, ManagedRegister scratch) OVERRIDE; 197 void Call(FrameOffset base, Offset offset, ManagedRegister scratch) OVERRIDE; 198 void CallFromThread64(ThreadOffset<8> offset, ManagedRegister scratch) OVERRIDE; 258 XRegister base, int32_t offset); 259 void StoreToOffset(XRegister source, XRegister base, int32_t offset); 260 void StoreSToOffset(SRegister source, XRegister base, int32_t offset); 261 void StoreDToOffset(DRegister source, XRegister base, int32_t offset); [all...] |
H A D | assembler_arm64.cc | 59 void Arm64Assembler::GetCurrentThread(FrameOffset offset, ManagedRegister /* scratch */) { argument 60 StoreToOffset(TR, SP, offset.Int32Value()); 98 XRegister base, int32_t offset) { 101 ___ Strb(reg_w(source), MEM_OP(reg_x(base), offset)); local 104 ___ Strh(reg_w(source), MEM_OP(reg_x(base), offset)); local 107 ___ Str(reg_w(source), MEM_OP(reg_x(base), offset)); local 114 void Arm64Assembler::StoreToOffset(XRegister source, XRegister base, int32_t offset) { argument 116 ___ Str(reg_x(source), MEM_OP(reg_x(base), offset)); local 119 void Arm64Assembler::StoreSToOffset(SRegister source, XRegister base, int32_t offset) { argument 120 ___ Str(reg_s(source), MEM_OP(reg_x(base), offset)); local 97 StoreWToOffset(StoreOperandType type, WRegister source, XRegister base, int32_t offset) argument 123 StoreDToOffset(DRegister source, XRegister base, int32_t offset) argument 124 ___ Str(reg_d(source), MEM_OP(reg_x(base), offset)); local 220 LoadWFromOffset(LoadOperandType type, WRegister dest, XRegister base, int32_t offset) argument 224 ___ Ldrsb(reg_w(dest), MEM_OP(reg_x(base), offset)); local 227 ___ Ldrsh(reg_w(dest), MEM_OP(reg_x(base), offset)); local 230 ___ Ldrb(reg_w(dest), MEM_OP(reg_x(base), offset)); local 233 ___ Ldrh(reg_w(dest), MEM_OP(reg_x(base), offset)); local 236 ___ Ldr(reg_w(dest), MEM_OP(reg_x(base), offset)); local 245 LoadFromOffset(XRegister dest, XRegister base, int32_t offset) argument 248 ___ Ldr(reg_x(dest), MEM_OP(reg_x(base), offset)); local 251 LoadSFromOffset(SRegister dest, XRegister base, int32_t offset) argument 253 ___ Ldr(reg_s(dest), MEM_OP(reg_x(base), offset)); local 256 LoadDFromOffset(DRegister dest, XRegister base, int32_t offset) argument 258 ___ Ldr(reg_d(dest), MEM_OP(reg_x(base), offset)); local 261 Load(Arm64ManagedRegister dest, XRegister base, int32_t offset, size_t size) argument 267 ___ Ldr(reg_w(dest.AsWRegister()), MEM_OP(reg_x(base), offset)); local 271 ___ Ldr(reg_w(dest.AsOverlappingWRegister()), MEM_OP(reg_x(base), offset)); local 274 ___ Ldr(reg_x(dest.AsXRegister()), MEM_OP(reg_x(base), offset)); local 277 ___ Ldr(reg_s(dest.AsSRegister()), MEM_OP(reg_x(base), offset)); local 280 ___ Ldr(reg_d(dest.AsDRegister()), MEM_OP(reg_x(base), offset)); local 648 SpillRegisters(vixl::CPURegList registers, int offset) argument 655 cfi_.RelOffset(DWARFReg(dst0), offset); local 662 cfi_.RelOffset(DWARFReg(dst0), offset); local 667 UnspillRegisters(vixl::CPURegList registers, int offset) argument 720 int32_t offset = frame_size + kArm64PointerSize; local [all...] |
/art/compiler/debug/dwarf/ |
H A D | dwarf_test.cc | 37 // Pick offset value which would catch Uleb vs Sleb errors. 38 const int offset = 40000; local 39 ASSERT_EQ(UnsignedLeb128Size(offset / 4), 2u); 40 ASSERT_EQ(SignedLeb128Size(offset / 4), 3u); 60 opcodes.DefCFA(reg, offset); 62 opcodes.DefCFA(reg, -offset); 66 opcodes.DefCFAOffset(offset); 68 opcodes.DefCFAOffset(-offset); 77 opcodes.Offset(Reg(0x3F), -offset); 81 opcodes.Offset(Reg(0x40), -offset); [all...] |