/art/compiler/linker/ |
H A D | relative_patcher_test.h | 103 static_assert(kTrampolineOffset == 0u, "Unexpected trampoline offset."); 104 uint32_t offset = kTrampolineSize; local 107 offset = patcher_->ReserveSpace(offset, compiled_method.get(), compiled_method_refs_[idx]); 109 uint32_t aligned_offset = compiled_method->AlignCode(offset); 110 uint32_t aligned_code_delta = aligned_offset - offset; 111 offset += aligned_code_delta; 113 offset += sizeof(OatQuickMethodHeader); 114 uint32_t quick_code_offset = offset + compiled_method->CodeDelta(); 116 offset 199 size_t offset = result.second - compiled_methods_[idx]->CodeDelta(); local [all...] |
H A D | error_delaying_output_stream.h | 41 << " bytes to " << GetLocation() << " at offset " << output_offset_; 51 off_t Seek(off_t offset, Whence whence) OVERRIDE { 52 // We keep shadow copy of the offset so that we return 57 new_offset = offset; 60 new_offset = output_offset_ + offset; 67 off_t actual_offset = output_->Seek(offset, whence); 69 PLOG(ERROR) << "Failed to seek in " << GetLocation() << ". Offset=" << offset
|
H A D | relative_patcher.cc | 44 uint32_t ReserveSpace(uint32_t offset, 47 return offset; // No space reserved; no patches expected. 50 uint32_t ReserveSpaceEnd(uint32_t offset) OVERRIDE { 51 return offset; // No space reserved; no patches expected. 54 uint32_t WriteThunks(OutputStream* out ATTRIBUTE_UNUSED, uint32_t offset) OVERRIDE { 55 return offset; // No thunks added; no patches expected.
|
/art/runtime/interpreter/mterp/mips64/ |
H A D | op_goto_32.S | 2 * Unconditional branch, 32-bit offset. 4 * The branch distance is a signed code-unit offset, which we need to 5 * double to get a byte offset. 14 ins rINST, a1, 16, 16 # rINST <- offset (sign-extended AAAAaaaa) 20 jal MterpProfileBranch # (self, shadow_frame, offset) 21 bnezc v0, MterpOnStackReplacement # Note: offset must be in rINST 23 dlsa rPC, rINST, rPC, 1 # rPC <- rPC + offset * 2 25 move a0, rINST # a0 <- offset
|
H A D | op_packed_switch.S | 20 jal $func # v0 <- code-unit branch offset 27 jal MterpProfileBranch # (self, shadow_frame, offset) 28 bnezc v0, MterpOnStackReplacement # Note: offset must be in rINST 30 dlsa rPC, rINST, rPC, 1 # rPC <- rPC + offset * 2 32 move a0, rINST # a0 <- offset
|
/art/compiler/trampolines/ |
H A D | trampoline_compiler.cc | 53 ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) { 58 __ LoadFromOffset(kLoadWord, PC, R0, offset.Int32Value()); 62 __ LoadFromOffset(kLoadWord, PC, IP, offset.Int32Value()); 65 __ LoadFromOffset(kLoadWord, PC, R9, offset.Int32Value()); 83 ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<8> offset) { 88 __ JumpTo(Arm64ManagedRegister::FromXRegister(X0), Offset(offset.Int32Value()), 97 __ JumpTo(Arm64ManagedRegister::FromXRegister(IP1), Offset(offset.Int32Value()), 102 __ JumpTo(Arm64ManagedRegister::FromXRegister(TR), Offset(offset.Int32Value()), 122 ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) { 127 __ LoadFromOffset(kLoadWord, T9, A0, offset 52 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) argument 82 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<8> offset) argument 121 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) argument 153 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<8> offset) argument 185 CreateTrampoline(ArenaAllocator* arena, ThreadOffset<4> offset) argument 206 CreateTrampoline(ArenaAllocator* arena, ThreadOffset<8> offset) argument 225 CreateTrampoline64(InstructionSet isa, EntryPointCallingConvention abi, ThreadOffset<8> offset) argument 251 CreateTrampoline32(InstructionSet isa, EntryPointCallingConvention abi, ThreadOffset<4> offset) argument [all...] |
/art/compiler/debug/dwarf/ |
H A D | debug_frame_opcode_writer.h | 73 void ALWAYS_INLINE RelOffset(Reg reg, int offset) { argument 74 Offset(reg, offset - current_cfa_offset_); 83 void ALWAYS_INLINE RelOffsetForMany(Reg reg_base, int offset, argument 92 RelOffset(Reg(reg_base.num() + i), offset); local 93 offset += reg_size; 117 void ALWAYS_INLINE Offset(Reg reg, int offset) { argument 120 int factored_offset = FactorDataOffset(offset); // May change sign. 191 void ALWAYS_INLINE DefCFA(Reg reg, int offset) { argument 194 if (offset >= 0) { 197 this->PushUleb128(offset); // No 216 DefCFAOffset(int offset) argument 234 ValOffset(Reg reg, int offset) argument 296 SetCurrentCFAOffset(int offset) argument [all...] |
H A D | writer.h | 129 void UpdateUint32(size_t offset, uint32_t value) { argument 130 DCHECK_LT(offset + 3, data_->size()); 131 (*data_)[offset + 0] = (value >> 0) & 0xFF; 132 (*data_)[offset + 1] = (value >> 8) & 0xFF; 133 (*data_)[offset + 2] = (value >> 16) & 0xFF; 134 (*data_)[offset + 3] = (value >> 24) & 0xFF; 137 void UpdateUint64(size_t offset, uint64_t value) { argument 138 DCHECK_LT(offset + 7, data_->size()); 139 (*data_)[offset + 0] = (value >> 0) & 0xFF; 140 (*data_)[offset 149 UpdateUleb128(size_t offset, uint32_t value) argument [all...] |
/art/runtime/ |
H A D | oat.cc | 131 return "Executable offset not page-aligned."; 230 void OatHeader::SetInterpreterToInterpreterBridgeOffset(uint32_t offset) { argument 231 CHECK(offset == 0 || offset >= executable_offset_); 233 DCHECK_EQ(interpreter_to_interpreter_bridge_offset_, 0U) << offset; 235 interpreter_to_interpreter_bridge_offset_ = offset; 248 void OatHeader::SetInterpreterToCompiledCodeBridgeOffset(uint32_t offset) { argument 249 CHECK(offset == 0 || offset >= interpreter_to_interpreter_bridge_offset_); 251 DCHECK_EQ(interpreter_to_compiled_code_bridge_offset_, 0U) << offset; 266 SetJniDlsymLookupOffset(uint32_t offset) argument 284 SetQuickGenericJniTrampolineOffset(uint32_t offset) argument 302 SetQuickImtConflictTrampolineOffset(uint32_t offset) argument 320 SetQuickResolutionTrampolineOffset(uint32_t offset) argument 338 SetQuickToInterpreterBridgeOffset(uint32_t offset) argument [all...] |
H A D | memory_region.h | 52 // Load value of type `T` at `offset`. The memory address corresponding 53 // to `offset` should be word-aligned (on ARM, this is a requirement). 55 ALWAYS_INLINE T Load(uintptr_t offset) const { 56 T* address = ComputeInternalPointer<T>(offset); 61 // Store `value` (of type `T`) at `offset`. The memory address 62 // corresponding to `offset` should be word-aligned (on ARM, this is 65 ALWAYS_INLINE void Store(uintptr_t offset, T value) const { argument 66 T* address = ComputeInternalPointer<T>(offset); 71 // Load value of type `T` at `offset`. The memory address corresponding 72 // to `offset` doe 89 StoreUnaligned(uintptr_t offset, T value) const argument 151 Subregion(uintptr_t offset, uintptr_t size_in) const argument [all...] |
/art/compiler/linker/arm64/ |
H A D | relative_patcher_arm64.h | 31 uint32_t ReserveSpace(uint32_t offset, 34 uint32_t ReserveSpaceEnd(uint32_t offset) OVERRIDE; 35 uint32_t WriteThunks(OutputStream* out, uint32_t offset) OVERRIDE; 51 void SetInsn(std::vector<uint8_t>* code, uint32_t offset, uint32_t value); 52 static uint32_t GetInsn(ArrayRef<const uint8_t> code, uint32_t offset); 55 static uint32_t GetInsn(std::vector<uint8_t, Alloc>* code, uint32_t offset); 67 // Map original patch_offset to thunk offset.
|
H A D | relative_patcher_arm64.cc | 55 uint32_t Arm64RelativePatcher::ReserveSpace(uint32_t offset, argument 60 return ReserveSpaceInternal(offset, compiled_method, method_ref, 0u); 66 offset = CompiledMethod::AlignCode(offset, kArm64) + kAdrpThunkSize * num_adrp_thunks; 79 offset = ReserveSpaceInternal(offset, compiled_method, method_ref, kAdrpThunkSize * num_adrp); 81 return offset; 84 // Now that we have the actual offset where the code will be placed, locate the ADRP insns 86 uint32_t quick_code_offset = compiled_method->AlignCode(offset) + sizeof(OatQuickMethodHeader); 99 return offset; 102 ReserveSpaceEnd(uint32_t offset) argument 116 WriteThunks(OutputStream* out, uint32_t offset) argument 255 assembler.JumpTo(ManagedRegister(arm64::X0), offset, ManagedRegister(arm64::IP0)); local 328 SetInsn(std::vector<uint8_t>* code, uint32_t offset, uint32_t value) argument 338 GetInsn(ArrayRef<const uint8_t> code, uint32_t offset) argument 350 GetInsn(std::vector<uint8_t, Alloc>* code, uint32_t offset) argument [all...] |
/art/compiler/linker/arm/ |
H A D | relative_patcher_arm_base.cc | 27 uint32_t ArmBaseRelativePatcher::ReserveSpace(uint32_t offset, argument 30 return ReserveSpaceInternal(offset, compiled_method, method_ref, 0u); 33 uint32_t ArmBaseRelativePatcher::ReserveSpaceEnd(uint32_t offset) { argument 37 // offset after reserving of writing any chunk. 38 uint32_t aligned_offset = CompiledMethod::AlignCode(offset, instruction_set_); 49 offset = CompiledMethod::AlignCode(aligned_offset + thunk_code_.size(), instruction_set_); 51 return offset; 54 uint32_t ArmBaseRelativePatcher::WriteThunks(OutputStream* out, uint32_t offset) { argument 56 return offset; 58 uint32_t aligned_offset = CompiledMethod::AlignCode(offset, instruction_set 90 ReserveSpaceInternal(uint32_t offset, const CompiledMethod* compiled_method, MethodReference method_ref, uint32_t max_extra_space) argument [all...] |
H A D | relative_patcher_arm_base.h | 30 uint32_t ReserveSpace(uint32_t offset, 33 uint32_t ReserveSpaceEnd(uint32_t offset) OVERRIDE; 34 uint32_t WriteThunks(OutputStream* out, uint32_t offset) OVERRIDE; 43 uint32_t ReserveSpaceInternal(uint32_t offset,
|
H A D | relative_patcher_thumb2.cc | 96 void Thumb2RelativePatcher::SetInsn32(std::vector<uint8_t>* code, uint32_t offset, uint32_t value) { argument 97 DCHECK_LE(offset + 4u, code->size()); 98 DCHECK_EQ(offset & 1u, 0u); 99 uint8_t* addr = &(*code)[offset]; 106 uint32_t Thumb2RelativePatcher::GetInsn32(ArrayRef<const uint8_t> code, uint32_t offset) { argument 107 DCHECK_LE(offset + 4u, code.size()); 108 DCHECK_EQ(offset & 1u, 0u); 109 const uint8_t* addr = &code[offset]; 118 uint32_t Thumb2RelativePatcher::GetInsn32(Vector* code, uint32_t offset) { argument 120 return GetInsn32(ArrayRef<const uint8_t>(*code), offset); [all...] |
/art/runtime/interpreter/mterp/mips/ |
H A D | op_goto_16.S | 2 * Unconditional branch, 16-bit offset. 4 * The branch distance is a signed code-unit offset, which we need to 5 * double to get a byte offset. 14 JAL(MterpProfileBranch) # (self, shadow_frame, offset) 15 bnez v0, MterpOnStackReplacement # Note: offset must be in rINST 16 addu a1, rINST, rINST # a1 <- byte offset, flags set 26 addu a1, rINST, rINST # a1 <- byte offset, flags set
|
H A D | bincmp.S | 14 FETCH_S(rINST, 1) # rINST<- branch offset, in code units 24 JAL(MterpProfileBranch) # (self, shadow_frame, offset) 25 bnez v0, MterpOnStackReplacement # Note: offset must be in rINST
|
H A D | zcmp.S | 11 FETCH_S(rINST, 1) # rINST <- branch offset, in code units 22 JAL(MterpProfileBranch) # (self, shadow_frame, offset) 23 bnez v0, MterpOnStackReplacement # Note: offset must be in rINST
|
/art/test/528-long-hint/src/ |
H A D | Main.java | 29 long offset = unsafe.objectFieldOffset(Main.class.getDeclaredField("instanceField")); 30 getUnsafe(); // spill offset 36 unsafe.compareAndSwapLong(f, offset, a, b);
|
/art/runtime/base/unix_file/ |
H A D | fd_file.cc | 154 int64_t FdFile::Read(char* buf, int64_t byte_count, int64_t offset) const { 156 int rc = TEMP_FAILURE_RETRY(pread64(fd_, buf, byte_count, offset)); 158 int rc = TEMP_FAILURE_RETRY(pread(fd_, buf, byte_count, offset)); 180 int64_t FdFile::Write(const char* buf, int64_t byte_count, int64_t offset) { argument 183 int rc = TEMP_FAILURE_RETRY(pwrite64(fd_, buf, byte_count, offset)); 185 int rc = TEMP_FAILURE_RETRY(pwrite(fd_, buf, byte_count, offset)); 207 static ssize_t ReadIgnoreOffset(int fd, void *buf, size_t count, off_t offset) { argument 208 DCHECK_EQ(offset, 0); 213 static bool ReadFullyGeneric(int fd, void* buffer, size_t byte_count, size_t offset) { argument 216 ssize_t bytes_read = TEMP_FAILURE_RETRY(read_func(fd, ptr, byte_count, offset)); 233 PreadFully(void* buffer, size_t byte_count, size_t offset) argument 238 WriteFullyGeneric(const void* buffer, size_t byte_count, size_t offset) argument 257 PwriteFully(const void* buffer, size_t byte_count, size_t offset) argument 265 Copy(FdFile* input_file, int64_t offset, int64_t size) argument [all...] |
H A D | fd_file.h | 55 int64_t Read(char* buf, int64_t byte_count, int64_t offset) const OVERRIDE WARN_UNUSED; 58 int64_t Write(const char* buf, int64_t byte_count, int64_t offset) OVERRIDE WARN_UNUSED; 80 bool PreadFully(void* buffer, size_t byte_count, size_t offset) WARN_UNUSED; 82 bool PwriteFully(const void* buffer, size_t byte_count, size_t offset) WARN_UNUSED; 85 bool Copy(FdFile* input_file, int64_t offset, int64_t size); 86 // Clears the file content and resets the file offset to 0. 89 // Resets the file offset to the beginning of the file. 124 bool WriteFullyGeneric(const void* buffer, size_t byte_count, size_t offset);
|
/art/tools/dexfuzz/src/dexfuzz/rawdex/ |
H A D | MapItem.java | 43 public Offset offset; field in class:MapItem 51 offset = file.getOffsetTracker().getNewHeaderOffset(file.readUInt()); 53 offset = file.getOffsetTracker().getNewOffset(file.readUInt()); 62 file.getOffsetTracker().tryToWriteOffset(offset, file, false /* ULEB128 */);
|
/art/runtime/native/ |
H A D | java_lang_StringFactory.cc | 31 jint high, jint offset, jint byte_count) { 40 if ((offset | byte_count) < 0 || byte_count > data_size - offset) { 43 offset, byte_count); 48 byte_array, offset, high, 54 static jstring StringFactory_newStringFromChars(JNIEnv* env, jclass, jint offset, argument 62 char_array, offset, 30 StringFactory_newStringFromBytes(JNIEnv* env, jclass, jbyteArray java_data, jint high, jint offset, jint byte_count) argument
|
H A D | sun_misc_Unsafe.cc | 33 static jboolean Unsafe_compareAndSwapInt(JNIEnv* env, jobject, jobject javaObj, jlong offset, argument 38 bool success = obj->CasFieldStrongSequentiallyConsistent32<false>(MemberOffset(offset), 43 static jboolean Unsafe_compareAndSwapLong(JNIEnv* env, jobject, jobject javaObj, jlong offset, argument 48 bool success = obj->CasFieldStrongSequentiallyConsistent64<false>(MemberOffset(offset), 53 static jboolean Unsafe_compareAndSwapObject(JNIEnv* env, jobject, jobject javaObj, jlong offset, argument 65 reinterpret_cast<uint8_t*>(obj) + static_cast<size_t>(offset)); 68 MemberOffset(offset), 71 bool success = obj->CasFieldStrongSequentiallyConsistentObject<false>(MemberOffset(offset), 76 static jint Unsafe_getInt(JNIEnv* env, jobject, jobject javaObj, jlong offset) { argument 79 return obj->GetField32(MemberOffset(offset)); 82 Unsafe_getIntVolatile(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 88 Unsafe_putInt(JNIEnv* env, jobject, jobject javaObj, jlong offset, jint newValue) argument 95 Unsafe_putIntVolatile(JNIEnv* env, jobject, jobject javaObj, jlong offset, jint newValue) argument 103 Unsafe_putOrderedInt(JNIEnv* env, jobject, jobject javaObj, jlong offset, jint newValue) argument 112 Unsafe_getLong(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 118 Unsafe_getLongVolatile(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 124 Unsafe_putLong(JNIEnv* env, jobject, jobject javaObj, jlong offset, jlong newValue) argument 131 Unsafe_putLongVolatile(JNIEnv* env, jobject, jobject javaObj, jlong offset, jlong newValue) argument 139 Unsafe_putOrderedLong(JNIEnv* env, jobject, jobject javaObj, jlong offset, jlong newValue) argument 148 Unsafe_getObjectVolatile(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 155 Unsafe_getObject(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 162 Unsafe_putObject(JNIEnv* env, jobject, jobject javaObj, jlong offset, jobject javaNewValue) argument 171 Unsafe_putObjectVolatile(JNIEnv* env, jobject, jobject javaObj, jlong offset, jobject javaNewValue) argument 180 Unsafe_putOrderedObject(JNIEnv* env, jobject, jobject javaObj, jlong offset, jobject javaNewValue) argument 391 Unsafe_getBoolean(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 397 Unsafe_putBoolean(JNIEnv* env, jobject, jobject javaObj, jlong offset, jboolean newValue) argument 404 Unsafe_getByte(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 410 Unsafe_putByte(JNIEnv* env, jobject, jobject javaObj, jlong offset, jbyte newValue) argument 417 Unsafe_getChar(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 423 Unsafe_putChar(JNIEnv* env, jobject, jobject javaObj, jlong offset, jchar newValue) argument 430 Unsafe_getShort(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 436 Unsafe_putShort(JNIEnv* env, jobject, jobject javaObj, jlong offset, jshort newValue) argument 443 Unsafe_getFloat(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 451 Unsafe_putFloat(JNIEnv* env, jobject, jobject javaObj, jlong offset, jfloat newValue) argument 460 Unsafe_getDouble(JNIEnv* env, jobject, jobject javaObj, jlong offset) argument 468 Unsafe_putDouble(JNIEnv* env, jobject, jobject javaObj, jlong offset, jdouble newValue) argument [all...] |
/art/runtime/base/ |
H A D | hash_set.h | 213 size_t offset = 0; variable 214 offset = ReadFromBytes(ptr, offset, &temp); 216 offset = ReadFromBytes(ptr, offset, &temp); 219 offset = ReadFromBytes(ptr, offset, &temp); 221 offset = ReadFromBytes(ptr, offset, &min_load_factor_); 222 offset 242 size_t offset = 0; local 648 WriteToBytes(uint8_t* ptr, size_t offset, Elem n) argument 657 ReadFromBytes(const uint8_t* ptr, size_t offset, Elem* out) argument [all...] |