/art/tools/checker/file_format/c1visualizer/ |
H A D | parser.py | 28 def __parseC1Line(line, lineNo, state, fileName): 37 if state.currentState == C1ParserState.StartingCfgBlock: 43 state.currentState = C1ParserState.InsideCfgBlock 44 return (None, state.lastMethodName + " " + line.split("\"")[1], None) 48 elif state.currentState == C1ParserState.InsideCfgBlock: 50 state.currentState = C1ParserState.OutsideBlock 55 elif state.currentState == C1ParserState.InsideCompilationBlock: 61 state.lastMethodName = methodName 63 state.currentState = C1ParserState.OutsideBlock 67 assert state [all...] |
/art/test/123-compiler-regressions-mt/src/ |
H A D | Main.java | 59 private volatile int state = 0; field in class:B17689750TestVolatile 63 while (state != 1) { } // Busy loop. 65 state = 2; 70 state = 1; 71 while (state != 2) { } // Busy loop. 80 private int state = 0; field in class:B17689750TestMonitor 88 s = state; 94 state = 2; 102 state = 1; 108 s = state; [all...] |
/art/runtime/ |
H A D | backtrace_helper.h | 41 auto* const state = reinterpret_cast<BacktraceCollector*>(arg); local 44 if (ip != 0 && state->skip_count_ > 0) { 45 --state->skip_count_; 49 state->out_frames_[state->num_frames_] = ip; 50 state->num_frames_++; 51 return state->num_frames_ >= state->max_depth_ ? _URC_END_OF_STACK : _URC_NO_REASON;
|
H A D | thread-inl.h | 130 << Thread::Current() << ") changing state to " << new_state; 134 CHECK_NE(old_state_and_flags.as_struct.state, kRunnable); 135 tls32_.state_and_flags.as_struct.state = new_state; 136 return static_cast<ThreadState>(old_state_and_flags.as_struct.state); 191 // Change the state but keep the current flags (kCheckpointRequest is clear). 195 new_state_and_flags.as_struct.state = new_state; 226 // Change to non-runnable state, thereby appearing suspended to the system. 237 int16_t old_state = old_state_and_flags.as_struct.state; 242 DCHECK_EQ(old_state_and_flags.as_struct.state, old_state); 248 new_state_and_flags.as_struct.state [all...] |
H A D | runtime_callbacks_test.cc | 117 if (state == CallbackState::kBase) { 118 state = CallbackState::kStarted; 121 state = CallbackState::kWrongStart; 126 if (state == CallbackState::kStarted && self == stored_self) { 127 state = CallbackState::kDied; 129 state = CallbackState::kWrongDeath; 134 CallbackState state = CallbackState::kBase; member in struct:art::ThreadLifecycleCallbackRuntimeCallbacksTest::Callback 147 cb_.state = CallbackState::kBase; // Ignore main thread attach. 184 EXPECT_TRUE(cb_.state == CallbackState::kDied) << static_cast<int>(cb_.state); [all...] |
H A D | fault_handler.cc | 244 // is in Runnable state. 252 ThreadState state = thread->GetState(); local 253 if (state != kRunnable) {
|
H A D | monitor.cc | 56 * from the "thin" state to the "fat" state and this transition is referred to as inflation. Once 57 * a lock has been inflated it remains in the "fat" state indefinitely. 136 // Propagate the lock state. 157 LOG(FATAL) << "Invalid monitor state " << lw.GetState(); 646 // Update thread state. If the GC wakes up, it'll ignore us, knowing 647 // that we won't touch any references in this state, and we'll check 681 // that a thread in a waiting/sleeping state has a non-null wait_monitor_ for debugging 797 // Assume no concurrent read barrier state changes as mutators are suspended. 803 // Assume no concurrent read barrier state change 1149 ThreadState state = thread->GetState(); local [all...] |
H A D | thread.h | 164 // Reset internal state of child thread after fork. 194 // Dumps a one-line summary of thread state (used for operator<<). 197 // Dumps the detailed thread state and the thread stack (used for SIGQUIT). 218 DCHECK_GE(tls32_.state_and_flags.as_struct.state, kTerminated); 219 DCHECK_LE(tls32_.state_and_flags.as_struct.state, kSuspended); 220 return static_cast<ThreadState>(tls32_.state_and_flags.as_struct.state); 236 return state_and_flags.as_struct.state != kRunnable && 275 // Transition from non-runnable to runnable state acquiring share on mutator_lock_. 280 // Transition from runnable into a state where mutator privileges are denied. Releases share of 1217 // Need to run pending checkpoint and suspend barriers. Run checkpoints in runnable state i 1323 volatile uint16_t state; variable [all...] |
H A D | primitive.h | 259 std::ostream& operator<<(std::ostream& os, const Primitive::Type& state);
|
/art/tools/dexfuzz/src/dexfuzz/ |
H A D | StreamConsumer.java | 37 private State state; field in class:StreamConsumer 58 state = State.WAITING; 172 Log.error("invalid state: StreamConsumer told about work, but not CONSUMING?"); 173 Log.error("state was: " + getCurrentState()); 183 return (expectedState == state); 187 if (state != previousState) { 188 Log.error("StreamConsumer Unexpected state: " + state + ", expected " + previousState); 189 state = State.ERROR; 191 state [all...] |
/art/runtime/base/ |
H A D | mutex_test.cc | 109 RecursiveLockWait* state = reinterpret_cast<RecursiveLockWait*>(arg); local 110 state->mu.Lock(Thread::Current()); 111 state->cv.Signal(Thread::Current()); 112 state->mu.Unlock(Thread::Current()); 118 RecursiveLockWait state; variable 119 state.mu.Lock(Thread::Current()); 120 state.mu.Lock(Thread::Current()); 123 int pthread_create_result = pthread_create(&pthread, nullptr, RecursiveLockWaitCallback, &state); 126 state.cv.Wait(Thread::Current()); 128 state [all...] |
H A D | mutex-inl.h | 160 // Reduce state by 1 and impose lock release load/store ordering. 162 // TODO: the ordering here is non-trivial as state is split across 3 fields, fix by placing 163 // a status bit into the state on contention. 211 int32_t state = state_.LoadRelaxed(); local 212 if (state == 0) { 214 } else if (state > 0) {
|
/art/runtime/jdwp/ |
H A D | jdwp_socket.cc | 42 * JDWP network state. 50 explicit JdwpSocketState(JdwpState* state) argument 51 : JdwpNetStateBase(state), 67 static JdwpSocketState* SocketStartup(JdwpState* state, uint16_t port, bool probe); 72 bool InitSocketTransport(JdwpState* state, const JdwpOptions* options) { argument 78 state->netState = SocketStartup(state, port, false); 82 state->netState = SocketStartup(state, port, true); 83 if (state 117 SocketStartup(JdwpState* state, uint16_t port, bool probe) argument [all...] |
H A D | jdwp_main.cc | 43 JdwpNetStateBase::JdwpNetStateBase(JdwpState* state) argument 44 : state_(state), socket_lock_("JdwpNetStateBase lock", kJdwpSocketLock) { 116 // Resets the state so we're ready to receive a new connection. 254 std::unique_ptr<JdwpState> state(new JdwpState(options)); 257 InitSocketTransport(state.get(), options); 261 InitAdbTransport(state.get(), options); 272 state->thread_start_lock_.AssertNotHeld(self); 273 MutexLock thread_start_locker(self, state->thread_start_lock_); 279 CHECK_PTHREAD_CALL(pthread_create, (&state->pthread_, nullptr, StartJdwpThread, state 441 JdwpState* state = reinterpret_cast<JdwpState*>(arg); local [all...] |
H A D | jdwp_adb.cc | 69 explicit JdwpAdbState(JdwpState* state) argument 70 : JdwpNetStateBase(state), 154 bool InitAdbTransport(JdwpState* state, const JdwpOptions*) { argument 156 state->netState = new JdwpAdbState(state); 157 return (state->netState != nullptr);
|
H A D | jdwp_handler.cc | 259 static JdwpError VM_Exit(JdwpState* state, Request* request, ExpandBuf*) 262 state->ExitAfterReplying(exit_status); 532 static JdwpError RT_Signature(JdwpState* state, Request* request, ExpandBuf* pReply) 534 return RT_Signature(state, request, pReply, false); 537 static JdwpError RT_SignatureWithGeneric(JdwpState* state, Request* request, ExpandBuf* pReply) 539 return RT_Signature(state, request, pReply, true); 650 static JdwpError CT_InvokeMethod(JdwpState* state, Request* request, 657 return RequestInvoke(state, request, thread_id, 0, class_id, method_id, false); 667 static JdwpError CT_NewInstance(JdwpState* state, Request* request, 679 return RequestInvoke(state, reques [all...] |
/art/runtime/mirror/ |
H A D | throwable.cc | 61 void Throwable::SetStackState(ObjPtr<Object> state) REQUIRES_SHARED(Locks::mutator_lock_) { 62 CHECK(state != nullptr); 64 SetFieldObjectVolatile<true>(OFFSET_OF_OBJECT_MEMBER(Throwable, backtrace_), state); local 66 SetFieldObjectVolatile<false>(OFFSET_OF_OBJECT_MEMBER(Throwable, backtrace_), state); local 98 // check stack state isn't missing or corrupt
|
H A D | throwable.h | 46 void SetStackState(ObjPtr<Object> state) REQUIRES_SHARED(Locks::mutator_lock_);
|
/art/test/924-threads/ |
H A D | threads.cc | 101 jint state; local 102 jvmtiError result = jvmti_env->GetThreadState(thread, &state); 106 return state;
|
/art/runtime/openjdkjvmti/ |
H A D | events.cc | 377 art::ThreadState state = thread->GetState(); local 378 if (state == art::ThreadState::kStarting || 379 state == art::ThreadState::kTerminated ||
|
H A D | ti_stack.cc | 204 art::ThreadState state = thread->GetState(); local 205 if (state == art::ThreadState::kStarting || 206 state == art::ThreadState::kTerminated || 335 stack_info.state = JVMTI_THREAD_STATE_SUSPENDED; 502 stack_info.state = JVMTI_THREAD_STATE_SUSPENDED; 560 stack_info[i].state = started ? kTerminatedState : kStartedState;
|
/art/test/924-threads/src/art/ |
H A D | Test924.java | 32 // Run the test on its own thread, so we have a known state for the "current" thread. 163 // messy. Wait till we get the right Java-side Thread state. 362 int state = getThreadState(t); 367 if ((state & i) != 0) { 379 System.out.println(Integer.toHexString(state) + " = " + sb.toString());
|
/art/runtime/gc/space/ |
H A D | region_space.cc | 202 RegionState state = r->State(); local 207 DCHECK((state == RegionState::kRegionStateAllocated || 208 state == RegionState::kRegionStateLarge) && 218 if (UNLIKELY(state == RegionState::kRegionStateLarge && 225 DCHECK(state == RegionState::kRegionStateLargeTail && 525 << " state=" << static_cast<uint>(state_) << " type=" << static_cast<uint>(type_)
|
/art/test/115-native-bridge/ |
H A D | nativebridge.cc | 255 void Set(TestStatus state) { argument 256 switch (state) { 278 printf("ERROR: unknown state\n"); 282 state_ = state; 290 printf("ERROR: unexpected state, was %d, expected %d\n", state_, expected);
|
/art/oatdump/ |
H A D | oatdump.cc | 2110 ImageDumper* state = reinterpret_cast<ImageDumper*>(arg); local 2111 if (!state->InDumpSpace(obj)) { 2117 state->stats_.object_bytes += object_bytes; 2118 state->stats_.alignment_bytes += alignment_bytes; 2120 std::ostream& os = state->vios_.Stream(); 2137 ScopedIndentation indent1(&state->vios_); 2139 const PointerSize image_pointer_size = state->image_header_.GetPointerSize(); 2166 ScopedIndentation indent2(&state->vios_); 2172 auto it = state->dex_caches_.find(obj); 2173 if (it != state [all...] |