Searched refs:LoadSequentiallyConsistent (Results 1 - 16 of 16) sorted by relevance

/art/runtime/gc/
H A Dtask_processor_test.cc79 while (counter.LoadSequentiallyConsistent() != kRecursion) {
87 ASSERT_TRUE(done_running.LoadSequentiallyConsistent());
99 ASSERT_TRUE(done_running.LoadSequentiallyConsistent());
100 ASSERT_EQ(counter.LoadSequentiallyConsistent(), kRecursion);
140 ASSERT_FALSE(done_running.LoadSequentiallyConsistent());
143 ASSERT_TRUE(done_running.LoadSequentiallyConsistent());
H A Dheap-inl.h182 AllocationListener* l = alloc_listener_.LoadSequentiallyConsistent();
388 size_t new_footprint = num_bytes_allocated_.LoadSequentiallyConsistent() + alloc_size;
H A Dheap.cc1210 old = storage->LoadSequentiallyConsistent();
1628 size_t bytes_freed = num_bytes_freed_revoke_.LoadSequentiallyConsistent();
1787 << count_requested_homogeneous_space_compaction_.LoadSequentiallyConsistent()
1789 << count_performed_homogeneous_space_compaction_.LoadSequentiallyConsistent()
1791 << count_ignored_homogeneous_space_compaction_.LoadSequentiallyConsistent()
1793 << count_delayed_oom_.LoadSequentiallyConsistent();
2100 uint32_t before_allocated = num_bytes_allocated_.LoadSequentiallyConsistent();
2235 int32_t after_allocated = num_bytes_allocated_.LoadSequentiallyConsistent();
2890 return fail_count_->LoadSequentiallyConsistent();
3062 return fail_count_->LoadSequentiallyConsistent();
[all...]
H A Dheap.h490 return num_bytes_allocated_.LoadSequentiallyConsistent();
540 size_t byte_allocated = num_bytes_allocated_.LoadSequentiallyConsistent();
/art/runtime/
H A Dthread_pool_test.cc74 EXPECT_EQ(num_tasks, count.LoadSequentiallyConsistent());
87 EXPECT_EQ(0, count.LoadSequentiallyConsistent());
96 EXPECT_EQ(0, bad_count.LoadSequentiallyConsistent());
160 EXPECT_EQ((1 << depth) - 1, count.LoadSequentiallyConsistent());
H A Datomic.h214 T LoadSequentiallyConsistent() const { function
H A Djava_vm_ext.cc685 allow_accessing_weak_globals_.LoadSequentiallyConsistent();
724 DCHECK(allow_accessing_weak_globals_.LoadSequentiallyConsistent());
/art/runtime/gc/space/
H A Dzygote_space.h70 return objects_allocated_.LoadSequentiallyConsistent();
/art/runtime/base/
H A Dmutex.cc371 if (num_contenders_.LoadSequentiallyConsistent() != 0) {
783 << " state=" << state_.LoadSequentiallyConsistent()
784 << " num_pending_writers=" << num_pending_writers_.LoadSequentiallyConsistent()
785 << " num_pending_readers=" << num_pending_readers_.LoadSequentiallyConsistent()
H A Dmutex.h208 return contention_log_data_->contention_count.LoadSequentiallyConsistent() > 0;
/art/runtime/gc/collector/
H A Dconcurrent_copying.cc1674 uint64_t to_bytes = bytes_moved_.LoadSequentiallyConsistent();
1676 uint64_t to_objects = objects_moved_.LoadSequentiallyConsistent();
1705 LOG(INFO) << "(before) num_bytes_allocated=" << heap_->num_bytes_allocated_.LoadSequentiallyConsistent();
1709 LOG(INFO) << "(after) num_bytes_allocated=" << heap_->num_bytes_allocated_.LoadSequentiallyConsistent();
1876 bool updated_all_immune_objects = updated_all_immune_objects_.LoadSequentiallyConsistent();
2205 << to_space_bytes_skipped_.LoadSequentiallyConsistent()
2206 << " skipped_objects=" << to_space_objects_skipped_.LoadSequentiallyConsistent();
H A Dmark_sweep.cc1385 CHECK_EQ(work_chunks_created_.LoadSequentiallyConsistent(),
1386 work_chunks_deleted_.LoadSequentiallyConsistent())
/art/runtime/jdwp/
H A Djdwp_main.cc619 int64_t last = last_activity_time_ms_.LoadSequentiallyConsistent();
/art/runtime/mirror/
H A Ddex_cache-inl.h156 return ref.LoadSequentiallyConsistent().Read();
H A Dobject.h660 return reinterpret_cast<const Atomic<kSize>*>(addr)->LoadSequentiallyConsistent();
/art/runtime/jit/
H A Djit_code_cache.cc481 : is_weak_access_enabled_.LoadSequentiallyConsistent();

Completed in 206 milliseconds