/external/compiler-rt/lib/sanitizer_common/ |
H A D | sanitizer_tls_get_addr.cc | 50 atomic_fetch_sub(&number_of_live_dtls, 1, memory_order_relaxed); 60 atomic_fetch_add(&number_of_live_dtls, 1, memory_order_relaxed); 93 atomic_load(&number_of_live_dtls, memory_order_relaxed));
|
H A D | sanitizer_atomic_msvc.h | 70 DCHECK(mo & (memory_order_relaxed | memory_order_consume 75 if (mo == memory_order_relaxed) { 87 DCHECK(mo & (memory_order_relaxed | memory_order_release 91 if (mo == memory_order_relaxed) {
|
H A D | sanitizer_persistent_allocator.h | 55 atomic_store(®ion_pos, 0, memory_order_relaxed);
|
H A D | sanitizer_stoptheworld_linux_libcdep.cc | 223 atomic_store(&inst->arg->done, 1, memory_order_relaxed); 300 atomic_store(&tracer_thread_argument->done, 1, memory_order_relaxed); 374 atomic_store(&tracer_thread_argument.done, 0, memory_order_relaxed); 430 while (atomic_load(&tracer_thread_argument.done, memory_order_relaxed) == 0)
|
H A D | sanitizer_stackdepotbase.h | 77 uptr cmp = atomic_load(p, memory_order_relaxed); 118 u32 id = atomic_fetch_add(&seq[part], 1, memory_order_relaxed) + 1; 170 uptr s = atomic_load(p, memory_order_relaxed);
|
H A D | sanitizer_deadlock_detector2.cc | 157 atomic_store(&m->owner, 0, memory_order_relaxed); 200 uptr owner = atomic_load(&m->owner, memory_order_relaxed); 279 uptr owner = atomic_load(&m->owner, memory_order_relaxed); 291 atomic_store(&m->owner, (uptr)cb->lt, memory_order_relaxed); 311 uptr owner = atomic_load(&m->owner, memory_order_relaxed); 317 atomic_store(&m->owner, 0, memory_order_relaxed);
|
H A D | sanitizer_allocator.cc | 61 if (atomic_load(&internal_allocator_initialized, memory_order_relaxed) ==
|
H A D | sanitizer_libignore.cc | 80 const uptr idx = atomic_load(&loaded_count_, memory_order_relaxed);
|
H A D | sanitizer_linux_libcdep.cc | 206 uptr val = atomic_load(&kThreadDescriptorSize, memory_order_relaxed); 238 atomic_store(&kThreadDescriptorSize, val, memory_order_relaxed); 246 atomic_store(&kThreadDescriptorSize, val, memory_order_relaxed);
|
H A D | sanitizer_common.cc | 287 atomic_fetch_add(&g_total_mmaped, size, memory_order_relaxed) + size; 295 atomic_fetch_sub(&g_total_mmaped, size, memory_order_relaxed);
|
/external/clang/test/CodeGen/ |
H A D | atomic-ops.c | 152 return __c11_atomic_load(d, memory_order_relaxed); 248 return __c11_atomic_fetch_add(p, 1, memory_order_relaxed); 256 return __atomic_fetch_sub(p, 4, memory_order_relaxed); 428 __c11_atomic_compare_exchange_strong(ptr, ptr2, 43, memory_order_acquire, memory_order_relaxed);
|
/external/compiler-rt/lib/dfsan/ |
H A D | dfsan.cc | 142 atomic_fetch_add(&__dfsan_last_label, 1, memory_order_relaxed) + 1; 205 atomic_fetch_add(&__dfsan_last_label, 1, memory_order_relaxed) + 1; 291 atomic_load(&__dfsan_last_label, memory_order_relaxed); 299 atomic_load(&__dfsan_last_label, memory_order_relaxed);
|
/external/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_mutex.cc | 217 atomic_store(&state_, kUnlocked, memory_order_relaxed); 221 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); 233 if (atomic_load(&state_, memory_order_relaxed) == kUnlocked) { 284 CHECK_NE(atomic_load(&state_, memory_order_relaxed), 0);
|
H A D | tsan_interceptors.cc | 294 atomic_store(&ctx->in_blocking_func, 1, memory_order_relaxed); 295 if (atomic_load(&ctx->have_pending_signals, memory_order_relaxed) == 0) 297 atomic_store(&ctx->in_blocking_func, 0, memory_order_relaxed); 310 atomic_store(&ctx->in_blocking_func, 0, memory_order_relaxed); 438 atomic_load(&sctx->in_blocking_func, memory_order_relaxed) : 441 memory_order_relaxed); 462 memory_order_relaxed); 465 memory_order_relaxed); 871 if (atomic_compare_exchange_strong(g, &cmp, 1<<16, memory_order_relaxed)) 890 atomic_store(g, 0, memory_order_relaxed); [all...] |
H A D | tsan_sync.cc | 60 atomic_store(&uid_gen_, 0, memory_order_relaxed); 232 const u64 uid = atomic_fetch_add(&uid_gen_, 1, memory_order_relaxed);
|
/external/libcxx/test/std/thread/futures/futures.shared_future/ |
H A D | wait_until.pass.cpp | 31 thread_state.store(state, std::memory_order_relaxed); 36 while (thread_state.load(std::memory_order_relaxed) != state);
|
/external/libcxx/test/std/thread/futures/futures.unique_future/ |
H A D | wait_until.pass.cpp | 31 thread_state.store(state, std::memory_order_relaxed); 36 while (thread_state.load(std::memory_order_relaxed) != state);
|
/external/libcxx/test/std/atomics/atomics.types.generic/ |
H A D | bool.pass.cpp | 80 assert(obj.exchange(true, std::memory_order_relaxed) == false); 136 assert(obj.exchange(true, std::memory_order_relaxed) == false); 192 assert(obj.exchange(true, std::memory_order_relaxed) == false);
|
H A D | address.pass.cpp | 98 assert(obj.exchange(T(3), std::memory_order_relaxed) == T(2));
|
H A D | integral.pass.cpp | 114 assert(obj.exchange(T(3), std::memory_order_relaxed) == T(2));
|
/external/compiler-rt/lib/ubsan/ |
H A D | ubsan_value.h | 63 __sanitizer::memory_order_relaxed);
|
/external/compiler-rt/lib/lsan/ |
H A D | lsan_allocator.cc | 78 atomic_store(reinterpret_cast<atomic_uint8_t *>(m), 1, memory_order_relaxed); local 85 atomic_store(reinterpret_cast<atomic_uint8_t *>(m), 0, memory_order_relaxed); local
|
/external/compiler-rt/lib/tsan/tests/rtl/ |
H A D | tsan_test_util_linux.cc | 84 uintptr_t addr = atomic_fetch_add(&uniq, size, memory_order_relaxed); 348 CHECK_EQ(atomic_load(&event, memory_order_relaxed), 0); 359 atomic_store(&impl_->event, 0, memory_order_relaxed);
|
/external/clang/lib/Headers/ |
H A D | stdatomic.h | 62 memory_order_relaxed = __ATOMIC_RELAXED, enumerator in enum:memory_order
|
/external/compiler-rt/lib/tsan/dd/ |
H A D | dd_rtl.cc | 102 uptr id = atomic_fetch_add(&id_gen, 1, memory_order_relaxed);
|