Home
last modified time | relevance | path

Searched refs:atomic_load_explicit (Results 1 – 11 of 11) sorted by relevance

/bionic/libc/bionic/
Dpthread_key.cpp80 uintptr_t seq = atomic_load_explicit(&key_map[i].seq, memory_order_relaxed); in pthread_key_clean_all()
90 atomic_load_explicit(&key_map[i].key_destructor, memory_order_relaxed)); in pthread_key_clean_all()
95 if (atomic_load_explicit(&key_map[i].seq, memory_order_relaxed) != seq) { in pthread_key_clean_all()
120 uintptr_t seq = atomic_load_explicit(&key_map[i].seq, memory_order_relaxed); in pthread_key_create()
142 uintptr_t seq = atomic_load_explicit(&key_map[key].seq, memory_order_relaxed); in pthread_key_delete()
156 uintptr_t seq = atomic_load_explicit(&key_map[key].seq, memory_order_relaxed); in pthread_getspecific()
174 uintptr_t seq = atomic_load_explicit(&key_map[key].seq, memory_order_relaxed); in pthread_setspecific()
Dpthread_barrier.cpp120 while(atomic_load_explicit(&barrier->state, memory_order_acquire) == RELEASE) { in pthread_barrier_wait()
124 uint32_t prev_wait_count = atomic_load_explicit(&barrier->wait_count, memory_order_relaxed); in pthread_barrier_wait()
154 while (atomic_load_explicit(&barrier->state, memory_order_acquire) == WAIT) { in pthread_barrier_wait()
175 while (atomic_load_explicit(&barrier->state, memory_order_acquire) == RELEASE) { in pthread_barrier_destroy()
178 if (atomic_load_explicit(&barrier->wait_count, memory_order_relaxed) != 0) { in pthread_barrier_destroy()
Dpthread_once.cpp54 int old_value = atomic_load_explicit(once_control_ptr, memory_order_acquire); in pthread_once()
83 old_value = atomic_load_explicit(once_control_ptr, memory_order_acquire); in pthread_once()
Dpthread_rwlock.cpp257 if (atomic_load_explicit(&rwlock->state, memory_order_relaxed) != 0) { in pthread_rwlock_destroy()
273 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed); in __pthread_rwlock_tryrdlock()
292 if (atomic_load_explicit(&rwlock->writer_tid, memory_order_relaxed) == __get_thread()->tid) { in __pthread_rwlock_timedrdlock()
306 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed); in __pthread_rwlock_timedrdlock()
349 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed); in __pthread_rwlock_trywrlock()
365 if (atomic_load_explicit(&rwlock->writer_tid, memory_order_relaxed) == __get_thread()->tid) { in __pthread_rwlock_timedwrlock()
378 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed); in __pthread_rwlock_timedwrlock()
453 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed); in pthread_rwlock_unlock()
455 if (atomic_load_explicit(&rwlock->writer_tid, memory_order_relaxed) != __get_thread()->tid) { in pthread_rwlock_unlock()
D__cxa_guard.cpp82 int old_value = atomic_load_explicit(&gv->state, memory_order_relaxed); in __cxa_guard_acquire()
113 old_value = atomic_load_explicit(&gv->state, memory_order_relaxed); in __cxa_guard_acquire()
Dsemaphore.cpp101 return (atomic_load_explicit(sem_count_ptr, memory_order_relaxed) & SEMCOUNT_SHARED_MASK); in SEM_GET_SHARED()
145 unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed); in __sem_dec()
165 unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed); in __sem_trydec()
189 unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed); in __sem_inc()
Dpthread_mutex.cpp415 uint32_t owner_tid = atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed); in __recursive_or_errorcheck_mutex_wait()
424 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed); in __pthread_mutex_lock_with_timeout()
435 if (tid == atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed)) { in __pthread_mutex_lock_with_timeout()
499 old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed); in __pthread_mutex_lock_with_timeout()
512 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed); in pthread_mutex_lock()
533 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed); in pthread_mutex_unlock()
545 if ( tid != atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed) ) { in pthread_mutex_unlock()
577 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed); in pthread_mutex_trylock()
591 if (tid == atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed)) { in pthread_mutex_trylock()
636 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed); in pthread_mutex_destroy()
Dsystem_properties.cpp396 uint_least32_t off = atomic_load_explicit(off_p, memory_order_consume); in to_prop_bt()
401 uint_least32_t off = atomic_load_explicit(off_p, memory_order_consume); in to_prop_info()
437 uint_least32_t left_offset = atomic_load_explicit(&current->left, memory_order_relaxed); in find_prop_bt()
453 … uint_least32_t right_offset = atomic_load_explicit(&current->right, memory_order_relaxed); in find_prop_bt()
491 … uint_least32_t children_offset = atomic_load_explicit(&current->children, memory_order_relaxed); in find_property()
517 uint_least32_t prop_offset = atomic_load_explicit(&current->prop, memory_order_relaxed); in find_property()
604 uint_least32_t left_offset = atomic_load_explicit(&trie->left, memory_order_relaxed); in foreach_property()
610 uint_least32_t prop_offset = atomic_load_explicit(&trie->prop, memory_order_relaxed); in foreach_property()
617 uint_least32_t children_offset = atomic_load_explicit(&trie->children, memory_order_relaxed); in foreach_property()
623 uint_least32_t right_offset = atomic_load_explicit(&trie->right, memory_order_relaxed); in foreach_property()
[all …]
Dpthread_cond.cpp111 return COND_IS_SHARED(atomic_load_explicit(&state, memory_order_relaxed)); in process_shared()
115 return COND_GET_CLOCK(atomic_load_explicit(&state, memory_order_relaxed)) == CLOCK_REALTIME; in use_realtime_clock()
180 unsigned int old_state = atomic_load_explicit(&cond->state, memory_order_relaxed); in __pthread_cond_timedwait()
/bionic/tests/
Dstdatomic_test.cpp100 ASSERT_EQ(123, atomic_load_explicit(&i, memory_order_relaxed)); in TEST()
207 yval = atomic_load_explicit(&a->y, memory_order_acquire); in reader()
208 zval = atomic_load_explicit(&a->z, memory_order_relaxed); in reader()
209 xval = atomic_load_explicit(&a->x, memory_order_relaxed); in reader()
247 EXPECT_EQ(atomic_load_explicit(&a.x, memory_order_consume), BIG + 1); in TEST()
248 EXPECT_EQ(atomic_load_explicit(&a.y, memory_order_seq_cst), BIG + 1); in TEST()
/bionic/libc/include/
Dstdatomic.h75 using std::atomic_load_explicit;
423 #define atomic_load_explicit(object, order) \ macro
448 #define atomic_load_explicit(object, order) \ macro
498 #define atomic_load_explicit(object, order) \ macro
531 atomic_load_explicit(object, memory_order_seq_cst)