am e375ecaa: Merge "Switch pthread_cond_t to <stdatomic.h>."

* commit 'e375ecaa39213ec01464dac9d0ffebaac70b4e9d':
  Switch pthread_cond_t to <stdatomic.h>.
This commit is contained in:
Yabin Cui 2015-02-10 00:18:47 +00:00 committed by Android Git Automerger
commit fa592b2a90
2 changed files with 59 additions and 39 deletions

View File

@ -30,13 +30,13 @@
#include <errno.h> #include <errno.h>
#include <limits.h> #include <limits.h>
#include <stdatomic.h>
#include <sys/mman.h> #include <sys/mman.h>
#include <time.h> #include <time.h>
#include <unistd.h> #include <unistd.h>
#include "pthread_internal.h" #include "pthread_internal.h"
#include "private/bionic_atomic_inline.h"
#include "private/bionic_futex.h" #include "private/bionic_futex.h"
#include "private/bionic_time_conversions.h" #include "private/bionic_time_conversions.h"
#include "private/bionic_tls.h" #include "private/bionic_tls.h"
@ -98,6 +98,14 @@ int pthread_condattr_destroy(pthread_condattr_t* attr) {
return 0; return 0;
} }
static inline atomic_uint* COND_TO_ATOMIC_POINTER(pthread_cond_t* cond) {
static_assert(sizeof(atomic_uint) == sizeof(cond->value),
"cond->value should actually be atomic_uint in implementation.");
// We prefer casting to atomic_uint instead of declaring cond->value to be atomic_uint directly.
// Because using the second method pollutes pthread.h, and causes an error when compiling libcxx.
return reinterpret_cast<atomic_uint*>(&cond->value);
}
// XXX *technically* there is a race condition that could allow // XXX *technically* there is a race condition that could allow
// XXX a signal to be missed. If thread A is preempted in _wait() // XXX a signal to be missed. If thread A is preempted in _wait()
@ -107,53 +115,54 @@ int pthread_condattr_destroy(pthread_condattr_t* attr) {
// XXX then the signal will be lost. // XXX then the signal will be lost.
int pthread_cond_init(pthread_cond_t* cond, const pthread_condattr_t* attr) { int pthread_cond_init(pthread_cond_t* cond, const pthread_condattr_t* attr) {
atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
unsigned int init_value = 0;
if (attr != NULL) { if (attr != NULL) {
cond->value = (*attr & COND_FLAGS_MASK); init_value = (*attr & COND_FLAGS_MASK);
} else {
cond->value = 0;
} }
atomic_init(cond_value_ptr, init_value);
return 0; return 0;
} }
int pthread_cond_destroy(pthread_cond_t* cond) { int pthread_cond_destroy(pthread_cond_t* cond) {
cond->value = 0xdeadc04d; atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
atomic_store_explicit(cond_value_ptr, 0xdeadc04d, memory_order_relaxed);
return 0; return 0;
} }
// This function is used by pthread_cond_broadcast and // This function is used by pthread_cond_broadcast and
// pthread_cond_signal to atomically decrement the counter // pthread_cond_signal to atomically decrement the counter
// then wake up 'counter' threads. // then wake up thread_count threads.
static int __pthread_cond_pulse(pthread_cond_t* cond, int counter) { static int __pthread_cond_pulse(atomic_uint* cond_value_ptr, int thread_count) {
int flags = (cond->value & COND_FLAGS_MASK); unsigned int old_value = atomic_load_explicit(cond_value_ptr, memory_order_relaxed);
while (true) { bool shared = COND_IS_SHARED(old_value);
int old_value = cond->value;
int new_value = ((old_value - COND_COUNTER_STEP) & COND_COUNTER_MASK) | flags;
if (__bionic_cmpxchg(old_value, new_value, &cond->value) == 0) {
break;
}
}
// Ensure that all memory accesses previously made by this thread are // We don't use a release/seq_cst fence here. Because pthread_cond_wait/signal can't be
// visible to the woken thread(s). On the other side, the "wait" // used as a method for memory synchronization by itself. It should always be used with
// code will issue any necessary barriers when locking the mutex. // pthread mutexes. Note that Spurious wakeups from pthread_cond_wait/timedwait may occur,
// // so when using condition variables there is always a boolean predicate involving shared
// This may not strictly be necessary -- if the caller follows // variables associated with each condition wait that is true if the thread should proceed.
// recommended practice and holds the mutex before signaling the cond // If the predicate is seen true before a condition wait, pthread_cond_wait/timedwait will
// var, the mutex ops will provide correct semantics. If they don't // not be called. That's why pthread_wait/signal pair can't be used as a method for memory
// hold the mutex, they're subject to race conditions anyway. // synchronization. And it doesn't help even if we use any fence here.
ANDROID_MEMBAR_FULL();
__futex_wake_ex(&cond->value, COND_IS_SHARED(cond->value), counter); // The increase of value should leave flags alone, even if the value can overflows.
atomic_fetch_add_explicit(cond_value_ptr, COND_COUNTER_STEP, memory_order_relaxed);
__futex_wake_ex(cond_value_ptr, shared, thread_count);
return 0; return 0;
} }
__LIBC_HIDDEN__ __LIBC_HIDDEN__
int __pthread_cond_timedwait_relative(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* reltime) { int __pthread_cond_timedwait_relative(atomic_uint* cond_value_ptr, pthread_mutex_t* mutex,
int old_value = cond->value; const timespec* reltime) {
unsigned int old_value = atomic_load_explicit(cond_value_ptr, memory_order_relaxed);
bool shared = COND_IS_SHARED(old_value);
pthread_mutex_unlock(mutex); pthread_mutex_unlock(mutex);
int status = __futex_wait_ex(&cond->value, COND_IS_SHARED(cond->value), old_value, reltime); int status = __futex_wait_ex(cond_value_ptr, shared, old_value, reltime);
pthread_mutex_lock(mutex); pthread_mutex_lock(mutex);
if (status == -ETIMEDOUT) { if (status == -ETIMEDOUT) {
@ -163,7 +172,8 @@ int __pthread_cond_timedwait_relative(pthread_cond_t* cond, pthread_mutex_t* mut
} }
__LIBC_HIDDEN__ __LIBC_HIDDEN__
int __pthread_cond_timedwait(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* abs_ts, clockid_t clock) { int __pthread_cond_timedwait(atomic_uint* cond_value_ptr, pthread_mutex_t* mutex,
const timespec* abs_ts, clockid_t clock) {
timespec ts; timespec ts;
timespec* tsp; timespec* tsp;
@ -176,42 +186,52 @@ int __pthread_cond_timedwait(pthread_cond_t* cond, pthread_mutex_t* mutex, const
tsp = NULL; tsp = NULL;
} }
return __pthread_cond_timedwait_relative(cond, mutex, tsp); return __pthread_cond_timedwait_relative(cond_value_ptr, mutex, tsp);
} }
int pthread_cond_broadcast(pthread_cond_t* cond) { int pthread_cond_broadcast(pthread_cond_t* cond) {
return __pthread_cond_pulse(cond, INT_MAX); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_pulse(cond_value_ptr, INT_MAX);
} }
int pthread_cond_signal(pthread_cond_t* cond) { int pthread_cond_signal(pthread_cond_t* cond) {
return __pthread_cond_pulse(cond, 1); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_pulse(cond_value_ptr, 1);
} }
int pthread_cond_wait(pthread_cond_t* cond, pthread_mutex_t* mutex) { int pthread_cond_wait(pthread_cond_t* cond, pthread_mutex_t* mutex) {
return __pthread_cond_timedwait(cond, mutex, NULL, COND_GET_CLOCK(cond->value)); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_timedwait(cond_value_ptr, mutex, NULL,
COND_GET_CLOCK(atomic_load_explicit(cond_value_ptr, memory_order_relaxed)));
} }
int pthread_cond_timedwait(pthread_cond_t *cond, pthread_mutex_t * mutex, const timespec *abstime) { int pthread_cond_timedwait(pthread_cond_t *cond, pthread_mutex_t * mutex, const timespec *abstime) {
return __pthread_cond_timedwait(cond, mutex, abstime, COND_GET_CLOCK(cond->value)); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_timedwait(cond_value_ptr, mutex, abstime,
COND_GET_CLOCK(atomic_load_explicit(cond_value_ptr, memory_order_relaxed)));
} }
#if !defined(__LP64__) #if !defined(__LP64__)
// TODO: this exists only for backward binary compatibility on 32 bit platforms. // TODO: this exists only for backward binary compatibility on 32 bit platforms.
extern "C" int pthread_cond_timedwait_monotonic(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* abstime) { extern "C" int pthread_cond_timedwait_monotonic(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* abstime) {
return __pthread_cond_timedwait(cond, mutex, abstime, CLOCK_MONOTONIC); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_timedwait(cond_value_ptr, mutex, abstime, CLOCK_MONOTONIC);
} }
extern "C" int pthread_cond_timedwait_monotonic_np(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* abstime) { extern "C" int pthread_cond_timedwait_monotonic_np(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* abstime) {
return __pthread_cond_timedwait(cond, mutex, abstime, CLOCK_MONOTONIC); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_timedwait(cond_value_ptr, mutex, abstime, CLOCK_MONOTONIC);
} }
extern "C" int pthread_cond_timedwait_relative_np(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* reltime) { extern "C" int pthread_cond_timedwait_relative_np(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* reltime) {
return __pthread_cond_timedwait_relative(cond, mutex, reltime); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_timedwait_relative(cond_value_ptr, mutex, reltime);
} }
extern "C" int pthread_cond_timeout_np(pthread_cond_t* cond, pthread_mutex_t* mutex, unsigned ms) { extern "C" int pthread_cond_timeout_np(pthread_cond_t* cond, pthread_mutex_t* mutex, unsigned ms) {
timespec ts; timespec ts;
timespec_from_ms(ts, ms); timespec_from_ms(ts, ms);
return __pthread_cond_timedwait_relative(cond, mutex, &ts); atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond);
return __pthread_cond_timedwait_relative(cond_value_ptr, mutex, &ts);
} }
#endif // !defined(__LP64__) #endif // !defined(__LP64__)

View File

@ -73,7 +73,7 @@ enum {
}; };
typedef struct { typedef struct {
int volatile value; unsigned int value;
#ifdef __LP64__ #ifdef __LP64__
char __reserved[44]; char __reserved[44];
#endif #endif