summaryrefslogtreecommitdiffstats
path: root/libc/bionic/semaphore.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'libc/bionic/semaphore.cpp')
-rw-r--r--libc/bionic/semaphore.cpp310
1 files changed, 310 insertions, 0 deletions
diff --git a/libc/bionic/semaphore.cpp b/libc/bionic/semaphore.cpp
new file mode 100644
index 0000000..0b04650
--- /dev/null
+++ b/libc/bionic/semaphore.cpp
@@ -0,0 +1,310 @@
+/*
+ * Copyright (C) 2008 The Android Open Source Project
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in
+ * the documentation and/or other materials provided with the
+ * distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
+ * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
+ * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
+ * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
+ * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
+ * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
+ * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
+ * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
+ * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
+ * SUCH DAMAGE.
+ */
+
+// Memory order requirements for POSIX semaphores appear unclear and are
+// currently interpreted inconsistently.
+// We conservatively prefer sequentially consistent operations for now.
+// CAUTION: This is more conservative than some other major implementations,
+// and may change if and when the issue is resolved.
+
+#include <semaphore.h>
+#include <errno.h>
+#include <limits.h>
+#include <stdatomic.h>
+#include <sys/time.h>
+#include <time.h>
+
+#include "private/bionic_constants.h"
+#include "private/bionic_futex.h"
+#include "private/bionic_time_conversions.h"
+
+// In this implementation, a semaphore contains a
+// 31-bit signed value and a 1-bit 'shared' flag
+// (for process-sharing purpose).
+//
+// We use the value -1 to indicate contention on the
+// semaphore, 0 or more to indicate uncontended state,
+// any value lower than -2 is invalid at runtime.
+//
+// State diagram:
+//
+// post(1) ==> 2
+// post(0) ==> 1
+// post(-1) ==> 1, then wake all waiters
+//
+// wait(2) ==> 1
+// wait(1) ==> 0
+// wait(0) ==> -1 then wait for a wake up + loop
+// wait(-1) ==> -1 then wait for a wake up + loop
+
+// Use the upper 31-bits for the counter, and the lower one
+// for the shared flag.
+#define SEMCOUNT_SHARED_MASK 0x00000001
+#define SEMCOUNT_VALUE_MASK 0xfffffffe
+#define SEMCOUNT_VALUE_SHIFT 1
+
+// Convert a value into the corresponding sem->count bit pattern.
+#define SEMCOUNT_FROM_VALUE(val) (((val) << SEMCOUNT_VALUE_SHIFT) & SEMCOUNT_VALUE_MASK)
+
+// Convert a sem->count bit pattern into the corresponding signed value.
+static inline int SEMCOUNT_TO_VALUE(unsigned int sval) {
+ return (static_cast<int>(sval) >> SEMCOUNT_VALUE_SHIFT);
+}
+
+// The value +1 as a sem->count bit-pattern.
+#define SEMCOUNT_ONE SEMCOUNT_FROM_VALUE(1)
+
+// The value -1 as a sem->count bit-pattern.
+#define SEMCOUNT_MINUS_ONE SEMCOUNT_FROM_VALUE(-1)
+
+#define SEMCOUNT_DECREMENT(sval) (((sval) - (1U << SEMCOUNT_VALUE_SHIFT)) & SEMCOUNT_VALUE_MASK)
+#define SEMCOUNT_INCREMENT(sval) (((sval) + (1U << SEMCOUNT_VALUE_SHIFT)) & SEMCOUNT_VALUE_MASK)
+
+static inline atomic_uint* SEM_TO_ATOMIC_POINTER(sem_t* sem) {
+ static_assert(sizeof(atomic_uint) == sizeof(sem->count),
+ "sem->count should actually be atomic_uint in implementation.");
+
+ // We prefer casting to atomic_uint instead of declaring sem->count to be atomic_uint directly.
+ // Because using the second method pollutes semaphore.h.
+ return reinterpret_cast<atomic_uint*>(&sem->count);
+}
+
+// Return the shared bitflag from a semaphore counter.
+static inline unsigned int SEM_GET_SHARED(atomic_uint* sem_count_ptr) {
+ // memory_order_relaxed is used as SHARED flag will not be changed after init.
+ return (atomic_load_explicit(sem_count_ptr, memory_order_relaxed) & SEMCOUNT_SHARED_MASK);
+}
+
+int sem_init(sem_t* sem, int pshared, unsigned int value) {
+ // Ensure that 'value' can be stored in the semaphore.
+ if (value > SEM_VALUE_MAX) {
+ errno = EINVAL;
+ return -1;
+ }
+
+ unsigned int count = SEMCOUNT_FROM_VALUE(value);
+ if (pshared != 0) {
+ count |= SEMCOUNT_SHARED_MASK;
+ }
+
+ atomic_uint* sem_count_ptr = SEM_TO_ATOMIC_POINTER(sem);
+ atomic_init(sem_count_ptr, count);
+ return 0;
+}
+
+int sem_destroy(sem_t*) {
+ return 0;
+}
+
+sem_t* sem_open(const char*, int, ...) {
+ errno = ENOSYS;
+ return SEM_FAILED;
+}
+
+int sem_close(sem_t*) {
+ errno = ENOSYS;
+ return -1;
+}
+
+int sem_unlink(const char*) {
+ errno = ENOSYS;
+ return -1;
+}
+
+// Decrement a semaphore's value atomically,
+// and return the old one. As a special case,
+// this returns immediately if the value is
+// negative (i.e. -1)
+static int __sem_dec(atomic_uint* sem_count_ptr) {
+ unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed);
+ unsigned int shared = old_value & SEMCOUNT_SHARED_MASK;
+
+ // Use memory_order_seq_cst in atomic_compare_exchange operation to ensure all
+ // memory access made by other threads can be seen in current thread.
+ // An acquire fence may be sufficient, but it is still in discussion whether
+ // POSIX semaphores should provide sequential consistency.
+ do {
+ if (SEMCOUNT_TO_VALUE(old_value) < 0) {
+ break;
+ }
+ } while (!atomic_compare_exchange_weak(sem_count_ptr, &old_value,
+ SEMCOUNT_DECREMENT(old_value) | shared));
+
+ return SEMCOUNT_TO_VALUE(old_value);
+}
+
+// Same as __sem_dec, but will not touch anything if the
+// value is already negative *or* 0. Returns the old value.
+static int __sem_trydec(atomic_uint* sem_count_ptr) {
+ unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed);
+ unsigned int shared = old_value & SEMCOUNT_SHARED_MASK;
+
+ // Use memory_order_seq_cst in atomic_compare_exchange operation to ensure all
+ // memory access made by other threads can be seen in current thread.
+ // An acquire fence may be sufficient, but it is still in discussion whether
+ // POSIX semaphores should provide sequential consistency.
+ do {
+ if (SEMCOUNT_TO_VALUE(old_value) <= 0) {
+ break;
+ }
+ } while (!atomic_compare_exchange_weak(sem_count_ptr, &old_value,
+ SEMCOUNT_DECREMENT(old_value) | shared));
+
+ return SEMCOUNT_TO_VALUE(old_value);
+}
+
+// "Increment" the value of a semaphore atomically and
+// return its old value. Note that this implements
+// the special case of "incrementing" any negative
+// value to +1 directly.
+//
+// NOTE: The value will _not_ wrap above SEM_VALUE_MAX
+static int __sem_inc(atomic_uint* sem_count_ptr) {
+ unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed);
+ unsigned int shared = old_value & SEMCOUNT_SHARED_MASK;
+ unsigned int new_value;
+
+ // Use memory_order_seq_cst in atomic_compare_exchange operation to ensure all
+ // memory access made before can be seen in other threads.
+ // A release fence may be sufficient, but it is still in discussion whether
+ // POSIX semaphores should provide sequential consistency.
+ do {
+ // Can't go higher than SEM_VALUE_MAX.
+ if (SEMCOUNT_TO_VALUE(old_value) == SEM_VALUE_MAX) {
+ break;
+ }
+
+ // If the counter is negative, go directly to one, otherwise just increment.
+ if (SEMCOUNT_TO_VALUE(old_value) < 0) {
+ new_value = SEMCOUNT_ONE | shared;
+ } else {
+ new_value = SEMCOUNT_INCREMENT(old_value) | shared;
+ }
+ } while (!atomic_compare_exchange_weak(sem_count_ptr, &old_value,
+ new_value));
+
+ return SEMCOUNT_TO_VALUE(old_value);
+}
+
+int sem_wait(sem_t* sem) {
+ atomic_uint* sem_count_ptr = SEM_TO_ATOMIC_POINTER(sem);
+ unsigned int shared = SEM_GET_SHARED(sem_count_ptr);
+
+ while (true) {
+ if (__sem_dec(sem_count_ptr) > 0) {
+ return 0;
+ }
+
+ __futex_wait_ex(sem_count_ptr, shared, shared | SEMCOUNT_MINUS_ONE, NULL);
+ }
+}
+
+int sem_timedwait(sem_t* sem, const timespec* abs_timeout) {
+ atomic_uint* sem_count_ptr = SEM_TO_ATOMIC_POINTER(sem);
+
+ // POSIX says we need to try to decrement the semaphore
+ // before checking the timeout value. Note that if the
+ // value is currently 0, __sem_trydec() does nothing.
+ if (__sem_trydec(sem_count_ptr) > 0) {
+ return 0;
+ }
+
+ // Check it as per POSIX.
+ if (abs_timeout == NULL || abs_timeout->tv_sec < 0 || abs_timeout->tv_nsec < 0 || abs_timeout->tv_nsec >= NS_PER_S) {
+ errno = EINVAL;
+ return -1;
+ }
+
+ unsigned int shared = SEM_GET_SHARED(sem_count_ptr);
+
+ while (true) {
+ // POSIX mandates CLOCK_REALTIME here.
+ timespec ts;
+ if (!timespec_from_absolute_timespec(ts, *abs_timeout, CLOCK_REALTIME)) {
+ errno = ETIMEDOUT;
+ return -1;
+ }
+
+ // Try to grab the semaphore. If the value was 0, this will also change it to -1.
+ if (__sem_dec(sem_count_ptr) > 0) {
+ break;
+ }
+
+ // Contention detected. Wait for a wakeup event.
+ int ret = __futex_wait_ex(sem_count_ptr, shared, shared | SEMCOUNT_MINUS_ONE, &ts);
+
+ // Return in case of timeout or interrupt.
+ if (ret == -ETIMEDOUT || ret == -EINTR) {
+ errno = -ret;
+ return -1;
+ }
+ }
+ return 0;
+}
+
+int sem_post(sem_t* sem) {
+ atomic_uint* sem_count_ptr = SEM_TO_ATOMIC_POINTER(sem);
+ unsigned int shared = SEM_GET_SHARED(sem_count_ptr);
+
+ int old_value = __sem_inc(sem_count_ptr);
+ if (old_value < 0) {
+ // Contention on the semaphore. Wake up all waiters.
+ __futex_wake_ex(sem_count_ptr, shared, INT_MAX);
+ } else if (old_value == SEM_VALUE_MAX) {
+ // Overflow detected.
+ errno = EOVERFLOW;
+ return -1;
+ }
+
+ return 0;
+}
+
+int sem_trywait(sem_t* sem) {
+ atomic_uint* sem_count_ptr = SEM_TO_ATOMIC_POINTER(sem);
+ if (__sem_trydec(sem_count_ptr) > 0) {
+ return 0;
+ } else {
+ errno = EAGAIN;
+ return -1;
+ }
+}
+
+int sem_getvalue(sem_t* sem, int* sval) {
+ atomic_uint* sem_count_ptr = SEM_TO_ATOMIC_POINTER(sem);
+
+ // Use memory_order_seq_cst in atomic_load operation.
+ // memory_order_relaxed may be fine here, but it is still in discussion
+ // whether POSIX semaphores should provide sequential consistency.
+ int val = SEMCOUNT_TO_VALUE(atomic_load(sem_count_ptr));
+ if (val < 0) {
+ val = 0;
+ }
+
+ *sval = val;
+ return 0;
+}