mirror of
https://github.com/jart/cosmopolitan.git
synced 2025-02-14 10:18:02 +00:00
Recursive mutexes now go as fast as normal mutexes. The tradeoff is they are no longer safe to use in signal handlers. However you can still have signal safe mutexes if you set your mutex to both recursive and pshared. You can also make functions that use recursive mutexes signal safe using sigprocmask to ensure recursion doesn't happen due to any signal handler The impact of this change is that, on Windows, many functions which edit the file descriptor table rely on recursive mutexes, e.g. open(). If you develop your app so it uses pread() and pwrite() then your app should go very fast when performing a heavily multithreaded and contended workload For example, when scaling to 40+ cores, *NSYNC mutexes can go as much as 1000x faster (in CPU time) than the naive recursive lock implementation. Now recursive will use *NSYNC under the hood when it's possible to do so
243 lines
8.8 KiB
C
243 lines
8.8 KiB
C
/*-*- mode:c;indent-tabs-mode:nil;c-basic-offset:2;tab-width:8;coding:utf-8 -*-│
|
|
│ vi: set et ft=c ts=2 sts=2 sw=2 fenc=utf-8 :vi │
|
|
╞══════════════════════════════════════════════════════════════════════════════╡
|
|
│ Copyright 2022 Justine Alexandra Roberts Tunney │
|
|
│ │
|
|
│ Permission to use, copy, modify, and/or distribute this software for │
|
|
│ any purpose with or without fee is hereby granted, provided that the │
|
|
│ above copyright notice and this permission notice appear in all copies. │
|
|
│ │
|
|
│ THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL │
|
|
│ WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED │
|
|
│ WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE │
|
|
│ AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL │
|
|
│ DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR │
|
|
│ PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER │
|
|
│ TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR │
|
|
│ PERFORMANCE OF THIS SOFTWARE. │
|
|
╚─────────────────────────────────────────────────────────────────────────────*/
|
|
#include "libc/calls/blockcancel.internal.h"
|
|
#include "libc/calls/calls.h"
|
|
#include "libc/calls/state.internal.h"
|
|
#include "libc/dce.h"
|
|
#include "libc/errno.h"
|
|
#include "libc/intrin/atomic.h"
|
|
#include "libc/intrin/describeflags.h"
|
|
#include "libc/intrin/strace.h"
|
|
#include "libc/intrin/weaken.h"
|
|
#include "libc/runtime/internal.h"
|
|
#include "libc/thread/lock.h"
|
|
#include "libc/thread/thread.h"
|
|
#include "third_party/nsync/futex.internal.h"
|
|
#include "third_party/nsync/mu.h"
|
|
|
|
static void pthread_mutex_lock_spin(atomic_int *word) {
|
|
int backoff = 0;
|
|
if (atomic_exchange_explicit(word, 1, memory_order_acquire)) {
|
|
LOCKTRACE("acquiring pthread_mutex_lock_spin(%t)...", word);
|
|
for (;;) {
|
|
for (;;) {
|
|
if (!atomic_load_explicit(word, memory_order_relaxed))
|
|
break;
|
|
backoff = pthread_delay_np(word, backoff);
|
|
}
|
|
if (!atomic_exchange_explicit(word, 1, memory_order_acquire))
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
// see "take 3" algorithm in "futexes are tricky" by ulrich drepper
|
|
// slightly improved to attempt acquiring multiple times b4 syscall
|
|
static void pthread_mutex_lock_drepper(atomic_int *futex, char pshare) {
|
|
int word = 0;
|
|
if (atomic_compare_exchange_strong_explicit(
|
|
futex, &word, 1, memory_order_acquire, memory_order_acquire))
|
|
return;
|
|
LOCKTRACE("acquiring pthread_mutex_lock_drepper(%t)...", futex);
|
|
if (word == 1)
|
|
word = atomic_exchange_explicit(futex, 2, memory_order_acquire);
|
|
BLOCK_CANCELATION;
|
|
while (word > 0) {
|
|
_weaken(nsync_futex_wait_)(futex, 2, pshare, 0, 0);
|
|
word = atomic_exchange_explicit(futex, 2, memory_order_acquire);
|
|
}
|
|
ALLOW_CANCELATION;
|
|
}
|
|
|
|
static errno_t pthread_mutex_lock_recursive(pthread_mutex_t *mutex,
|
|
uint64_t word) {
|
|
uint64_t lock;
|
|
int backoff = 0;
|
|
int me = gettid();
|
|
bool once = false;
|
|
for (;;) {
|
|
if (MUTEX_OWNER(word) == me) {
|
|
if (MUTEX_TYPE(word) != PTHREAD_MUTEX_ERRORCHECK) {
|
|
if (MUTEX_DEPTH(word) < MUTEX_DEPTH_MAX) {
|
|
if (atomic_compare_exchange_weak_explicit(
|
|
&mutex->_word, &word, MUTEX_INC_DEPTH(word),
|
|
memory_order_relaxed, memory_order_relaxed))
|
|
return 0;
|
|
continue;
|
|
} else {
|
|
return EAGAIN;
|
|
}
|
|
} else {
|
|
return EDEADLK;
|
|
}
|
|
}
|
|
word = MUTEX_UNLOCK(word);
|
|
lock = MUTEX_LOCK(word);
|
|
lock = MUTEX_SET_OWNER(lock, me);
|
|
if (atomic_compare_exchange_weak_explicit(&mutex->_word, &word, lock,
|
|
memory_order_acquire,
|
|
memory_order_relaxed)) {
|
|
mutex->_pid = __pid;
|
|
return 0;
|
|
}
|
|
if (!once) {
|
|
LOCKTRACE("acquiring pthread_mutex_lock_recursive(%t)...", mutex);
|
|
once = true;
|
|
}
|
|
for (;;) {
|
|
word = atomic_load_explicit(&mutex->_word, memory_order_relaxed);
|
|
if (MUTEX_OWNER(word) == me)
|
|
break;
|
|
if (word == MUTEX_UNLOCK(word))
|
|
break;
|
|
backoff = pthread_delay_np(mutex, backoff);
|
|
}
|
|
}
|
|
}
|
|
|
|
#if PTHREAD_USE_NSYNC
|
|
static errno_t pthread_mutex_lock_recursive_nsync(pthread_mutex_t *mutex,
|
|
uint64_t word) {
|
|
int me = gettid();
|
|
for (;;) {
|
|
if (MUTEX_OWNER(word) == me) {
|
|
if (MUTEX_TYPE(word) != PTHREAD_MUTEX_ERRORCHECK) {
|
|
if (MUTEX_DEPTH(word) < MUTEX_DEPTH_MAX) {
|
|
if (atomic_compare_exchange_weak_explicit(
|
|
&mutex->_word, &word, MUTEX_INC_DEPTH(word),
|
|
memory_order_relaxed, memory_order_relaxed))
|
|
return 0;
|
|
continue;
|
|
} else {
|
|
return EAGAIN;
|
|
}
|
|
} else {
|
|
return EDEADLK;
|
|
}
|
|
}
|
|
_weaken(nsync_mu_lock)((nsync_mu *)mutex->_nsyncx);
|
|
word = MUTEX_UNLOCK(word);
|
|
word = MUTEX_LOCK(word);
|
|
word = MUTEX_SET_OWNER(word, me);
|
|
mutex->_word = word;
|
|
mutex->_pid = __pid;
|
|
return 0;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static errno_t pthread_mutex_lock_impl(pthread_mutex_t *mutex) {
|
|
uint64_t word;
|
|
|
|
// get current state of lock
|
|
word = atomic_load_explicit(&mutex->_word, memory_order_relaxed);
|
|
|
|
#if PTHREAD_USE_NSYNC
|
|
// use superior mutexes if possible
|
|
if (MUTEX_TYPE(word) == PTHREAD_MUTEX_NORMAL && //
|
|
MUTEX_PSHARED(word) == PTHREAD_PROCESS_PRIVATE && //
|
|
_weaken(nsync_mu_lock)) {
|
|
// on apple silicon we should just put our faith in ulock
|
|
// otherwise *nsync gets struck down by the eye of sauron
|
|
if (!IsXnuSilicon()) {
|
|
_weaken(nsync_mu_lock)((nsync_mu *)mutex);
|
|
return 0;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
// handle normal mutexes
|
|
if (MUTEX_TYPE(word) == PTHREAD_MUTEX_NORMAL) {
|
|
if (_weaken(nsync_futex_wait_)) {
|
|
pthread_mutex_lock_drepper(&mutex->_futex, MUTEX_PSHARED(word));
|
|
} else {
|
|
pthread_mutex_lock_spin(&mutex->_futex);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
// handle recursive and error checking mutexes
|
|
#if PTHREAD_USE_NSYNC
|
|
if (_weaken(nsync_mu_lock) &&
|
|
MUTEX_PSHARED(word) == PTHREAD_PROCESS_PRIVATE) {
|
|
return pthread_mutex_lock_recursive_nsync(mutex, word);
|
|
} else {
|
|
return pthread_mutex_lock_recursive(mutex, word);
|
|
}
|
|
#else
|
|
return pthread_mutex_lock_recursive(mutex, word);
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* Locks mutex.
|
|
*
|
|
* Here's an example of using a normal mutex:
|
|
*
|
|
* pthread_mutex_t lock = PTHREAD_MUTEX_INITIALIZER;
|
|
* pthread_mutex_lock(&lock);
|
|
* // do work...
|
|
* pthread_mutex_unlock(&lock);
|
|
* pthread_mutex_destroy(&lock);
|
|
*
|
|
* Cosmopolitan permits succinct notation for normal mutexes:
|
|
*
|
|
* pthread_mutex_t lock = {0};
|
|
* pthread_mutex_lock(&lock);
|
|
* // do work...
|
|
* pthread_mutex_unlock(&lock);
|
|
*
|
|
* Here's an example of the proper way to do recursive mutexes:
|
|
*
|
|
* pthread_mutex_t lock;
|
|
* pthread_mutexattr_t attr;
|
|
* pthread_mutexattr_init(&attr);
|
|
* pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_RECURSIVE);
|
|
* pthread_mutex_init(&lock, &attr);
|
|
* pthread_mutexattr_destroy(&attr);
|
|
* pthread_mutex_lock(&lock);
|
|
* // do work...
|
|
* pthread_mutex_unlock(&lock);
|
|
* pthread_mutex_destroy(&lock);
|
|
*
|
|
* This function does nothing in vfork() children.
|
|
*
|
|
* You can debug locks the acquisition of locks by building your program
|
|
* with `cosmocc -mdbg` and passing the `--strace` flag to your program.
|
|
* This will cause a line to be logged each time a mutex or spin lock is
|
|
* locked or unlocked. When locking, this is printed after the lock gets
|
|
* acquired. The entry to the lock operation will be logged too but only
|
|
* if the lock couldn't be immediately acquired. Lock logging works best
|
|
* when `mutex` refers to a static variable, in which case its name will
|
|
* be printed in the log.
|
|
*
|
|
* @return 0 on success, or error number on failure
|
|
* @see pthread_spin_lock()
|
|
* @vforksafe
|
|
*/
|
|
errno_t pthread_mutex_lock(pthread_mutex_t *mutex) {
|
|
if (!__vforked) {
|
|
errno_t err = pthread_mutex_lock_impl(mutex);
|
|
LOCKTRACE("pthread_mutex_lock(%t) → %s", mutex, DescribeErrno(err));
|
|
return err;
|
|
} else {
|
|
LOCKTRACE("skipping pthread_mutex_lock(%t) due to vfork", mutex);
|
|
return 0;
|
|
}
|
|
}
|