mirror of
https://github.com/jart/cosmopolitan.git
synced 2025-01-31 19:43:32 +00:00
69f4152f38
We had previously not enabled TLS in MODE=tiny in order to keep the smallest example programs (e.g. life.com) just 16kb in size. But it was error prone doing that, so now we just always enable it because this change uses hacks to ensure it won't increase life.com's size. This change also fixes a bug on NetBSD, where signal handlers would break thread local storage if SA_SIGINFO was being used. This looks like it might be a bug in NetBSD, but it's got a simple workaround.
133 lines
4.7 KiB
C
133 lines
4.7 KiB
C
/*-*- mode:c;indent-tabs-mode:nil;c-basic-offset:2;tab-width:8;coding:utf-8 -*-│
|
|
│vi: set net ft=c ts=2 sts=2 sw=2 fenc=utf-8 :vi│
|
|
╞══════════════════════════════════════════════════════════════════════════════╡
|
|
│ Copyright 2020 Justine Alexandra Roberts Tunney │
|
|
│ │
|
|
│ Permission to use, copy, modify, and/or distribute this software for │
|
|
│ any purpose with or without fee is hereby granted, provided that the │
|
|
│ above copyright notice and this permission notice appear in all copies. │
|
|
│ │
|
|
│ THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL │
|
|
│ WARRANTIES WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED │
|
|
│ WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE │
|
|
│ AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL │
|
|
│ DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR │
|
|
│ PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER │
|
|
│ TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR │
|
|
│ PERFORMANCE OF THIS SOFTWARE. │
|
|
╚─────────────────────────────────────────────────────────────────────────────*/
|
|
#include "libc/bits/atomic.h"
|
|
#include "libc/calls/calls.h"
|
|
#include "libc/thread/thread.h"
|
|
|
|
#define CTHREAD_THREAD_VAL_BITS 32
|
|
|
|
static void Pause(int attempt) {
|
|
if (attempt < 16) {
|
|
for (int i = 0; i < (1 << attempt); ++i) {
|
|
__builtin_ia32_pause();
|
|
}
|
|
} else {
|
|
sched_yield();
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Initializes semaphore.
|
|
*/
|
|
int cthread_sem_init(cthread_sem_t* sem, int count) {
|
|
sem->linux.count = count;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Destroys semaphore.
|
|
*/
|
|
int cthread_sem_destroy(cthread_sem_t* sem) {
|
|
(void)sem;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Notifies a thread waiting on semaphore.
|
|
*/
|
|
int cthread_sem_signal(cthread_sem_t* sem) {
|
|
uint64_t count;
|
|
count = atomic_fetch_add(&sem->linux.count, 1);
|
|
if ((count >> CTHREAD_THREAD_VAL_BITS)) {
|
|
// WARNING: an offset of 4 bytes would be required on little-endian archs
|
|
void* wait_address = &sem->linux.count;
|
|
cthread_memory_wake32(wait_address, 1);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Waits on semaphore with kernel assistance.
|
|
*/
|
|
int cthread_sem_wait_futex(cthread_sem_t* sem, const struct timespec* timeout) {
|
|
uint64_t count;
|
|
|
|
// record current thread as waiter
|
|
count = atomic_fetch_add(&sem->linux.count,
|
|
(uint64_t)1 << CTHREAD_THREAD_VAL_BITS);
|
|
|
|
for (;;) {
|
|
// try to acquire the semaphore, as well as remove itself from waiters
|
|
while ((uint32_t)count > 0) {
|
|
// without spin, we could miss a futex wake
|
|
if (atomic_compare_exchange_weak(
|
|
&sem->linux.count, &count,
|
|
count - 1 - ((uint64_t)1 << CTHREAD_THREAD_VAL_BITS))) {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
// WARNING: an offset of 4 bytes would be required on little-endian archs
|
|
void* wait_address = &sem->linux.count;
|
|
cthread_memory_wait32(wait_address, count, timeout);
|
|
count = atomic_load(&sem->linux.count);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Waits on semaphore without kernel assistance.
|
|
*/
|
|
int cthread_sem_wait_spin(cthread_sem_t* sem, uint64_t count, int spin,
|
|
const struct timespec* timeout) {
|
|
// spin on pause
|
|
for (int attempt = 0; attempt < spin; ++attempt) {
|
|
// if ((count >> CTHREAD_THREAD_VAL_BITS) != 0) break;
|
|
while ((uint32_t)count > 0) {
|
|
// spin is useful if multiple waiters can acquire the semaphore at the
|
|
// same time
|
|
if (atomic_compare_exchange_weak(&sem->linux.count, &count, count - 1)) {
|
|
return 0;
|
|
}
|
|
}
|
|
Pause(attempt);
|
|
}
|
|
|
|
return cthread_sem_wait_futex(sem, timeout);
|
|
}
|
|
|
|
/**
|
|
* Waits on semaphore.
|
|
*/
|
|
int cthread_sem_wait(cthread_sem_t* sem, int spin,
|
|
const struct timespec* timeout) {
|
|
uint64_t count = atomic_load(&sem->linux.count);
|
|
|
|
// uncontended
|
|
while ((uint32_t)count > 0) {
|
|
// spin is useful if multiple waiters can acquire the semaphore at the same
|
|
// time
|
|
if (atomic_compare_exchange_weak(&sem->linux.count, &count, count - 1)) {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return cthread_sem_wait_spin(sem, count, spin, timeout);
|
|
}
|