posix: pthread: replace irq_lock with spinlock

We shouldn't use swapping with an interrupt lock held
as it works incorrectly on SMP platforms.

Fix that by replacing irq_lock with spinlock for pthread
subsystem.

NOTE: we fix that in a simple way with single spinlock
for mutex / cond_var / barrier. That could be improved
later (i.e. split it for several spinlocks).

Signed-off-by: Eugeniy Paltsev <Eugeniy.Paltsev@synopsys.com>
Signed-off-by: Evgeniy Paltsev <PaltsevEvgeniy@gmail.com>
This commit is contained in:
Evgeniy Paltsev 2021-08-06 22:58:20 +03:00 committed by Anas Nashif
commit 4e0f7ea540
3 changed files with 27 additions and 19 deletions

View file

@ -9,9 +9,11 @@
#include <ksched.h> #include <ksched.h>
#include <wait_q.h> #include <wait_q.h>
extern struct k_spinlock z_pthread_spinlock;
int pthread_barrier_wait(pthread_barrier_t *b) int pthread_barrier_wait(pthread_barrier_t *b)
{ {
unsigned int key = irq_lock(); k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
int ret = 0; int ret = 0;
b->count++; b->count++;
@ -22,10 +24,10 @@ int pthread_barrier_wait(pthread_barrier_t *b)
while (z_waitq_head(&b->wait_q)) { while (z_waitq_head(&b->wait_q)) {
_ready_one_thread(&b->wait_q); _ready_one_thread(&b->wait_q);
} }
z_reschedule_irqlock(key); z_reschedule(&z_pthread_spinlock, key);
ret = PTHREAD_BARRIER_SERIAL_THREAD; ret = PTHREAD_BARRIER_SERIAL_THREAD;
} else { } else {
(void) z_pend_curr_irqlock(key, &b->wait_q, K_FOREVER); (void) z_pend_curr(&z_pthread_spinlock, key, &b->wait_q, K_FOREVER);
} }
return ret; return ret;

View file

@ -9,6 +9,8 @@
#include <wait_q.h> #include <wait_q.h>
#include <posix/pthread.h> #include <posix/pthread.h>
extern struct k_spinlock z_pthread_spinlock;
int64_t timespec_to_timeoutms(const struct timespec *abstime); int64_t timespec_to_timeoutms(const struct timespec *abstime);
static int cond_wait(pthread_cond_t *cv, pthread_mutex_t *mut, static int cond_wait(pthread_cond_t *cv, pthread_mutex_t *mut,
@ -16,12 +18,13 @@ static int cond_wait(pthread_cond_t *cv, pthread_mutex_t *mut,
{ {
__ASSERT(mut->lock_count == 1U, ""); __ASSERT(mut->lock_count == 1U, "");
int ret, key = irq_lock(); int ret;
k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
mut->lock_count = 0U; mut->lock_count = 0U;
mut->owner = NULL; mut->owner = NULL;
_ready_one_thread(&mut->wait_q); _ready_one_thread(&mut->wait_q);
ret = z_pend_curr_irqlock(key, &cv->wait_q, timeout); ret = z_pend_curr(&z_pthread_spinlock, key, &cv->wait_q, timeout);
/* FIXME: this extra lock (and the potential context switch it /* FIXME: this extra lock (and the potential context switch it
* can cause) could be optimized out. At the point of the * can cause) could be optimized out. At the point of the
@ -49,23 +52,23 @@ static int cond_wait(pthread_cond_t *cv, pthread_mutex_t *mut,
int pthread_cond_signal(pthread_cond_t *cv) int pthread_cond_signal(pthread_cond_t *cv)
{ {
int key = irq_lock(); k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
_ready_one_thread(&cv->wait_q); _ready_one_thread(&cv->wait_q);
z_reschedule_irqlock(key); z_reschedule(&z_pthread_spinlock, key);
return 0; return 0;
} }
int pthread_cond_broadcast(pthread_cond_t *cv) int pthread_cond_broadcast(pthread_cond_t *cv)
{ {
int key = irq_lock(); k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
while (z_waitq_head(&cv->wait_q)) { while (z_waitq_head(&cv->wait_q)) {
_ready_one_thread(&cv->wait_q); _ready_one_thread(&cv->wait_q);
} }
z_reschedule_irqlock(key); z_reschedule(&z_pthread_spinlock, key);
return 0; return 0;
} }

View file

@ -9,6 +9,8 @@
#include <wait_q.h> #include <wait_q.h>
#include <posix/pthread.h> #include <posix/pthread.h>
struct k_spinlock z_pthread_spinlock;
int64_t timespec_to_timeoutms(const struct timespec *abstime); int64_t timespec_to_timeoutms(const struct timespec *abstime);
#define MUTEX_MAX_REC_LOCK 32767 #define MUTEX_MAX_REC_LOCK 32767
@ -22,13 +24,14 @@ static const pthread_mutexattr_t def_attr = {
static int acquire_mutex(pthread_mutex_t *m, k_timeout_t timeout) static int acquire_mutex(pthread_mutex_t *m, k_timeout_t timeout)
{ {
int rc = 0, key = irq_lock(); int rc = 0;
k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
if (m->lock_count == 0U && m->owner == NULL) { if (m->lock_count == 0U && m->owner == NULL) {
m->lock_count++; m->lock_count++;
m->owner = pthread_self(); m->owner = pthread_self();
irq_unlock(key); k_spin_unlock(&z_pthread_spinlock, key);
return 0; return 0;
} else if (m->owner == pthread_self()) { } else if (m->owner == pthread_self()) {
if (m->type == PTHREAD_MUTEX_RECURSIVE && if (m->type == PTHREAD_MUTEX_RECURSIVE &&
@ -41,16 +44,16 @@ static int acquire_mutex(pthread_mutex_t *m, k_timeout_t timeout)
rc = EINVAL; rc = EINVAL;
} }
irq_unlock(key); k_spin_unlock(&z_pthread_spinlock, key);
return rc; return rc;
} }
if (K_TIMEOUT_EQ(timeout, K_NO_WAIT)) { if (K_TIMEOUT_EQ(timeout, K_NO_WAIT)) {
irq_unlock(key); k_spin_unlock(&z_pthread_spinlock, key);
return EINVAL; return EINVAL;
} }
rc = z_pend_curr_irqlock(key, &m->wait_q, timeout); rc = z_pend_curr(&z_pthread_spinlock, key, &m->wait_q, timeout);
if (rc != 0) { if (rc != 0) {
rc = ETIMEDOUT; rc = ETIMEDOUT;
} }
@ -121,17 +124,17 @@ int pthread_mutex_lock(pthread_mutex_t *m)
*/ */
int pthread_mutex_unlock(pthread_mutex_t *m) int pthread_mutex_unlock(pthread_mutex_t *m)
{ {
unsigned int key = irq_lock(); k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
k_tid_t thread; k_tid_t thread;
if (m->owner != pthread_self()) { if (m->owner != pthread_self()) {
irq_unlock(key); k_spin_unlock(&z_pthread_spinlock, key);
return EPERM; return EPERM;
} }
if (m->lock_count == 0U) { if (m->lock_count == 0U) {
irq_unlock(key); k_spin_unlock(&z_pthread_spinlock, key);
return EINVAL; return EINVAL;
} }
@ -144,13 +147,13 @@ int pthread_mutex_unlock(pthread_mutex_t *m)
m->lock_count++; m->lock_count++;
arch_thread_return_value_set(thread, 0); arch_thread_return_value_set(thread, 0);
z_ready_thread(thread); z_ready_thread(thread);
z_reschedule_irqlock(key); z_reschedule(&z_pthread_spinlock, key);
return 0; return 0;
} }
m->owner = NULL; m->owner = NULL;
} }
irq_unlock(key); k_spin_unlock(&z_pthread_spinlock, key);
return 0; return 0;
} }