source: mainline/kernel/generic/src/synch/condvar.c

Last change on this file was 0b47781, checked in by Jiří Zárevúcky <zarevucky.jiri@…>, 3 months ago

Extend kernel condvar_t to work with spinlocks

MUTEX_ACTIVE is a poor copy of the spinlock, made solely to allow
using condvars with it. Solve that issue the other way around,
by making condvar_wait() generic, so that we can get rid of that
hack.

  • Property mode set to 100644
File size: 5.1 KB
RevLine 
[f761f1eb]1/*
[df4ed85]2 * Copyright (c) 2001-2004 Jakub Jermar
[f761f1eb]3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 *
9 * - Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * - Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 * - The name of the author may not be used to endorse or promote products
15 * derived from this software without specific prior written permission.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 */
28
[e88eb48]29/** @addtogroup kernel_sync
[b45c443]30 * @{
31 */
32
[cf26ba9]33/**
[b45c443]34 * @file
[cf26ba9]35 * @brief Condition variables.
36 */
37
[f761f1eb]38#include <synch/condvar.h>
39#include <synch/mutex.h>
[46a5b37]40#include <synch/spinlock.h>
[f761f1eb]41#include <synch/waitq.h>
[c0bc189]42#include <arch.h>
[f761f1eb]43
[c0bc189]44/** Initialize condition variable.
[df364582]45 *
[08a19ba]46 * @param cv Condition variable.
[df364582]47 */
[f761f1eb]48void condvar_initialize(condvar_t *cv)
49{
[597fa24]50 *cv = CONDVAR_INITIALIZER(*cv);
[f761f1eb]51}
52
[08a19ba]53/** Signal the condition has become true to the first waiting thread by waking
54 * it up.
[df364582]55 *
[08a19ba]56 * @param cv Condition variable.
[df364582]57 */
[f761f1eb]58void condvar_signal(condvar_t *cv)
59{
[111b9b9]60 waitq_signal(&cv->wq);
[f761f1eb]61}
62
[08a19ba]63/** Signal the condition has become true to all waiting threads by waking
64 * them up.
[df364582]65 *
[08a19ba]66 * @param cv Condition variable.
[df364582]67 */
[f761f1eb]68void condvar_broadcast(condvar_t *cv)
69{
[111b9b9]70 waitq_wake_all(&cv->wq);
[f761f1eb]71}
72
[c0bc189]73/** Wait for the condition becoming true.
[df364582]74 *
[08a19ba]75 * @param cv Condition variable.
76 * @param mtx Mutex.
77 * @param usec Timeout value in microseconds.
[a783ca4]78 *
[08a19ba]79 * @return See comment for waitq_sleep_timeout().
[df364582]80 */
[0b47781]81errno_t __condvar_wait_timeout_mutex(condvar_t *cv, mutex_t *mtx, uint32_t usec)
[5110d0a]82{
[111b9b9]83 wait_guard_t guard = waitq_sleep_prepare(&cv->wq);
[5110d0a]84
85 /* Unlock only after the waitq is locked so we don't miss a wakeup. */
86 mutex_unlock(mtx);
87
[111b9b9]88 errno_t rc = waitq_sleep_timeout_unsafe(&cv->wq, usec, SYNCH_FLAGS_NON_BLOCKING, guard);
[5110d0a]89
90 mutex_lock(mtx);
91 return rc;
92}
93
[0b47781]94errno_t __condvar_wait_mutex(condvar_t *cv, mutex_t *mtx)
[f761f1eb]95{
[111b9b9]96 wait_guard_t guard = waitq_sleep_prepare(&cv->wq);
[f761f1eb]97
[497bd656]98 /* Unlock only after the waitq is locked so we don't miss a wakeup. */
[f761f1eb]99 mutex_unlock(mtx);
[baafe71]100
[111b9b9]101 errno_t rc = waitq_sleep_unsafe(&cv->wq, guard);
[c0bc189]102
[497bd656]103 mutex_lock(mtx);
[f761f1eb]104 return rc;
105}
[b45c443]106
[0b47781]107/** Same as __condvar_wait_timeout_mutex(), except for spinlock_t. */
108errno_t __condvar_wait_timeout_spinlock(condvar_t *cv, spinlock_t *lock,
109 uint32_t usec)
[46a5b37]110{
[111b9b9]111 wait_guard_t guard = waitq_sleep_prepare(&cv->wq);
[46a5b37]112
[497bd656]113 /* Unlock only after the waitq is locked so we don't miss a wakeup. */
[46a5b37]114 spinlock_unlock(lock);
115
[0b47781]116 errno_t rc = waitq_sleep_timeout_unsafe(&cv->wq, usec,
117 SYNCH_FLAGS_NON_BLOCKING, guard);
[46a5b37]118
119 spinlock_lock(lock);
120 return rc;
121}
122
[0b47781]123errno_t __condvar_wait_spinlock(condvar_t *cv, spinlock_t *mtx)
124{
125 wait_guard_t guard = waitq_sleep_prepare(&cv->wq);
126
127 /* Unlock only after the waitq is locked so we don't miss a wakeup. */
128 spinlock_unlock(mtx);
129
130 errno_t rc = waitq_sleep_unsafe(&cv->wq, guard);
131
132 spinlock_lock(mtx);
133 return rc;
134}
135
136/** Same as __condvar_wait_timeout_mutex(), except for irq_spinlock_t. */
137errno_t __condvar_wait_timeout_irq_spinlock(condvar_t *cv,
138 irq_spinlock_t *irq_lock, uint32_t usec)
[46a5b37]139{
[b7fd2a0]140 errno_t rc;
[46a5b37]141 /* Save spinlock's state so we can restore it correctly later on. */
142 ipl_t ipl = irq_lock->ipl;
143 bool guard = irq_lock->guard;
[a35b458]144
[46a5b37]145 irq_lock->guard = false;
[a35b458]146
[1b20da0]147 /*
148 * waitq_prepare() restores interrupts to the current state,
149 * ie disabled. Therefore, interrupts will remain disabled while
150 * it spins waiting for a pending timeout handler to complete.
[46a5b37]151 * Although it spins with interrupts disabled there can only
152 * be a pending timeout if we failed to cancel an imminent
[1b20da0]153 * timeout (on another cpu) during a wakeup. As a result the
154 * timeout handler is guaranteed to run (it is most likely already
[46a5b37]155 * running) and there is no danger of a deadlock.
156 */
[0b47781]157 rc = __condvar_wait_timeout_spinlock(cv, &irq_lock->lock, usec);
158
159 irq_lock->guard = guard;
160 irq_lock->ipl = ipl;
161
162 return rc;
163}
164
165/** Same as __condvar_wait_mutex(), except for irq_spinlock_t. */
166errno_t __condvar_wait_irq_spinlock(condvar_t *cv, irq_spinlock_t *irq_lock)
167{
168 errno_t rc;
169 /* Save spinlock's state so we can restore it correctly later on. */
170 ipl_t ipl = irq_lock->ipl;
171 bool guard = irq_lock->guard;
172
173 irq_lock->guard = false;
174
175 rc = __condvar_wait_spinlock(cv, &irq_lock->lock);
[a35b458]176
[46a5b37]177 irq_lock->guard = guard;
178 irq_lock->ipl = ipl;
[a35b458]179
[46a5b37]180 return rc;
181}
182
[cc73a8a1]183/** @}
[b45c443]184 */
Note: See TracBrowser for help on using the repository browser.