source: mainline/kernel/generic/src/synch/condvar.c@ a36f442

lfn serial ticket/834-toolchain-update topic/msim-upgrade topic/simplify-dev-export
Last change on this file since a36f442 was a35b458, checked in by Jiří Zárevúcky <zarevucky.jiri@…>, 8 years ago

style: Remove trailing whitespace on _all_ lines, including empty ones, for particular file types.

Command used: tools/srepl '\s\+$' '' -- *.c *.h *.py *.sh *.s *.S *.ag

Currently, whitespace on empty lines is very inconsistent.
There are two basic choices: Either remove the whitespace, or keep empty lines
indented to the level of surrounding code. The former is AFAICT more common,
and also much easier to do automatically.

Alternatively, we could write script for automatic indentation, and use that
instead. However, if such a script exists, it's possible to use the indented
style locally, by having the editor apply relevant conversions on load/save,
without affecting remote repository. IMO, it makes more sense to adopt
the simpler rule.

  • Property mode set to 100644
File size: 5.8 KB
Line 
1/*
2 * Copyright (c) 2001-2004 Jakub Jermar
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 *
9 * - Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * - Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 * - The name of the author may not be used to endorse or promote products
15 * derived from this software without specific prior written permission.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 */
28
29/** @addtogroup sync
30 * @{
31 */
32
33/**
34 * @file
35 * @brief Condition variables.
36 */
37
38#include <synch/condvar.h>
39#include <synch/mutex.h>
40#include <synch/spinlock.h>
41#include <synch/waitq.h>
42#include <arch.h>
43
44/** Initialize condition variable.
45 *
46 * @param cv Condition variable.
47 */
48void condvar_initialize(condvar_t *cv)
49{
50 waitq_initialize(&cv->wq);
51}
52
53/** Signal the condition has become true to the first waiting thread by waking
54 * it up.
55 *
56 * @param cv Condition variable.
57 */
58void condvar_signal(condvar_t *cv)
59{
60 waitq_wakeup(&cv->wq, WAKEUP_FIRST);
61}
62
63/** Signal the condition has become true to all waiting threads by waking
64 * them up.
65 *
66 * @param cv Condition variable.
67 */
68void condvar_broadcast(condvar_t *cv)
69{
70 waitq_wakeup(&cv->wq, WAKEUP_ALL);
71}
72
73/** Wait for the condition becoming true.
74 *
75 * @param cv Condition variable.
76 * @param mtx Mutex.
77 * @param usec Timeout value in microseconds.
78 * @param flags Select mode of operation.
79 *
80 * For exact description of meaning of possible combinations of usec and flags,
81 * see comment for waitq_sleep_timeout(). Note that when
82 * SYNCH_FLAGS_NON_BLOCKING is specified here, EAGAIN is always
83 * returned.
84 *
85 * @return See comment for waitq_sleep_timeout().
86 */
87errno_t _condvar_wait_timeout(condvar_t *cv, mutex_t *mtx, uint32_t usec, int flags)
88{
89 errno_t rc;
90 ipl_t ipl;
91 bool blocked;
92
93 ipl = waitq_sleep_prepare(&cv->wq);
94 /* Unlock only after the waitq is locked so we don't miss a wakeup. */
95 mutex_unlock(mtx);
96
97 cv->wq.missed_wakeups = 0; /* Enforce blocking. */
98 rc = waitq_sleep_timeout_unsafe(&cv->wq, usec, flags, &blocked);
99 assert(blocked || rc != EOK);
100
101 waitq_sleep_finish(&cv->wq, blocked, ipl);
102 /* Lock only after releasing the waitq to avoid a possible deadlock. */
103 mutex_lock(mtx);
104
105 return rc;
106}
107
108/** Wait for the condition to become true with a locked spinlock.
109 *
110 * The function is not aware of irq_spinlock. Therefore do not even
111 * try passing irq_spinlock_t to it. Use _condvar_wait_timeout_irq_spinlock()
112 * instead.
113 *
114 * @param cv Condition variable.
115 * @param lock Locked spinlock.
116 * @param usec Timeout value in microseconds.
117 * @param flags Select mode of operation.
118 *
119 * For exact description of meaning of possible combinations of usec and flags,
120 * see comment for waitq_sleep_timeout(). Note that when
121 * SYNCH_FLAGS_NON_BLOCKING is specified here, EAGAIN is always
122 * returned.
123 *
124 * @return See comment for waitq_sleep_timeout().
125 */
126errno_t _condvar_wait_timeout_spinlock_impl(condvar_t *cv, spinlock_t *lock,
127 uint32_t usec, int flags)
128{
129 errno_t rc;
130 ipl_t ipl;
131 bool blocked;
132
133 ipl = waitq_sleep_prepare(&cv->wq);
134
135 /* Unlock only after the waitq is locked so we don't miss a wakeup. */
136 spinlock_unlock(lock);
137
138 cv->wq.missed_wakeups = 0; /* Enforce blocking. */
139 rc = waitq_sleep_timeout_unsafe(&cv->wq, usec, flags, &blocked);
140 assert(blocked || rc != EOK);
141
142 waitq_sleep_finish(&cv->wq, blocked, ipl);
143 /* Lock only after releasing the waitq to avoid a possible deadlock. */
144 spinlock_lock(lock);
145
146 return rc;
147}
148
149/** Wait for the condition to become true with a locked irq spinlock.
150 *
151 * @param cv Condition variable.
152 * @param lock Locked irq spinlock.
153 * @param usec Timeout value in microseconds.
154 * @param flags Select mode of operation.
155 *
156 * For exact description of meaning of possible combinations of usec and flags,
157 * see comment for waitq_sleep_timeout(). Note that when
158 * SYNCH_FLAGS_NON_BLOCKING is specified here, EAGAIN is always
159 * returned.
160 *
161 * @return See comment for waitq_sleep_timeout().
162 */
163errno_t _condvar_wait_timeout_irq_spinlock(condvar_t *cv, irq_spinlock_t *irq_lock,
164 uint32_t usec, int flags)
165{
166 errno_t rc;
167 /* Save spinlock's state so we can restore it correctly later on. */
168 ipl_t ipl = irq_lock->ipl;
169 bool guard = irq_lock->guard;
170
171 irq_lock->guard = false;
172
173 /*
174 * waitq_prepare() restores interrupts to the current state,
175 * ie disabled. Therefore, interrupts will remain disabled while
176 * it spins waiting for a pending timeout handler to complete.
177 * Although it spins with interrupts disabled there can only
178 * be a pending timeout if we failed to cancel an imminent
179 * timeout (on another cpu) during a wakeup. As a result the
180 * timeout handler is guaranteed to run (it is most likely already
181 * running) and there is no danger of a deadlock.
182 */
183 rc = _condvar_wait_timeout_spinlock(cv, &irq_lock->lock, usec, flags);
184
185 irq_lock->guard = guard;
186 irq_lock->ipl = ipl;
187
188 return rc;
189}
190
191
192/** @}
193 */
Note: See TracBrowser for help on using the repository browser.