source: mainline/kernel/generic/include/proc/task.h@ 8e8c1a5

lfn serial ticket/834-toolchain-update topic/msim-upgrade topic/simplify-dev-export
Last change on this file since 8e8c1a5 was 8e8c1a5, checked in by Martin Decky <martin@…>, 18 years ago

cleanup garbage

  • Property mode set to 100644
File size: 9.8 KB
Line 
1/*
2 * Copyright (c) 2001-2004 Jakub Jermar
3 * All rights reserved.
4 *
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
7 * are met:
8 *
9 * - Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * - Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 * - The name of the author may not be used to endorse or promote products
15 * derived from this software without specific prior written permission.
16 *
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27 */
28
29/** @addtogroup genericproc
30 * @{
31 */
32/** @file
33 */
34
35#ifndef KERN_TASK_H_
36#define KERN_TASK_H_
37
38#include <synch/spinlock.h>
39#include <synch/mutex.h>
40#include <synch/rwlock.h>
41#include <synch/futex.h>
42#include <adt/btree.h>
43#include <adt/list.h>
44#include <security/cap.h>
45#include <arch/proc/task.h>
46#include <arch/proc/thread.h>
47#include <arch/context.h>
48#include <arch/fpu_context.h>
49#include <arch/cpu.h>
50#include <mm/tlb.h>
51#include <proc/scheduler.h>
52
53#define IPC_MAX_PHONES 16
54#define THREAD_NAME_BUFLEN 20
55
56struct answerbox;
57struct task;
58struct thread;
59
60typedef enum {
61 IPC_PHONE_FREE = 0, /**< Phone is free and can be allocated */
62 IPC_PHONE_CONNECTING, /**< Phone is connecting somewhere */
63 IPC_PHONE_CONNECTED, /**< Phone is connected */
64 IPC_PHONE_HUNGUP, /**< Phone is hung up, waiting for answers to come */
65 IPC_PHONE_SLAMMED /**< Phone was hungup from server */
66} ipc_phone_state_t;
67
68/** Structure identifying phone (in TASK structure) */
69typedef struct {
70 SPINLOCK_DECLARE(lock);
71 link_t link;
72 struct answerbox *callee;
73 ipc_phone_state_t state;
74 atomic_t active_calls;
75} phone_t;
76
77typedef struct answerbox {
78 SPINLOCK_DECLARE(lock);
79
80 struct task *task;
81
82 waitq_t wq;
83
84 link_t connected_phones; /**< Phones connected to this answerbox */
85 link_t calls; /**< Received calls */
86 link_t dispatched_calls; /* Should be hash table in the future */
87
88 link_t answers; /**< Answered calls */
89
90 SPINLOCK_DECLARE(irq_lock);
91 link_t irq_notifs; /**< Notifications from IRQ handlers */
92 link_t irq_head; /**< IRQs with notifications to this answerbox. */
93} answerbox_t;
94
95/** Task structure. */
96typedef struct task {
97 /** Task lock.
98 *
99 * Must be acquired before threads_lock and thread lock of any of its threads.
100 */
101 SPINLOCK_DECLARE(lock);
102
103 char *name;
104 struct thread *main_thread; /**< Pointer to the main thread. */
105 link_t th_head; /**< List of threads contained in this task. */
106 as_t *as; /**< Address space. */
107 task_id_t taskid; /**< Unique identity of task */
108 context_id_t context; /**< Task security context */
109
110 /** If this is true, new threads can become part of the task. */
111 bool accept_new_threads;
112
113 count_t refcount; /**< Number of references (i.e. threads). */
114
115 cap_t capabilities; /**< Task capabilities. */
116
117 /* IPC stuff */
118 answerbox_t answerbox; /**< Communication endpoint */
119 phone_t phones[IPC_MAX_PHONES];
120 atomic_t active_calls; /**< Active asynchronous messages.
121 * It is used for limiting uspace to
122 * certain extent. */
123
124 task_arch_t arch; /**< Architecture specific task data. */
125
126 /**
127 * Serializes access to the B+tree of task's futexes. This mutex is
128 * independent on the task spinlock.
129 */
130 mutex_t futexes_lock;
131 btree_t futexes; /**< B+tree of futexes referenced by this task. */
132
133 uint64_t cycles; /**< Accumulated accounting. */
134} task_t;
135
136/** CPU structure.
137 *
138 * There is one structure like this for every processor.
139 */
140typedef struct {
141 SPINLOCK_DECLARE(lock);
142
143 tlb_shootdown_msg_t tlb_messages[TLB_MESSAGE_QUEUE_LEN];
144 count_t tlb_messages_count;
145
146 context_t saved_context;
147
148 atomic_t nrdy;
149 runq_t rq[RQ_COUNT];
150 volatile count_t needs_relink;
151
152 SPINLOCK_DECLARE(timeoutlock);
153 link_t timeout_active_head;
154
155 count_t missed_clock_ticks; /**< When system clock loses a tick, it is recorded here
156 so that clock() can react. This variable is
157 CPU-local and can be only accessed when interrupts
158 are disabled. */
159
160 /**
161 * Processor ID assigned by kernel.
162 */
163 unsigned int id;
164
165 int active;
166 int tlb_active;
167
168 uint16_t frequency_mhz;
169 uint32_t delay_loop_const;
170
171 cpu_arch_t arch;
172
173 struct thread *fpu_owner;
174
175 /**
176 * Stack used by scheduler when there is no running thread.
177 */
178 uint8_t *stack;
179} cpu_t;
180
181typedef void (* timeout_handler_t)(void *arg);
182
183typedef struct {
184 SPINLOCK_DECLARE(lock);
185
186 link_t link; /**< Link to the list of active timeouts on THE->cpu */
187
188 uint64_t ticks; /**< Timeout will be activated in this amount of clock() ticks. */
189
190 timeout_handler_t handler; /**< Function that will be called on timeout activation. */
191 void *arg; /**< Argument to be passed to handler() function. */
192
193 cpu_t *cpu; /**< On which processor is this timeout registered. */
194} timeout_t;
195
196/** Thread states. */
197typedef enum {
198 Invalid, /**< It is an error, if thread is found in this state. */
199 Running, /**< State of a thread that is currently executing on some CPU. */
200 Sleeping, /**< Thread in this state is waiting for an event. */
201 Ready, /**< State of threads in a run queue. */
202 Entering, /**< Threads are in this state before they are first readied. */
203 Exiting, /**< After a thread calls thread_exit(), it is put into Exiting state. */
204 Undead /**< Threads that were not detached but exited are in the Undead state. */
205} state_t;
206
207/** Join types. */
208typedef enum {
209 None,
210 TaskClnp, /**< The thread will be joined by ktaskclnp thread. */
211 TaskGC /**< The thread will be joined by ktaskgc thread. */
212} thread_join_type_t;
213
214/** Thread structure. There is one per thread. */
215typedef struct thread {
216 link_t rq_link; /**< Run queue link. */
217 link_t wq_link; /**< Wait queue link. */
218 link_t th_link; /**< Links to threads within containing task. */
219
220 /** Lock protecting thread structure.
221 *
222 * Protects the whole thread structure except list links above.
223 */
224 SPINLOCK_DECLARE(lock);
225
226 char name[THREAD_NAME_BUFLEN];
227
228 void (* thread_code)(void *); /**< Function implementing the thread. */
229 void *thread_arg; /**< Argument passed to thread_code() function. */
230
231 /** From here, the stored context is restored when the thread is scheduled. */
232 context_t saved_context;
233 /** From here, the stored timeout context is restored when sleep times out. */
234 context_t sleep_timeout_context;
235 /** From here, the stored interruption context is restored when sleep is interrupted. */
236 context_t sleep_interruption_context;
237
238 bool sleep_interruptible; /**< If true, the thread can be interrupted from sleep. */
239 waitq_t *sleep_queue; /**< Wait queue in which this thread sleeps. */
240 timeout_t sleep_timeout; /**< Timeout used for timeoutable sleeping. */
241 volatile int timeout_pending; /**< Flag signalling sleep timeout in progress. */
242
243 /** True if this thread is executing copy_from_uspace(). False otherwise. */
244 bool in_copy_from_uspace;
245 /** True if this thread is executing copy_to_uspace(). False otherwise. */
246 bool in_copy_to_uspace;
247
248 /**
249 * If true, the thread will not go to sleep at all and will
250 * call thread_exit() before returning to userspace.
251 */
252 bool interrupted;
253
254 thread_join_type_t join_type; /**< Who joinins the thread. */
255 bool detached; /**< If true, thread_join_timeout() cannot be used on this thread. */
256 waitq_t join_wq; /**< Waitq for thread_join_timeout(). */
257
258 fpu_context_t *saved_fpu_context;
259 int fpu_context_exists;
260
261 /*
262 * Defined only if thread doesn't run.
263 * It means that fpu context is in CPU that last time executes this thread.
264 * This disables migration.
265 */
266 int fpu_context_engaged;
267
268 rwlock_type_t rwlock_holder_type;
269
270 void (* call_me)(void *); /**< Funtion to be called in scheduler before the thread is put asleep. */
271 void *call_me_with; /**< Argument passed to call_me(). */
272
273 state_t state; /**< Thread's state. */
274 int flags; /**< Thread's flags. */
275
276 cpu_t *cpu; /**< Thread's CPU. */
277 task_t *task; /**< Containing task. */
278
279 uint64_t ticks; /**< Ticks before preemption. */
280
281 uint64_t cycles; /**< Thread accounting. */
282 uint64_t last_cycle; /**< Last sampled cycle. */
283 bool uncounted; /**< Thread doesn't affect accumulated accounting. */
284
285 int priority; /**< Thread's priority. Implemented as index to CPU->rq */
286 uint32_t tid; /**< Thread ID. */
287
288 thread_arch_t arch; /**< Architecture-specific data. */
289
290 uint8_t *kstack; /**< Thread's kernel stack. */
291} thread_t;
292
293extern spinlock_t tasks_lock;
294extern btree_t tasks_btree;
295
296extern void task_init(void);
297extern task_t *task_create(as_t *as, char *name);
298extern void task_destroy(task_t *t);
299extern task_t *task_run_program(void *program_addr, char *name);
300extern task_t *task_find_by_id(task_id_t id);
301extern int task_kill(task_id_t id);
302extern uint64_t task_get_accounting(task_t *t);
303
304extern void cap_set(task_t *t, cap_t caps);
305extern cap_t cap_get(task_t *t);
306
307
308#ifndef task_create_arch
309extern void task_create_arch(task_t *t);
310#endif
311
312#ifndef task_destroy_arch
313extern void task_destroy_arch(task_t *t);
314#endif
315
316extern unative_t sys_task_get_id(task_id_t *uspace_task_id);
317
318#endif
319
320/** @}
321 */
Note: See TracBrowser for help on using the repository browser.