2 * Copyright (c) 1995-1998 John Birrell <jb@cimlogic.com.au>.
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
13 * 3. All advertising materials mentioning features or use of this software
14 * must display the following acknowledgement:
15 * This product includes software developed by John Birrell.
16 * 4. Neither the name of the author nor the names of any co-contributors
17 * may be used to endorse or promote products derived from this software
18 * without specific prior written permission.
20 * THIS SOFTWARE IS PROVIDED BY JOHN BIRRELL AND CONTRIBUTORS ``AS IS'' AND
21 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
24 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
25 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
26 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
27 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
28 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
29 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 * Private thread definitions for the uthread kernel.
37 #ifndef _THR_PRIVATE_H
38 #define _THR_PRIVATE_H
46 #include <sys/queue.h>
47 #include <sys/types.h>
49 #include <sys/cdefs.h>
55 #include <pthread_np.h>
59 #include "pthread_md.h"
63 * Evaluate the storage class specifier.
65 #ifdef GLOBAL_PTHREAD_PRIVATE
67 #define SCLASS_PRESET(x...) = x
70 #define SCLASS_PRESET(x...)
74 * Kernel fatal error handler macro.
76 #define PANIC(string) _thr_exit(__FILE__,__LINE__,string)
79 /* Output debug messages like this: */
80 #define stdout_debug(args...) _thread_printf(STDOUT_FILENO, ##args)
81 #define stderr_debug(args...) _thread_printf(STDOUT_FILENO, ##args)
83 #define DBG_MUTEX 0x0001
84 #define DBG_SIG 0x0002
86 #ifdef _PTHREADS_INVARIANTS
87 #define THR_ASSERT(cond, msg) do { \
92 #define THR_ASSERT(cond, msg)
96 * State change macro without scheduling queue change:
98 #define THR_SET_STATE(thrd, newstate) do { \
99 (thrd)->state = newstate; \
100 (thrd)->fname = __FILE__; \
101 (thrd)->lineno = __LINE__; \
105 #define TIMESPEC_ADD(dst, src, val) \
107 (dst)->tv_sec = (src)->tv_sec + (val)->tv_sec; \
108 (dst)->tv_nsec = (src)->tv_nsec + (val)->tv_nsec; \
109 if ((dst)->tv_nsec > 1000000000) { \
111 (dst)->tv_nsec -= 1000000000; \
115 #define TIMESPEC_SUB(dst, src, val) \
117 (dst)->tv_sec = (src)->tv_sec - (val)->tv_sec; \
118 (dst)->tv_nsec = (src)->tv_nsec - (val)->tv_nsec; \
119 if ((dst)->tv_nsec < 0) { \
121 (dst)->tv_nsec += 1000000000; \
128 * XXX It'd be nice if these were contained in uthread_priority_queue.[ch].
130 typedef struct pq_list {
131 TAILQ_HEAD(, pthread) pl_head; /* list of threads at this priority */
132 TAILQ_ENTRY(pq_list) pl_link; /* link for queue of priority lists */
133 int pl_prio; /* the priority of this list */
134 int pl_queued; /* is this in the priority queue */
137 typedef struct pq_queue {
138 TAILQ_HEAD(, pq_list) pq_queue; /* queue of priority lists */
139 pq_list_t *pq_lists; /* array of all priority lists */
140 int pq_size; /* number of priority lists */
141 #define PQF_ACTIVE 0x0001
147 * Each KSEG has a scheduling queue. For now, threads that exist in their
148 * own KSEG (system scope) will get a full priority queue. In the future
149 * this can be optimized for the single thread per KSEG case.
153 TAILQ_HEAD(, pthread) sq_waitq; /* waiting in userland */
156 typedef struct kse_thr_mailbox *kse_critical_t;
160 #define MAX_KSE_LOCKLEVEL 5
162 /* -- location and order specific items for gdb -- */
164 struct pthread *k_curthread; /* current thread */
165 struct kse_group *k_kseg; /* parent KSEG */
166 struct sched_queue *k_schedq; /* scheduling queue */
167 /* -- end of location and order specific items -- */
168 TAILQ_ENTRY(kse) k_qe; /* KSE list link entry */
169 TAILQ_ENTRY(kse) k_kgqe; /* KSEG's KSE list entry */
171 * Items that are only modified by the kse, or that otherwise
172 * don't need to be locked when accessed
175 struct lockuser k_lockusers[MAX_KSE_LOCKLEVEL];
179 #define KF_STARTED 0x0001 /* kernel kse created */
180 #define KF_INITIALIZED 0x0002 /* initialized on 1st upcall */
181 #define KF_TERMINATED 0x0004 /* kse is terminated */
182 #define KF_IDLE 0x0008 /* kse is idle */
183 #define KF_SWITCH 0x0010 /* thread switch in UTS */
184 int k_error; /* syscall errno in critical */
185 int k_cpu; /* CPU ID when bound */
186 int k_sigseqno; /* signal buffered count */
189 #define KSE_SET_IDLE(kse) ((kse)->k_flags |= KF_IDLE)
190 #define KSE_CLEAR_IDLE(kse) ((kse)->k_flags &= ~KF_IDLE)
191 #define KSE_IS_IDLE(kse) (((kse)->k_flags & KF_IDLE) != 0)
192 #define KSE_SET_SWITCH(kse) ((kse)->k_flags |= KF_SWITCH)
193 #define KSE_CLEAR_SWITCH(kse) ((kse)->k_flags &= ~KF_SWITCH)
194 #define KSE_IS_SWITCH(kse) (((kse)->k_flags & KF_SWITCH) != 0)
197 * Each KSE group contains one or more KSEs in which threads can run.
198 * At least for now, there is one scheduling queue per KSE group; KSEs
199 * within the same KSE group compete for threads from the same scheduling
200 * queue. A scope system thread has one KSE in one KSE group; the group
201 * does not use its scheduling queue.
204 TAILQ_HEAD(, kse) kg_kseq; /* list of KSEs in group */
205 TAILQ_HEAD(, pthread) kg_threadq; /* list of threads in group */
206 TAILQ_ENTRY(kse_group) kg_qe; /* link entry */
207 struct sched_queue kg_schedq; /* scheduling queue */
209 int kg_threadcount; /* # of assigned threads */
210 int kg_ksecount; /* # of assigned KSEs */
213 #define KGF_SINGLE_THREAD 0x0001 /* scope system kse group */
214 #define KGF_SCHEDQ_INITED 0x0002 /* has an initialized schedq */
218 * Add/remove threads from a KSE's scheduling queue.
219 * For now the scheduling queue is hung off the KSEG.
221 #define KSEG_THRQ_ADD(kseg, thr) \
223 TAILQ_INSERT_TAIL(&(kseg)->kg_threadq, thr, kle);\
224 (kseg)->kg_threadcount++; \
227 #define KSEG_THRQ_REMOVE(kseg, thr) \
229 TAILQ_REMOVE(&(kseg)->kg_threadq, thr, kle); \
230 (kseg)->kg_threadcount--; \
235 * Lock acquire and release for KSEs.
237 #define KSE_LOCK_ACQUIRE(kse, lck) \
239 if ((kse)->k_locklevel < MAX_KSE_LOCKLEVEL) { \
240 (kse)->k_locklevel++; \
241 _lock_acquire((lck), \
242 &(kse)->k_lockusers[(kse)->k_locklevel - 1], 0); \
245 PANIC("Exceeded maximum lock level"); \
248 #define KSE_LOCK_RELEASE(kse, lck) \
250 if ((kse)->k_locklevel > 0) { \
251 _lock_release((lck), \
252 &(kse)->k_lockusers[(kse)->k_locklevel - 1]); \
253 (kse)->k_locklevel--; \
260 #define KSE_LOCK(curkse) \
261 KSE_LOCK_ACQUIRE(curkse, &(curkse)->k_kseg->kg_lock)
262 #define KSE_UNLOCK(curkse) \
263 KSE_LOCK_RELEASE(curkse, &(curkse)->k_kseg->kg_lock)
266 * Lock a potentially different KSEG.
268 #define KSE_SCHED_LOCK(curkse, kseg) \
269 KSE_LOCK_ACQUIRE(curkse, &(kseg)->kg_lock)
270 #define KSE_SCHED_UNLOCK(curkse, kseg) \
271 KSE_LOCK_RELEASE(curkse, &(kseg)->kg_lock)
274 * Waiting queue manipulation macros (using pqe link):
276 #define KSE_WAITQ_REMOVE(kse, thrd) \
278 if (((thrd)->flags & THR_FLAGS_IN_WAITQ) != 0) { \
279 TAILQ_REMOVE(&(kse)->k_schedq->sq_waitq, thrd, pqe); \
280 (thrd)->flags &= ~THR_FLAGS_IN_WAITQ; \
283 #define KSE_WAITQ_INSERT(kse, thrd) kse_waitq_insert(thrd)
284 #define KSE_WAITQ_FIRST(kse) TAILQ_FIRST(&(kse)->k_schedq->sq_waitq)
286 #define KSE_WAKEUP(kse) kse_wakeup(&(kse)->k_kcb->kcb_kmbx)
289 * TailQ initialization values.
291 #define TAILQ_INITIALIZER { NULL, NULL }
294 * lock initialization values.
296 #define LCK_INITIALIZER { NULL, NULL, LCK_DEFAULT }
298 struct pthread_mutex {
300 * Lock for accesses to this structure.
303 enum pthread_mutextype m_type;
305 TAILQ_HEAD(mutex_head, pthread) m_queue;
306 struct pthread *m_owner;
312 * Used for priority inheritence and protection.
314 * m_prio - For priority inheritence, the highest active
315 * priority (threads locking the mutex inherit
316 * this priority). For priority protection, the
317 * ceiling priority of this mutex.
318 * m_saved_prio - mutex owners inherited priority before
319 * taking the mutex, restored when the owner
326 * Link for list of all mutexes a thread currently owns.
328 TAILQ_ENTRY(pthread_mutex) m_qe;
334 #define MUTEX_FLAGS_PRIVATE 0x01
335 #define MUTEX_FLAGS_INITED 0x02
336 #define MUTEX_FLAGS_BUSY 0x04
339 * Static mutex initialization values.
341 #define PTHREAD_MUTEX_STATIC_INITIALIZER \
342 { LCK_INITIALIZER, PTHREAD_MUTEX_DEFAULT, PTHREAD_PRIO_NONE, \
343 TAILQ_INITIALIZER, NULL, MUTEX_FLAGS_PRIVATE, 0, 0, 0, 0, \
346 struct pthread_mutex_attr {
347 enum pthread_mutextype m_type;
353 #define PTHREAD_MUTEXATTR_STATIC_INITIALIZER \
354 { PTHREAD_MUTEX_DEFAULT, PTHREAD_PRIO_NONE, 0, MUTEX_FLAGS_PRIVATE }
357 * Condition variable definitions.
359 enum pthread_cond_type {
364 struct pthread_cond {
366 * Lock for accesses to this structure.
369 enum pthread_cond_type c_type;
370 TAILQ_HEAD(cond_head, pthread) c_queue;
371 struct pthread_mutex *c_mutex;
376 struct pthread_cond_attr {
377 enum pthread_cond_type c_type;
381 struct pthread_barrier {
382 pthread_mutex_t b_lock;
383 pthread_cond_t b_cond;
389 struct pthread_barrierattr {
393 struct pthread_spinlock {
399 * Flags for condition variables.
401 #define COND_FLAGS_PRIVATE 0x01
402 #define COND_FLAGS_INITED 0x02
403 #define COND_FLAGS_BUSY 0x04
406 * Static cond initialization values.
408 #define PTHREAD_COND_STATIC_INITIALIZER \
409 { LCK_INITIALIZER, COND_TYPE_FAST, TAILQ_INITIALIZER, \
413 * Cleanup definitions.
415 struct pthread_cleanup {
416 struct pthread_cleanup *next;
422 #define THR_CLEANUP_PUSH(td, func, arg) { \
423 struct pthread_cleanup __cup; \
425 __cup.routine = func; \
426 __cup.routine_arg = arg; \
428 __cup.next = (td)->cleanup; \
429 (td)->cleanup = &__cup;
431 #define THR_CLEANUP_POP(td, exec) \
432 (td)->cleanup = __cup.next; \
434 __cup.routine(__cup.routine_arg); \
437 struct pthread_atfork {
438 TAILQ_ENTRY(pthread_atfork) qe;
439 void (*prepare)(void);
440 void (*parent)(void);
444 struct pthread_attr {
450 #define THR_STACK_USER 0x100 /* 0xFF reserved for <pthread.h> */
451 #define THR_SIGNAL_THREAD 0x200 /* This is a signal thread */
454 void (*cleanup_attr) ();
455 void *stackaddr_attr;
456 size_t stacksize_attr;
457 size_t guardsize_attr;
461 * Thread creation state attributes.
463 #define THR_CREATE_RUNNING 0
464 #define THR_CREATE_SUSPENDED 1
467 * Miscellaneous definitions.
469 #define THR_STACK_DEFAULT 65536
472 * Maximum size of initial thread's stack. This perhaps deserves to be larger
473 * than the stacks of other threads, since many applications are likely to run
474 * almost entirely on this stack.
476 #define THR_STACK_INITIAL 0x100000
479 * Define the different priority ranges. All applications have thread
480 * priorities constrained within 0-31. The threads library raises the
481 * priority when delivering signals in order to ensure that signal
482 * delivery happens (from the POSIX spec) "as soon as possible".
483 * In the future, the threads library will also be able to map specific
484 * threads into real-time (cooperating) processes or kernel threads.
485 * The RT and SIGNAL priorities will be used internally and added to
486 * thread base priorities so that the scheduling queue can handle both
487 * normal and RT priority threads with and without signal handling.
489 * The approach taken is that, within each class, signal delivery
490 * always has priority over thread execution.
492 #define THR_DEFAULT_PRIORITY 15
493 #define THR_MIN_PRIORITY 0
494 #define THR_MAX_PRIORITY 31 /* 0x1F */
495 #define THR_SIGNAL_PRIORITY 32 /* 0x20 */
496 #define THR_RT_PRIORITY 64 /* 0x40 */
497 #define THR_FIRST_PRIORITY THR_MIN_PRIORITY
498 #define THR_LAST_PRIORITY \
499 (THR_MAX_PRIORITY + THR_SIGNAL_PRIORITY + THR_RT_PRIORITY)
500 #define THR_BASE_PRIORITY(prio) ((prio) & THR_MAX_PRIORITY)
503 * Clock resolution in microseconds.
505 #define CLOCK_RES_USEC 10000
508 * Time slice period in microseconds.
510 #define TIMESLICE_USEC 20000
513 * XXX - Define a thread-safe macro to get the current time of day
514 * which is updated at regular intervals by something.
516 * For now, we just make the system call to get the time.
518 #define KSE_GET_TOD(curkse, tsp) \
520 *tsp = (curkse)->k_kcb->kcb_kmbx.km_timeofday; \
521 if ((tsp)->tv_sec == 0) \
522 clock_gettime(CLOCK_REALTIME, tsp); \
525 struct pthread_rwlockattr {
529 struct pthread_rwlock {
530 pthread_mutex_t lock; /* monitor lock */
531 pthread_cond_t read_signal;
532 pthread_cond_t write_signal;
533 int state; /* 0 = idle >0 = # of readers -1 = writer */
555 struct sigwait_data {
557 siginfo_t *siginfo; /* used to save siginfo for sigwaitinfo() */
560 union pthread_wait_data {
561 pthread_mutex_t mutex;
564 struct sigwait_data *sigwait;
568 * Define a continuation routine that can be used to perform a
569 * transfer of control:
571 typedef void (*thread_continuation_t) (void *);
574 * This stores a thread's state prior to running a signal handler.
575 * It is used when a signal is delivered to a thread blocked in
576 * userland. If the signal handler returns normally, the thread's
577 * state is restored from here.
579 struct pthread_sigframe {
586 enum pthread_state psf_state;
587 union pthread_wait_data psf_wait_data;
588 struct timespec psf_wakeup_time;
590 sigset_t psf_sigmask;
592 thread_continuation_t psf_continuation;
596 struct pthread *thread;
601 struct pthread_specific_elem {
607 volatile int allocated;
610 void (*destructor) (void *);
613 #define MAX_THR_LOCKLEVEL 5
618 /* Thread control block */
622 * Magic value to help recognize a valid thread structure
623 * from an invalid one:
625 #define THR_MAGIC ((u_int32_t) 0xd09ba115)
628 u_int64_t uniqueid; /* for gdb */
630 /* Queue entry for list of all threads: */
631 TAILQ_ENTRY(pthread) tle; /* link for all threads in process */
632 TAILQ_ENTRY(pthread) kle; /* link for all threads in KSE/KSEG */
634 /* Queue entry for GC lists: */
635 TAILQ_ENTRY(pthread) gcle;
637 /* Hash queue entry */
638 LIST_ENTRY(pthread) hle;
641 * Lock for accesses to this thread structure.
644 struct lockuser lockusers[MAX_THR_LOCKLEVEL];
646 kse_critical_t critical[MAX_KSE_LOCKLEVEL];
648 struct kse_group *kseg;
651 * Thread start routine, argument, stack pointer and thread
654 void *(*start_routine)(void *);
656 struct pthread_attr attr;
658 int active; /* thread running */
659 int blocked; /* thread blocked in kernel */
663 * Used for tracking delivery of signal handlers.
666 thread_continuation_t sigbackout;
669 * Cancelability flags - the lower 2 bits are used by cancel
670 * definitions in pthread.h
672 #define THR_AT_CANCEL_POINT 0x0004
673 #define THR_CANCELLING 0x0008
674 #define THR_CANCEL_NEEDED 0x0010
677 thread_continuation_t continuation;
680 * The thread's base and pending signal masks. The active
681 * signal mask is stored in the thread's context (in mailbox).
685 sigset_t *oldsigmask;
686 volatile int check_pending;
690 enum pthread_state state;
691 volatile int lock_switch;
694 * Number of microseconds accumulated by this thread when
695 * time slicing is active.
700 * Time to wake up thread. This is used for sleeping threads and
701 * for any operation which may time out (such as select).
703 struct timespec wakeup_time;
705 /* TRUE if operation has timed out. */
709 * Error variable used instead of errno. The function __error()
710 * returns a pointer to this.
715 * The joiner is the thread that is joining to this thread. The
716 * join status keeps track of a join operation to another thread.
718 struct pthread *joiner;
719 struct join_status join_status;
722 * The current thread can belong to only one scheduling queue at
723 * a time (ready or waiting queue). It can also belong to:
725 * o A queue of threads waiting for a mutex
726 * o A queue of threads waiting for a condition variable
728 * It is possible for a thread to belong to more than one of the
729 * above queues if it is handling a signal. A thread may only
730 * enter a mutex or condition variable queue when it is not
731 * being called from a signal handler. If a thread is a member
732 * of one of these queues when a signal handler is invoked, it
733 * must be removed from the queue before invoking the handler
734 * and then added back to the queue after return from the handler.
736 * Use pqe for the scheduling queue link (both ready and waiting),
737 * sqe for synchronization (mutex, condition variable, and join)
738 * queue links, and qe for all other links.
740 TAILQ_ENTRY(pthread) pqe; /* priority, wait queues link */
741 TAILQ_ENTRY(pthread) sqe; /* synchronization queue link */
744 union pthread_wait_data data;
747 * Set to TRUE if a blocking operation was
748 * interrupted by a signal:
753 * Set to non-zero when this thread has entered a critical
754 * region. We allow for recursive entries into critical regions.
759 * Set to TRUE if this thread should yield after leaving a
760 * critical region to check for signals, messages, etc.
765 #define THR_FLAGS_IN_SYNCQ 0x0001
767 /* Miscellaneous flags; only set with scheduling lock held. */
769 #define THR_FLAGS_PRIVATE 0x0001
770 #define THR_FLAGS_IN_WAITQ 0x0002 /* in waiting queue using pqe link */
771 #define THR_FLAGS_IN_RUNQ 0x0004 /* in run queue using pqe link */
772 #define THR_FLAGS_EXITING 0x0008 /* thread is exiting */
773 #define THR_FLAGS_SUSPENDED 0x0010 /* thread is suspended */
775 /* Thread list flags; only set with thread list lock held. */
776 #define TLFLAGS_GC_SAFE 0x0001 /* thread safe for cleaning */
777 #define TLFLAGS_IN_TDLIST 0x0002 /* thread in all thread list */
778 #define TLFLAGS_IN_GCLIST 0x0004 /* thread in gc list */
782 * Base priority is the user setable and retrievable priority
783 * of the thread. It is only affected by explicit calls to
784 * set thread priority and upon thread creation via a thread
785 * attribute or default priority.
790 * Inherited priority is the priority a thread inherits by
791 * taking a priority inheritence or protection mutex. It
792 * is not affected by base priority changes. Inherited
793 * priority defaults to and remains 0 until a mutex is taken
794 * that is being waited on by any other thread whose priority
797 char inherited_priority;
800 * Active priority is always the maximum of the threads base
801 * priority and inherited priority. When there is a change
802 * in either the base or inherited priority, the active
803 * priority must be recalculated.
805 char active_priority;
807 /* Number of priority ceiling or protection mutexes owned. */
808 int priority_mutex_count;
810 /* Number rwlocks rdlocks held. */
814 * Queue of currently owned mutexes.
816 TAILQ_HEAD(, pthread_mutex) mutexq;
819 struct pthread_specific_elem *specific;
820 int specific_data_count;
822 /* Alternative stack for sigaltstack() */
826 * Current locks bitmap for rtld.
830 /* Cleanup handlers Link List */
831 struct pthread_cleanup *cleanup;
832 char *fname; /* Ptr to source file name */
833 int lineno; /* Source line number. */
837 * Critical regions can also be detected by looking at the threads
838 * current lock level. Ensure these macros increment and decrement
839 * the lock levels such that locks can not be held with a lock level
842 #define THR_IN_CRITICAL(thrd) \
843 (((thrd)->locklevel > 0) || \
844 ((thrd)->critical_count > 0))
846 #define THR_YIELD_CHECK(thrd) \
848 if (!THR_IN_CRITICAL(thrd)) { \
849 if (__predict_false(_libkse_debug)) \
850 _thr_debug_check_yield(thrd); \
851 if ((thrd)->critical_yield != 0) \
852 _thr_sched_switch(thrd); \
853 if ((thrd)->check_pending != 0) \
854 _thr_sig_check_pending(thrd); \
858 #define THR_LOCK_ACQUIRE(thrd, lck) \
860 if ((thrd)->locklevel < MAX_THR_LOCKLEVEL) { \
861 THR_DEACTIVATE_LAST_LOCK(thrd); \
862 (thrd)->locklevel++; \
863 _lock_acquire((lck), \
864 &(thrd)->lockusers[(thrd)->locklevel - 1], \
865 (thrd)->active_priority); \
867 PANIC("Exceeded maximum lock level"); \
870 #define THR_LOCK_RELEASE(thrd, lck) \
872 if ((thrd)->locklevel > 0) { \
873 _lock_release((lck), \
874 &(thrd)->lockusers[(thrd)->locklevel - 1]); \
875 (thrd)->locklevel--; \
876 THR_ACTIVATE_LAST_LOCK(thrd); \
877 if ((thrd)->locklevel == 0) \
878 THR_YIELD_CHECK(thrd); \
882 #define THR_ACTIVATE_LAST_LOCK(thrd) \
884 if ((thrd)->locklevel > 0) \
885 _lockuser_setactive( \
886 &(thrd)->lockusers[(thrd)->locklevel - 1], 1); \
889 #define THR_DEACTIVATE_LAST_LOCK(thrd) \
891 if ((thrd)->locklevel > 0) \
892 _lockuser_setactive( \
893 &(thrd)->lockusers[(thrd)->locklevel - 1], 0); \
897 * For now, threads will have their own lock separate from their
898 * KSE scheduling lock.
900 #define THR_LOCK(thr) THR_LOCK_ACQUIRE(thr, &(thr)->lock)
901 #define THR_UNLOCK(thr) THR_LOCK_RELEASE(thr, &(thr)->lock)
902 #define THR_THREAD_LOCK(curthrd, thr) THR_LOCK_ACQUIRE(curthrd, &(thr)->lock)
903 #define THR_THREAD_UNLOCK(curthrd, thr) THR_LOCK_RELEASE(curthrd, &(thr)->lock)
906 * Priority queue manipulation macros (using pqe link). We use
907 * the thread's kseg link instead of the kse link because a thread
908 * does not (currently) have a statically assigned kse.
910 #define THR_RUNQ_INSERT_HEAD(thrd) \
911 _pq_insert_head(&(thrd)->kseg->kg_schedq.sq_runq, thrd)
912 #define THR_RUNQ_INSERT_TAIL(thrd) \
913 _pq_insert_tail(&(thrd)->kseg->kg_schedq.sq_runq, thrd)
914 #define THR_RUNQ_REMOVE(thrd) \
915 _pq_remove(&(thrd)->kseg->kg_schedq.sq_runq, thrd)
918 * Macros to insert/remove threads to the all thread list and
921 #define THR_LIST_ADD(thrd) do { \
922 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) == 0) { \
923 TAILQ_INSERT_HEAD(&_thread_list, thrd, tle); \
924 _thr_hash_add(thrd); \
925 (thrd)->tlflags |= TLFLAGS_IN_TDLIST; \
928 #define THR_LIST_REMOVE(thrd) do { \
929 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) != 0) { \
930 TAILQ_REMOVE(&_thread_list, thrd, tle); \
931 _thr_hash_remove(thrd); \
932 (thrd)->tlflags &= ~TLFLAGS_IN_TDLIST; \
935 #define THR_GCLIST_ADD(thrd) do { \
936 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) == 0) { \
937 TAILQ_INSERT_HEAD(&_thread_gc_list, thrd, gcle);\
938 (thrd)->tlflags |= TLFLAGS_IN_GCLIST; \
942 #define THR_GCLIST_REMOVE(thrd) do { \
943 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) != 0) { \
944 TAILQ_REMOVE(&_thread_gc_list, thrd, gcle); \
945 (thrd)->tlflags &= ~TLFLAGS_IN_GCLIST; \
950 #define GC_NEEDED() (atomic_load_acq_int(&_gc_count) >= 5)
953 * Locking the scheduling queue for another thread uses that thread's
956 #define THR_SCHED_LOCK(curthr, thr) do { \
957 (curthr)->critical[(curthr)->locklevel] = _kse_critical_enter(); \
958 (curthr)->locklevel++; \
959 KSE_SCHED_LOCK((curthr)->kse, (thr)->kseg); \
962 #define THR_SCHED_UNLOCK(curthr, thr) do { \
963 KSE_SCHED_UNLOCK((curthr)->kse, (thr)->kseg); \
964 (curthr)->locklevel--; \
965 _kse_critical_leave((curthr)->critical[(curthr)->locklevel]); \
968 /* Take the scheduling lock with the intent to call the scheduler. */
969 #define THR_LOCK_SWITCH(curthr) do { \
970 (void)_kse_critical_enter(); \
971 KSE_SCHED_LOCK((curthr)->kse, (curthr)->kseg); \
973 #define THR_UNLOCK_SWITCH(curthr) do { \
974 KSE_SCHED_UNLOCK((curthr)->kse, (curthr)->kseg);\
977 #define THR_CRITICAL_ENTER(thr) (thr)->critical_count++
978 #define THR_CRITICAL_LEAVE(thr) do { \
979 (thr)->critical_count--; \
980 if (((thr)->critical_yield != 0) && \
981 ((thr)->critical_count == 0)) { \
982 (thr)->critical_yield = 0; \
983 _thr_sched_switch(thr); \
987 #define THR_IS_ACTIVE(thrd) \
988 ((thrd)->kse != NULL) && ((thrd)->kse->k_curthread == (thrd))
990 #define THR_IN_SYNCQ(thrd) (((thrd)->sflags & THR_FLAGS_IN_SYNCQ) != 0)
992 #define THR_IS_SUSPENDED(thrd) \
993 (((thrd)->state == PS_SUSPENDED) || \
994 (((thrd)->flags & THR_FLAGS_SUSPENDED) != 0))
995 #define THR_IS_EXITING(thrd) (((thrd)->flags & THR_FLAGS_EXITING) != 0)
996 #define DBG_CAN_RUN(thrd) (((thrd)->tcb->tcb_tmbx.tm_dflags & \
999 extern int __isthreaded;
1002 _kse_isthreaded(void)
1004 return (__isthreaded != 0);
1008 * Global variables for the pthread kernel.
1011 SCLASS void *_usrstack SCLASS_PRESET(NULL);
1012 SCLASS struct kse *_kse_initial SCLASS_PRESET(NULL);
1013 SCLASS struct pthread *_thr_initial SCLASS_PRESET(NULL);
1015 SCLASS int _libkse_debug SCLASS_PRESET(0);
1016 SCLASS int _thread_activated SCLASS_PRESET(0);
1017 SCLASS int _thread_scope_system SCLASS_PRESET(0);
1019 /* List of all threads: */
1020 SCLASS TAILQ_HEAD(, pthread) _thread_list
1021 SCLASS_PRESET(TAILQ_HEAD_INITIALIZER(_thread_list));
1023 /* List of threads needing GC: */
1024 SCLASS TAILQ_HEAD(, pthread) _thread_gc_list
1025 SCLASS_PRESET(TAILQ_HEAD_INITIALIZER(_thread_gc_list));
1027 SCLASS int _thread_active_threads SCLASS_PRESET(1);
1029 SCLASS TAILQ_HEAD(atfork_head, pthread_atfork) _thr_atfork_list;
1030 SCLASS pthread_mutex_t _thr_atfork_mutex;
1032 /* Default thread attributes: */
1033 SCLASS struct pthread_attr _pthread_attr_default
1035 SCHED_RR, 0, TIMESLICE_USEC, THR_DEFAULT_PRIORITY,
1036 THR_CREATE_RUNNING, PTHREAD_CREATE_JOINABLE, NULL,
1037 NULL, NULL, THR_STACK_DEFAULT, /* guardsize */0
1040 /* Default mutex attributes: */
1041 SCLASS struct pthread_mutex_attr _pthread_mutexattr_default
1042 SCLASS_PRESET({PTHREAD_MUTEX_DEFAULT, PTHREAD_PRIO_NONE, 0, 0 });
1044 /* Default condition variable attributes: */
1045 SCLASS struct pthread_cond_attr _pthread_condattr_default
1046 SCLASS_PRESET({COND_TYPE_FAST, 0});
1048 /* Clock resolution in usec. */
1049 SCLASS int _clock_res_usec SCLASS_PRESET(CLOCK_RES_USEC);
1051 /* Array of signal actions for this process: */
1052 SCLASS struct sigaction _thread_sigact[_SIG_MAXSIG];
1055 * Lock for above count of dummy handlers and for the process signal
1056 * mask and pending signal sets.
1058 SCLASS struct lock _thread_signal_lock;
1060 /* Pending signals and mask for this process: */
1061 SCLASS sigset_t _thr_proc_sigpending;
1062 SCLASS siginfo_t _thr_proc_siginfo[_SIG_MAXSIG];
1064 SCLASS pid_t _thr_pid SCLASS_PRESET(0);
1066 /* Garbage collector lock. */
1067 SCLASS struct lock _gc_lock;
1068 SCLASS int _gc_check SCLASS_PRESET(0);
1069 SCLASS int _gc_count SCLASS_PRESET(0);
1071 SCLASS struct lock _mutex_static_lock;
1072 SCLASS struct lock _rwlock_static_lock;
1073 SCLASS struct lock _keytable_lock;
1074 SCLASS struct lock _thread_list_lock;
1075 SCLASS int _thr_guard_default;
1076 SCLASS int _thr_page_size;
1077 SCLASS pthread_t _thr_sig_daemon;
1078 SCLASS int _thr_debug_flags SCLASS_PRESET(0);
1080 /* Undefine the storage class and preset specifiers: */
1082 #undef SCLASS_PRESET
1086 * Function prototype definitions.
1089 int _cond_reinit(pthread_cond_t *);
1090 struct kse *_kse_alloc(struct pthread *, int sys_scope);
1091 kse_critical_t _kse_critical_enter(void);
1092 void _kse_critical_leave(kse_critical_t);
1093 int _kse_in_critical(void);
1094 void _kse_free(struct pthread *, struct kse *);
1096 struct kse_group *_kseg_alloc(struct pthread *);
1097 void _kse_lock_wait(struct lock *, struct lockuser *lu);
1098 void _kse_lock_wakeup(struct lock *, struct lockuser *lu);
1099 void _kse_single_thread(struct pthread *);
1100 int _kse_setthreaded(int);
1101 void _kseg_free(struct kse_group *);
1102 int _mutex_cv_lock(pthread_mutex_t *);
1103 int _mutex_cv_unlock(pthread_mutex_t *);
1104 void _mutex_notify_priochange(struct pthread *, struct pthread *, int);
1105 int _mutex_reinit(struct pthread_mutex *);
1106 void _mutex_unlock_private(struct pthread *);
1107 void _libpthread_init(struct pthread *);
1108 int _pq_alloc(struct pq_queue *, int, int);
1109 void _pq_free(struct pq_queue *);
1110 int _pq_init(struct pq_queue *);
1111 void _pq_remove(struct pq_queue *pq, struct pthread *);
1112 void _pq_insert_head(struct pq_queue *pq, struct pthread *);
1113 void _pq_insert_tail(struct pq_queue *pq, struct pthread *);
1114 struct pthread *_pq_first(struct pq_queue *pq);
1115 struct pthread *_pq_first_debug(struct pq_queue *pq);
1116 void *_pthread_getspecific(pthread_key_t);
1117 int _pthread_key_create(pthread_key_t *, void (*) (void *));
1118 int _pthread_key_delete(pthread_key_t);
1119 int _pthread_mutex_destroy(pthread_mutex_t *);
1120 int _pthread_mutex_init(pthread_mutex_t *, const pthread_mutexattr_t *);
1121 int _pthread_mutex_lock(pthread_mutex_t *);
1122 int _pthread_mutex_trylock(pthread_mutex_t *);
1123 int _pthread_mutex_unlock(pthread_mutex_t *);
1124 int _pthread_mutexattr_init(pthread_mutexattr_t *);
1125 int _pthread_mutexattr_destroy(pthread_mutexattr_t *);
1126 int _pthread_mutexattr_settype(pthread_mutexattr_t *, int);
1127 int _pthread_once(pthread_once_t *, void (*) (void));
1128 int _pthread_rwlock_init(pthread_rwlock_t *, const pthread_rwlockattr_t *);
1129 int _pthread_rwlock_destroy (pthread_rwlock_t *);
1130 struct pthread *_pthread_self(void);
1131 int _pthread_setspecific(pthread_key_t, const void *);
1132 void _pthread_yield(void);
1133 void _pthread_cleanup_push(void (*routine) (void *), void *routine_arg);
1134 void _pthread_cleanup_pop(int execute);
1135 struct pthread *_thr_alloc(struct pthread *);
1136 void _thr_exit(char *, int, char *);
1137 void _thr_exit_cleanup(void);
1138 void _thr_lock_wait(struct lock *lock, struct lockuser *lu);
1139 void _thr_lock_wakeup(struct lock *lock, struct lockuser *lu);
1140 void _thr_mutex_reinit(pthread_mutex_t *);
1141 int _thr_ref_add(struct pthread *, struct pthread *, int);
1142 void _thr_ref_delete(struct pthread *, struct pthread *);
1143 void _thr_rtld_init(void);
1144 void _thr_rtld_fini(void);
1145 int _thr_schedule_add(struct pthread *, struct pthread *);
1146 void _thr_schedule_remove(struct pthread *, struct pthread *);
1147 void _thr_setrunnable(struct pthread *curthread, struct pthread *thread);
1148 struct kse_mailbox *_thr_setrunnable_unlocked(struct pthread *thread);
1149 struct kse_mailbox *_thr_sig_add(struct pthread *, int, siginfo_t *);
1150 void _thr_sig_dispatch(struct kse *, int, siginfo_t *);
1151 int _thr_stack_alloc(struct pthread_attr *);
1152 void _thr_stack_free(struct pthread_attr *);
1153 void _thr_exit_cleanup(void);
1154 void _thr_free(struct pthread *, struct pthread *);
1155 void _thr_gc(struct pthread *);
1156 void _thr_panic_exit(char *, int, char *);
1157 void _thread_cleanupspecific(void);
1158 void _thread_dump_info(void);
1159 void _thread_printf(int, const char *, ...);
1160 void _thr_sched_switch(struct pthread *);
1161 void _thr_sched_switch_unlocked(struct pthread *);
1162 void _thr_set_timeout(const struct timespec *);
1163 void _thr_seterrno(struct pthread *, int);
1164 void _thr_sig_handler(int, siginfo_t *, ucontext_t *);
1165 void _thr_sig_check_pending(struct pthread *);
1166 void _thr_sig_rundown(struct pthread *, ucontext_t *);
1167 void _thr_sig_send(struct pthread *pthread, int sig);
1168 void _thr_sigframe_restore(struct pthread *thread, struct pthread_sigframe *psf);
1169 void _thr_spinlock_init(void);
1170 void _thr_cancel_enter(struct pthread *);
1171 void _thr_cancel_leave(struct pthread *, int);
1172 int _thr_setconcurrency(int new_level);
1173 int _thr_setmaxconcurrency(void);
1174 void _thr_critical_enter(struct pthread *);
1175 void _thr_critical_leave(struct pthread *);
1176 int _thr_start_sig_daemon(void);
1177 int _thr_getprocsig(int sig, siginfo_t *siginfo);
1178 int _thr_getprocsig_unlocked(int sig, siginfo_t *siginfo);
1179 void _thr_signal_init(void);
1180 void _thr_signal_deinit(void);
1181 void _thr_hash_add(struct pthread *);
1182 void _thr_hash_remove(struct pthread *);
1183 struct pthread *_thr_hash_find(struct pthread *);
1184 void _thr_finish_cancellation(void *arg);
1185 int _thr_sigonstack(void *sp);
1186 void _thr_debug_check_yield(struct pthread *);
1189 * Aliases for _pthread functions. Should be called instead of
1190 * originals if PLT replocation is unwanted at runtme.
1192 int _thr_cond_broadcast(pthread_cond_t *);
1193 int _thr_cond_signal(pthread_cond_t *);
1194 int _thr_cond_wait(pthread_cond_t *, pthread_mutex_t *);
1195 int _thr_mutex_lock(pthread_mutex_t *);
1196 int _thr_mutex_unlock(pthread_mutex_t *);
1197 int _thr_rwlock_rdlock (pthread_rwlock_t *);
1198 int _thr_rwlock_wrlock (pthread_rwlock_t *);
1199 int _thr_rwlock_unlock (pthread_rwlock_t *);
1201 /* #include <sys/aio.h> */
1203 int __sys_aio_suspend(const struct aiocb * const[], int, const struct timespec *);
1206 /* #include <fcntl.h> */
1207 #ifdef _SYS_FCNTL_H_
1208 int __sys_fcntl(int, int, ...);
1209 int __sys_open(const char *, int, ...);
1212 /* #include <sys/ioctl.h> */
1213 #ifdef _SYS_IOCTL_H_
1214 int __sys_ioctl(int, unsigned long, ...);
1217 /* #inclde <sched.h> */
1219 int __sys_sched_yield(void);
1222 /* #include <signal.h> */
1224 int __sys_kill(pid_t, int);
1225 int __sys_sigaction(int, const struct sigaction *, struct sigaction *);
1226 int __sys_sigpending(sigset_t *);
1227 int __sys_sigprocmask(int, const sigset_t *, sigset_t *);
1228 int __sys_sigsuspend(const sigset_t *);
1229 int __sys_sigreturn(ucontext_t *);
1230 int __sys_sigaltstack(const struct sigaltstack *, struct sigaltstack *);
1233 /* #include <sys/socket.h> */
1234 #ifdef _SYS_SOCKET_H_
1235 int __sys_accept(int, struct sockaddr *, socklen_t *);
1236 int __sys_connect(int, const struct sockaddr *, socklen_t);
1237 int __sys_sendfile(int, int, off_t, size_t, struct sf_hdtr *,
1241 /* #include <sys/uio.h> */
1243 ssize_t __sys_readv(int, const struct iovec *, int);
1244 ssize_t __sys_writev(int, const struct iovec *, int);
1247 /* #include <time.h> */
1249 int __sys_nanosleep(const struct timespec *, struct timespec *);
1252 /* #include <unistd.h> */
1254 int __sys_close(int);
1255 int __sys_execve(const char *, char * const *, char * const *);
1256 int __sys_fork(void);
1257 int __sys_fsync(int);
1258 pid_t __sys_getpid(void);
1259 int __sys_select(int, fd_set *, fd_set *, fd_set *, struct timeval *);
1260 ssize_t __sys_read(int, void *, size_t);
1261 ssize_t __sys_write(int, const void *, size_t);
1262 void __sys_exit(int);
1263 int __sys_sigwait(const sigset_t *, int *);
1264 int __sys_sigtimedwait(sigset_t *, siginfo_t *, struct timespec *);
1267 /* #include <poll.h> */
1269 int __sys_poll(struct pollfd *, unsigned, int);
1272 /* #include <sys/mman.h> */
1274 int __sys_msync(void *, size_t, int);
1277 #endif /* !_THR_PRIVATE_H */