1/* 2 * Copyright (C) 2005 Daniel M. Eischen <deischen@freebsd.org> 3 * Copyright (c) 2005 David Xu <davidxu@freebsd.org> 4 * Copyright (c) 1995-1998 John Birrell <jb@cimlogic.com.au>. 5 * 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice unmodified, this list of conditions, and the following 13 * disclaimer. 14 * 2. Redistributions in binary form must reproduce the above copyright 15 * notice, this list of conditions and the following disclaimer in the 16 * documentation and/or other materials provided with the distribution. 17 * 18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 19 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 21 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 23 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 28 *
| 1/* 2 * Copyright (C) 2005 Daniel M. Eischen <deischen@freebsd.org> 3 * Copyright (c) 2005 David Xu <davidxu@freebsd.org> 4 * Copyright (c) 1995-1998 John Birrell <jb@cimlogic.com.au>. 5 * 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice unmodified, this list of conditions, and the following 13 * disclaimer. 14 * 2. Redistributions in binary form must reproduce the above copyright 15 * notice, this list of conditions and the following disclaimer in the 16 * documentation and/or other materials provided with the distribution. 17 * 18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 19 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 21 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 23 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 28 *
|
29 * $FreeBSD: head/lib/libthr/thread/thr_private.h 214334 2010-10-25 09:16:04Z davidxu $
| 29 * $FreeBSD: head/lib/libthr/thread/thr_private.h 216641 2010-12-22 05:01:52Z davidxu $
|
30 */ 31 32#ifndef _THR_PRIVATE_H 33#define _THR_PRIVATE_H 34 35/* 36 * Include files. 37 */ 38#include <sys/types.h> 39#include <sys/time.h> 40#include <sys/cdefs.h> 41#include <sys/queue.h> 42#include <sys/param.h> 43#include <sys/cpuset.h> 44#include <machine/atomic.h> 45#include <errno.h> 46#include <limits.h> 47#include <signal.h> 48#include <stddef.h> 49#include <stdio.h> 50#include <unistd.h> 51#include <ucontext.h> 52#include <sys/thr.h> 53#include <pthread.h> 54 55#define SYM_FB10(sym) __CONCAT(sym, _fb10) 56#define SYM_FBP10(sym) __CONCAT(sym, _fbp10) 57#define WEAK_REF(sym, alias) __weak_reference(sym, alias) 58#define SYM_COMPAT(sym, impl, ver) __sym_compat(sym, impl, ver) 59#define SYM_DEFAULT(sym, impl, ver) __sym_default(sym, impl, ver) 60 61#define FB10_COMPAT(func, sym) \ 62 WEAK_REF(func, SYM_FB10(sym)); \ 63 SYM_COMPAT(sym, SYM_FB10(sym), FBSD_1.0) 64 65#define FB10_COMPAT_PRIVATE(func, sym) \ 66 WEAK_REF(func, SYM_FBP10(sym)); \ 67 SYM_DEFAULT(sym, SYM_FBP10(sym), FBSDprivate_1.0) 68 69#include "pthread_md.h" 70#include "thr_umtx.h" 71#include "thread_db.h" 72 73#ifdef _PTHREAD_FORCED_UNWIND 74#define _BSD_SOURCE 75#include <unwind.h> 76#endif 77 78typedef TAILQ_HEAD(pthreadlist, pthread) pthreadlist; 79typedef TAILQ_HEAD(atfork_head, pthread_atfork) atfork_head; 80TAILQ_HEAD(mutex_queue, pthread_mutex); 81 82/* Signal to do cancellation */ 83#define SIGCANCEL 32 84 85/* 86 * Kernel fatal error handler macro. 87 */ 88#define PANIC(string) _thread_exit(__FILE__,__LINE__,string) 89 90/* Output debug messages like this: */ 91#define stdout_debug(args...) _thread_printf(STDOUT_FILENO, ##args) 92#define stderr_debug(args...) _thread_printf(STDERR_FILENO, ##args) 93 94#ifdef _PTHREADS_INVARIANTS 95#define THR_ASSERT(cond, msg) do { \ 96 if (__predict_false(!(cond))) \ 97 PANIC(msg); \ 98} while (0) 99#else 100#define THR_ASSERT(cond, msg) 101#endif 102 103#ifdef PIC 104# define STATIC_LIB_REQUIRE(name) 105#else 106# define STATIC_LIB_REQUIRE(name) __asm (".globl " #name) 107#endif 108 109#define TIMESPEC_ADD(dst, src, val) \ 110 do { \ 111 (dst)->tv_sec = (src)->tv_sec + (val)->tv_sec; \ 112 (dst)->tv_nsec = (src)->tv_nsec + (val)->tv_nsec; \ 113 if ((dst)->tv_nsec >= 1000000000) { \ 114 (dst)->tv_sec++; \ 115 (dst)->tv_nsec -= 1000000000; \ 116 } \ 117 } while (0) 118 119#define TIMESPEC_SUB(dst, src, val) \ 120 do { \ 121 (dst)->tv_sec = (src)->tv_sec - (val)->tv_sec; \ 122 (dst)->tv_nsec = (src)->tv_nsec - (val)->tv_nsec; \ 123 if ((dst)->tv_nsec < 0) { \ 124 (dst)->tv_sec--; \ 125 (dst)->tv_nsec += 1000000000; \ 126 } \ 127 } while (0) 128 129/* XXX These values should be same as those defined in pthread.h */ 130#define THR_MUTEX_INITIALIZER ((struct pthread_mutex *)NULL) 131#define THR_ADAPTIVE_MUTEX_INITIALIZER ((struct pthread_mutex *)1) 132#define THR_MUTEX_DESTROYED ((struct pthread_mutex *)2) 133#define THR_COND_INITIALIZER ((struct pthread_cond *)NULL) 134#define THR_COND_DESTROYED ((struct pthread_cond *)1) 135#define THR_RWLOCK_INITIALIZER ((struct pthread_rwlock *)NULL) 136#define THR_RWLOCK_DESTROYED ((struct pthread_rwlock *)1) 137
| 30 */ 31 32#ifndef _THR_PRIVATE_H 33#define _THR_PRIVATE_H 34 35/* 36 * Include files. 37 */ 38#include <sys/types.h> 39#include <sys/time.h> 40#include <sys/cdefs.h> 41#include <sys/queue.h> 42#include <sys/param.h> 43#include <sys/cpuset.h> 44#include <machine/atomic.h> 45#include <errno.h> 46#include <limits.h> 47#include <signal.h> 48#include <stddef.h> 49#include <stdio.h> 50#include <unistd.h> 51#include <ucontext.h> 52#include <sys/thr.h> 53#include <pthread.h> 54 55#define SYM_FB10(sym) __CONCAT(sym, _fb10) 56#define SYM_FBP10(sym) __CONCAT(sym, _fbp10) 57#define WEAK_REF(sym, alias) __weak_reference(sym, alias) 58#define SYM_COMPAT(sym, impl, ver) __sym_compat(sym, impl, ver) 59#define SYM_DEFAULT(sym, impl, ver) __sym_default(sym, impl, ver) 60 61#define FB10_COMPAT(func, sym) \ 62 WEAK_REF(func, SYM_FB10(sym)); \ 63 SYM_COMPAT(sym, SYM_FB10(sym), FBSD_1.0) 64 65#define FB10_COMPAT_PRIVATE(func, sym) \ 66 WEAK_REF(func, SYM_FBP10(sym)); \ 67 SYM_DEFAULT(sym, SYM_FBP10(sym), FBSDprivate_1.0) 68 69#include "pthread_md.h" 70#include "thr_umtx.h" 71#include "thread_db.h" 72 73#ifdef _PTHREAD_FORCED_UNWIND 74#define _BSD_SOURCE 75#include <unwind.h> 76#endif 77 78typedef TAILQ_HEAD(pthreadlist, pthread) pthreadlist; 79typedef TAILQ_HEAD(atfork_head, pthread_atfork) atfork_head; 80TAILQ_HEAD(mutex_queue, pthread_mutex); 81 82/* Signal to do cancellation */ 83#define SIGCANCEL 32 84 85/* 86 * Kernel fatal error handler macro. 87 */ 88#define PANIC(string) _thread_exit(__FILE__,__LINE__,string) 89 90/* Output debug messages like this: */ 91#define stdout_debug(args...) _thread_printf(STDOUT_FILENO, ##args) 92#define stderr_debug(args...) _thread_printf(STDERR_FILENO, ##args) 93 94#ifdef _PTHREADS_INVARIANTS 95#define THR_ASSERT(cond, msg) do { \ 96 if (__predict_false(!(cond))) \ 97 PANIC(msg); \ 98} while (0) 99#else 100#define THR_ASSERT(cond, msg) 101#endif 102 103#ifdef PIC 104# define STATIC_LIB_REQUIRE(name) 105#else 106# define STATIC_LIB_REQUIRE(name) __asm (".globl " #name) 107#endif 108 109#define TIMESPEC_ADD(dst, src, val) \ 110 do { \ 111 (dst)->tv_sec = (src)->tv_sec + (val)->tv_sec; \ 112 (dst)->tv_nsec = (src)->tv_nsec + (val)->tv_nsec; \ 113 if ((dst)->tv_nsec >= 1000000000) { \ 114 (dst)->tv_sec++; \ 115 (dst)->tv_nsec -= 1000000000; \ 116 } \ 117 } while (0) 118 119#define TIMESPEC_SUB(dst, src, val) \ 120 do { \ 121 (dst)->tv_sec = (src)->tv_sec - (val)->tv_sec; \ 122 (dst)->tv_nsec = (src)->tv_nsec - (val)->tv_nsec; \ 123 if ((dst)->tv_nsec < 0) { \ 124 (dst)->tv_sec--; \ 125 (dst)->tv_nsec += 1000000000; \ 126 } \ 127 } while (0) 128 129/* XXX These values should be same as those defined in pthread.h */ 130#define THR_MUTEX_INITIALIZER ((struct pthread_mutex *)NULL) 131#define THR_ADAPTIVE_MUTEX_INITIALIZER ((struct pthread_mutex *)1) 132#define THR_MUTEX_DESTROYED ((struct pthread_mutex *)2) 133#define THR_COND_INITIALIZER ((struct pthread_cond *)NULL) 134#define THR_COND_DESTROYED ((struct pthread_cond *)1) 135#define THR_RWLOCK_INITIALIZER ((struct pthread_rwlock *)NULL) 136#define THR_RWLOCK_DESTROYED ((struct pthread_rwlock *)1) 137
|
| 138#define PMUTEX_FLAG_TYPE_MASK 0x0ff 139#define PMUTEX_FLAG_PRIVATE 0x100 140#define PMUTEX_FLAG_DEFERED 0x200 141#define PMUTEX_TYPE(mtxflags) ((mtxflags) & PMUTEX_FLAG_TYPE_MASK) 142 143#define MAX_DEFER_WAITERS 50 144
|
138struct pthread_mutex { 139 /* 140 * Lock for accesses to this structure. 141 */ 142 struct umutex m_lock;
| 145struct pthread_mutex { 146 /* 147 * Lock for accesses to this structure. 148 */ 149 struct umutex m_lock;
|
143 enum pthread_mutextype m_type;
| 150 int m_flags;
|
144 struct pthread *m_owner; 145 int m_count;
| 151 struct pthread *m_owner; 152 int m_count;
|
146 int m_refcount;
| |
147 int m_spinloops; 148 int m_yieldloops;
| 153 int m_spinloops; 154 int m_yieldloops;
|
149 int m_private;
| |
150 /* 151 * Link for all mutexes a thread currently owns. 152 */ 153 TAILQ_ENTRY(pthread_mutex) m_qe; 154}; 155 156struct pthread_mutex_attr { 157 enum pthread_mutextype m_type; 158 int m_protocol; 159 int m_ceiling; 160}; 161 162#define PTHREAD_MUTEXATTR_STATIC_INITIALIZER \ 163 { PTHREAD_MUTEX_DEFAULT, PTHREAD_PRIO_NONE, 0, MUTEX_FLAGS_PRIVATE } 164 165struct pthread_cond {
| 155 /* 156 * Link for all mutexes a thread currently owns. 157 */ 158 TAILQ_ENTRY(pthread_mutex) m_qe; 159}; 160 161struct pthread_mutex_attr { 162 enum pthread_mutextype m_type; 163 int m_protocol; 164 int m_ceiling; 165}; 166 167#define PTHREAD_MUTEXATTR_STATIC_INITIALIZER \ 168 { PTHREAD_MUTEX_DEFAULT, PTHREAD_PRIO_NONE, 0, MUTEX_FLAGS_PRIVATE } 169 170struct pthread_cond {
|
166 struct umutex c_lock; 167 struct ucond c_kerncv; 168 int c_pshared; 169 int c_clockid;
| 171 __uint32_t __has_user_waiters; 172 __uint32_t __has_kern_waiters; 173 __uint32_t __flags; 174 __uint32_t __clock_id;
|
170}; 171 172struct pthread_cond_attr { 173 int c_pshared; 174 int c_clockid; 175}; 176 177struct pthread_barrier { 178 struct umutex b_lock; 179 struct ucond b_cv; 180 volatile int64_t b_cycle; 181 volatile int b_count; 182 volatile int b_waiters; 183}; 184 185struct pthread_barrierattr { 186 int pshared; 187}; 188 189struct pthread_spinlock { 190 struct umutex s_lock; 191}; 192 193/* 194 * Flags for condition variables. 195 */ 196#define COND_FLAGS_PRIVATE 0x01 197#define COND_FLAGS_INITED 0x02 198#define COND_FLAGS_BUSY 0x04 199 200/* 201 * Cleanup definitions. 202 */ 203struct pthread_cleanup { 204 struct pthread_cleanup *prev; 205 void (*routine)(void *); 206 void *routine_arg; 207 int onheap; 208}; 209 210#define THR_CLEANUP_PUSH(td, func, arg) { \ 211 struct pthread_cleanup __cup; \ 212 \ 213 __cup.routine = func; \ 214 __cup.routine_arg = arg; \ 215 __cup.onheap = 0; \ 216 __cup.prev = (td)->cleanup; \ 217 (td)->cleanup = &__cup; 218 219#define THR_CLEANUP_POP(td, exec) \ 220 (td)->cleanup = __cup.prev; \ 221 if ((exec) != 0) \ 222 __cup.routine(__cup.routine_arg); \ 223} 224 225struct pthread_atfork { 226 TAILQ_ENTRY(pthread_atfork) qe; 227 void (*prepare)(void); 228 void (*parent)(void); 229 void (*child)(void); 230}; 231 232struct pthread_attr { 233#define pthread_attr_start_copy sched_policy 234 int sched_policy; 235 int sched_inherit; 236 int prio; 237 int suspend; 238#define THR_STACK_USER 0x100 /* 0xFF reserved for <pthread.h> */ 239 int flags; 240 void *stackaddr_attr; 241 size_t stacksize_attr; 242 size_t guardsize_attr; 243#define pthread_attr_end_copy cpuset 244 cpuset_t *cpuset; 245 size_t cpusetsize; 246}; 247
| 175}; 176 177struct pthread_cond_attr { 178 int c_pshared; 179 int c_clockid; 180}; 181 182struct pthread_barrier { 183 struct umutex b_lock; 184 struct ucond b_cv; 185 volatile int64_t b_cycle; 186 volatile int b_count; 187 volatile int b_waiters; 188}; 189 190struct pthread_barrierattr { 191 int pshared; 192}; 193 194struct pthread_spinlock { 195 struct umutex s_lock; 196}; 197 198/* 199 * Flags for condition variables. 200 */ 201#define COND_FLAGS_PRIVATE 0x01 202#define COND_FLAGS_INITED 0x02 203#define COND_FLAGS_BUSY 0x04 204 205/* 206 * Cleanup definitions. 207 */ 208struct pthread_cleanup { 209 struct pthread_cleanup *prev; 210 void (*routine)(void *); 211 void *routine_arg; 212 int onheap; 213}; 214 215#define THR_CLEANUP_PUSH(td, func, arg) { \ 216 struct pthread_cleanup __cup; \ 217 \ 218 __cup.routine = func; \ 219 __cup.routine_arg = arg; \ 220 __cup.onheap = 0; \ 221 __cup.prev = (td)->cleanup; \ 222 (td)->cleanup = &__cup; 223 224#define THR_CLEANUP_POP(td, exec) \ 225 (td)->cleanup = __cup.prev; \ 226 if ((exec) != 0) \ 227 __cup.routine(__cup.routine_arg); \ 228} 229 230struct pthread_atfork { 231 TAILQ_ENTRY(pthread_atfork) qe; 232 void (*prepare)(void); 233 void (*parent)(void); 234 void (*child)(void); 235}; 236 237struct pthread_attr { 238#define pthread_attr_start_copy sched_policy 239 int sched_policy; 240 int sched_inherit; 241 int prio; 242 int suspend; 243#define THR_STACK_USER 0x100 /* 0xFF reserved for <pthread.h> */ 244 int flags; 245 void *stackaddr_attr; 246 size_t stacksize_attr; 247 size_t guardsize_attr; 248#define pthread_attr_end_copy cpuset 249 cpuset_t *cpuset; 250 size_t cpusetsize; 251}; 252
|
| 253struct wake_addr { 254 struct wake_addr *link; 255 unsigned int value; 256 char pad[12]; 257}; 258 259struct sleepqueue { 260 TAILQ_HEAD(, pthread) sq_blocked; 261 SLIST_HEAD(, sleepqueue) sq_freeq; 262 LIST_ENTRY(sleepqueue) sq_hash; 263 SLIST_ENTRY(sleepqueue) sq_flink; 264 void *sq_wchan; 265 int sq_type; 266}; 267
|
248/* 249 * Thread creation state attributes. 250 */ 251#define THR_CREATE_RUNNING 0 252#define THR_CREATE_SUSPENDED 1 253 254/* 255 * Miscellaneous definitions. 256 */ 257#define THR_STACK_DEFAULT (sizeof(void *) / 4 * 1024 * 1024) 258 259/* 260 * Maximum size of initial thread's stack. This perhaps deserves to be larger 261 * than the stacks of other threads, since many applications are likely to run 262 * almost entirely on this stack. 263 */ 264#define THR_STACK_INITIAL (THR_STACK_DEFAULT * 2) 265 266/* 267 * Define priorities returned by kernel. 268 */ 269#define THR_MIN_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_min) 270#define THR_MAX_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_max) 271#define THR_DEF_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_default) 272 273#define THR_MIN_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_min) 274#define THR_MAX_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_max) 275#define THR_DEF_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_default) 276 277/* XXX The SCHED_FIFO should have same priority range as SCHED_RR */ 278#define THR_MIN_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO_1].pri_min) 279#define THR_MAX_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO-1].pri_max) 280#define THR_DEF_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO-1].pri_default) 281 282struct pthread_prio { 283 int pri_min; 284 int pri_max; 285 int pri_default; 286}; 287 288struct pthread_rwlockattr { 289 int pshared; 290}; 291 292struct pthread_rwlock { 293 struct urwlock lock; 294 struct pthread *owner; 295}; 296 297/* 298 * Thread states. 299 */ 300enum pthread_state { 301 PS_RUNNING, 302 PS_DEAD 303}; 304 305struct pthread_specific_elem { 306 const void *data; 307 int seqno; 308}; 309 310struct pthread_key { 311 volatile int allocated; 312 int seqno; 313 void (*destructor)(void *); 314}; 315 316/* 317 * lwpid_t is 32bit but kernel thr API exports tid as long type 318 * in very earily date. 319 */ 320#define TID(thread) ((uint32_t) ((thread)->tid)) 321 322/* 323 * Thread structure. 324 */ 325struct pthread { 326 /* Kernel thread id. */ 327 long tid; 328#define TID_TERMINATED 1 329 330 /* 331 * Lock for accesses to this thread structure. 332 */ 333 struct umutex lock; 334 335 /* Internal condition variable cycle number. */ 336 uint32_t cycle; 337 338 /* How many low level locks the thread held. */ 339 int locklevel; 340 341 /* 342 * Set to non-zero when this thread has entered a critical 343 * region. We allow for recursive entries into critical regions. 344 */ 345 int critical_count; 346 347 /* Signal blocked counter. */ 348 int sigblock; 349 350 /* Queue entry for list of all threads. */ 351 TAILQ_ENTRY(pthread) tle; /* link for all threads in process */ 352 353 /* Queue entry for GC lists. */ 354 TAILQ_ENTRY(pthread) gcle; 355 356 /* Hash queue entry. */ 357 LIST_ENTRY(pthread) hle; 358
| 268/* 269 * Thread creation state attributes. 270 */ 271#define THR_CREATE_RUNNING 0 272#define THR_CREATE_SUSPENDED 1 273 274/* 275 * Miscellaneous definitions. 276 */ 277#define THR_STACK_DEFAULT (sizeof(void *) / 4 * 1024 * 1024) 278 279/* 280 * Maximum size of initial thread's stack. This perhaps deserves to be larger 281 * than the stacks of other threads, since many applications are likely to run 282 * almost entirely on this stack. 283 */ 284#define THR_STACK_INITIAL (THR_STACK_DEFAULT * 2) 285 286/* 287 * Define priorities returned by kernel. 288 */ 289#define THR_MIN_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_min) 290#define THR_MAX_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_max) 291#define THR_DEF_PRIORITY (_thr_priorities[SCHED_OTHER-1].pri_default) 292 293#define THR_MIN_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_min) 294#define THR_MAX_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_max) 295#define THR_DEF_RR_PRIORITY (_thr_priorities[SCHED_RR-1].pri_default) 296 297/* XXX The SCHED_FIFO should have same priority range as SCHED_RR */ 298#define THR_MIN_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO_1].pri_min) 299#define THR_MAX_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO-1].pri_max) 300#define THR_DEF_FIFO_PRIORITY (_thr_priorities[SCHED_FIFO-1].pri_default) 301 302struct pthread_prio { 303 int pri_min; 304 int pri_max; 305 int pri_default; 306}; 307 308struct pthread_rwlockattr { 309 int pshared; 310}; 311 312struct pthread_rwlock { 313 struct urwlock lock; 314 struct pthread *owner; 315}; 316 317/* 318 * Thread states. 319 */ 320enum pthread_state { 321 PS_RUNNING, 322 PS_DEAD 323}; 324 325struct pthread_specific_elem { 326 const void *data; 327 int seqno; 328}; 329 330struct pthread_key { 331 volatile int allocated; 332 int seqno; 333 void (*destructor)(void *); 334}; 335 336/* 337 * lwpid_t is 32bit but kernel thr API exports tid as long type 338 * in very earily date. 339 */ 340#define TID(thread) ((uint32_t) ((thread)->tid)) 341 342/* 343 * Thread structure. 344 */ 345struct pthread { 346 /* Kernel thread id. */ 347 long tid; 348#define TID_TERMINATED 1 349 350 /* 351 * Lock for accesses to this thread structure. 352 */ 353 struct umutex lock; 354 355 /* Internal condition variable cycle number. */ 356 uint32_t cycle; 357 358 /* How many low level locks the thread held. */ 359 int locklevel; 360 361 /* 362 * Set to non-zero when this thread has entered a critical 363 * region. We allow for recursive entries into critical regions. 364 */ 365 int critical_count; 366 367 /* Signal blocked counter. */ 368 int sigblock; 369 370 /* Queue entry for list of all threads. */ 371 TAILQ_ENTRY(pthread) tle; /* link for all threads in process */ 372 373 /* Queue entry for GC lists. */ 374 TAILQ_ENTRY(pthread) gcle; 375 376 /* Hash queue entry. */ 377 LIST_ENTRY(pthread) hle; 378
|
| 379 /* Sleep queue entry */ 380 TAILQ_ENTRY(pthread) wle; 381
|
359 /* Threads reference count. */ 360 int refcount; 361 362 /* 363 * Thread start routine, argument, stack pointer and thread 364 * attributes. 365 */ 366 void *(*start_routine)(void *); 367 void *arg; 368 struct pthread_attr attr; 369 370#define SHOULD_CANCEL(thr) \ 371 ((thr)->cancel_pending && (thr)->cancel_enable && \ 372 (thr)->no_cancel == 0) 373 374 /* Cancellation is enabled */ 375 int cancel_enable; 376 377 /* Cancellation request is pending */ 378 int cancel_pending; 379 380 /* Thread is at cancellation point */ 381 int cancel_point; 382 383 /* Cancellation is temporarily disabled */ 384 int no_cancel; 385 386 /* Asynchronouse cancellation is enabled */ 387 int cancel_async; 388 389 /* Cancellation is in progress */ 390 int cancelling; 391 392 /* Thread temporary signal mask. */ 393 sigset_t sigmask; 394 395 /* Thread should unblock SIGCANCEL. */ 396 int unblock_sigcancel; 397 398 /* In sigsuspend state */ 399 int in_sigsuspend; 400 401 /* deferred signal info */ 402 siginfo_t deferred_siginfo; 403 404 /* signal mask to restore. */ 405 sigset_t deferred_sigmask; 406 407 /* the sigaction should be used for deferred signal. */ 408 struct sigaction deferred_sigact; 409 410 /* Force new thread to exit. */ 411 int force_exit; 412 413 /* Thread state: */ 414 enum pthread_state state; 415 416 /* 417 * Error variable used instead of errno. The function __error() 418 * returns a pointer to this. 419 */ 420 int error; 421 422 /* 423 * The joiner is the thread that is joining to this thread. The 424 * join status keeps track of a join operation to another thread. 425 */ 426 struct pthread *joiner; 427 428 /* Miscellaneous flags; only set with scheduling lock held. */ 429 int flags; 430#define THR_FLAGS_PRIVATE 0x0001 431#define THR_FLAGS_NEED_SUSPEND 0x0002 /* thread should be suspended */ 432#define THR_FLAGS_SUSPENDED 0x0004 /* thread is suspended */ 433#define THR_FLAGS_DETACHED 0x0008 /* thread is detached */ 434 435 /* Thread list flags; only set with thread list lock held. */ 436 int tlflags; 437#define TLFLAGS_GC_SAFE 0x0001 /* thread safe for cleaning */ 438#define TLFLAGS_IN_TDLIST 0x0002 /* thread in all thread list */ 439#define TLFLAGS_IN_GCLIST 0x0004 /* thread in gc list */ 440 441 /* Queue of currently owned NORMAL or PRIO_INHERIT type mutexes. */ 442 struct mutex_queue mutexq; 443 444 /* Queue of all owned PRIO_PROTECT mutexes. */ 445 struct mutex_queue pp_mutexq; 446 447 void *ret; 448 struct pthread_specific_elem *specific; 449 int specific_data_count; 450 451 /* Number rwlocks rdlocks held. */ 452 int rdlock_count; 453 454 /* 455 * Current locks bitmap for rtld. */ 456 int rtld_bits; 457 458 /* Thread control block */ 459 struct tcb *tcb; 460 461 /* Cleanup handlers Link List */ 462 struct pthread_cleanup *cleanup; 463 464#ifdef _PTHREAD_FORCED_UNWIND 465 struct _Unwind_Exception ex; 466 void *unwind_stackend; 467 int unwind_disabled; 468#endif 469 470 /* 471 * Magic value to help recognize a valid thread structure 472 * from an invalid one: 473 */ 474#define THR_MAGIC ((u_int32_t) 0xd09ba115) 475 u_int32_t magic; 476 477 /* Enable event reporting */ 478 int report_events; 479 480 /* Event mask */ 481 int event_mask; 482 483 /* Event */ 484 td_event_msg_t event_buf;
| 382 /* Threads reference count. */ 383 int refcount; 384 385 /* 386 * Thread start routine, argument, stack pointer and thread 387 * attributes. 388 */ 389 void *(*start_routine)(void *); 390 void *arg; 391 struct pthread_attr attr; 392 393#define SHOULD_CANCEL(thr) \ 394 ((thr)->cancel_pending && (thr)->cancel_enable && \ 395 (thr)->no_cancel == 0) 396 397 /* Cancellation is enabled */ 398 int cancel_enable; 399 400 /* Cancellation request is pending */ 401 int cancel_pending; 402 403 /* Thread is at cancellation point */ 404 int cancel_point; 405 406 /* Cancellation is temporarily disabled */ 407 int no_cancel; 408 409 /* Asynchronouse cancellation is enabled */ 410 int cancel_async; 411 412 /* Cancellation is in progress */ 413 int cancelling; 414 415 /* Thread temporary signal mask. */ 416 sigset_t sigmask; 417 418 /* Thread should unblock SIGCANCEL. */ 419 int unblock_sigcancel; 420 421 /* In sigsuspend state */ 422 int in_sigsuspend; 423 424 /* deferred signal info */ 425 siginfo_t deferred_siginfo; 426 427 /* signal mask to restore. */ 428 sigset_t deferred_sigmask; 429 430 /* the sigaction should be used for deferred signal. */ 431 struct sigaction deferred_sigact; 432 433 /* Force new thread to exit. */ 434 int force_exit; 435 436 /* Thread state: */ 437 enum pthread_state state; 438 439 /* 440 * Error variable used instead of errno. The function __error() 441 * returns a pointer to this. 442 */ 443 int error; 444 445 /* 446 * The joiner is the thread that is joining to this thread. The 447 * join status keeps track of a join operation to another thread. 448 */ 449 struct pthread *joiner; 450 451 /* Miscellaneous flags; only set with scheduling lock held. */ 452 int flags; 453#define THR_FLAGS_PRIVATE 0x0001 454#define THR_FLAGS_NEED_SUSPEND 0x0002 /* thread should be suspended */ 455#define THR_FLAGS_SUSPENDED 0x0004 /* thread is suspended */ 456#define THR_FLAGS_DETACHED 0x0008 /* thread is detached */ 457 458 /* Thread list flags; only set with thread list lock held. */ 459 int tlflags; 460#define TLFLAGS_GC_SAFE 0x0001 /* thread safe for cleaning */ 461#define TLFLAGS_IN_TDLIST 0x0002 /* thread in all thread list */ 462#define TLFLAGS_IN_GCLIST 0x0004 /* thread in gc list */ 463 464 /* Queue of currently owned NORMAL or PRIO_INHERIT type mutexes. */ 465 struct mutex_queue mutexq; 466 467 /* Queue of all owned PRIO_PROTECT mutexes. */ 468 struct mutex_queue pp_mutexq; 469 470 void *ret; 471 struct pthread_specific_elem *specific; 472 int specific_data_count; 473 474 /* Number rwlocks rdlocks held. */ 475 int rdlock_count; 476 477 /* 478 * Current locks bitmap for rtld. */ 479 int rtld_bits; 480 481 /* Thread control block */ 482 struct tcb *tcb; 483 484 /* Cleanup handlers Link List */ 485 struct pthread_cleanup *cleanup; 486 487#ifdef _PTHREAD_FORCED_UNWIND 488 struct _Unwind_Exception ex; 489 void *unwind_stackend; 490 int unwind_disabled; 491#endif 492 493 /* 494 * Magic value to help recognize a valid thread structure 495 * from an invalid one: 496 */ 497#define THR_MAGIC ((u_int32_t) 0xd09ba115) 498 u_int32_t magic; 499 500 /* Enable event reporting */ 501 int report_events; 502 503 /* Event mask */ 504 int event_mask; 505 506 /* Event */ 507 td_event_msg_t event_buf;
|
| 508 509 struct wake_addr *wake_addr; 510#define WAKE_ADDR(td) ((td)->wake_addr) 511 512 /* Sleep queue */ 513 struct sleepqueue *sleepqueue; 514 515 /* Wait channel */ 516 void *wchan; 517 518 /* Referenced mutex. */ 519 struct pthread_mutex *mutex_obj; 520 521 /* Thread will sleep. */ 522 int will_sleep; 523 524 /* Number of threads deferred. */ 525 int nwaiter_defer; 526 527 /* Deferred threads from pthread_cond_signal. */ 528 unsigned int *defer_waiters[MAX_DEFER_WAITERS];
|
485}; 486 487#define THR_SHOULD_GC(thrd) \ 488 ((thrd)->refcount == 0 && (thrd)->state == PS_DEAD && \ 489 ((thrd)->flags & THR_FLAGS_DETACHED) != 0) 490 491#define THR_IN_CRITICAL(thrd) \ 492 (((thrd)->locklevel > 0) || \ 493 ((thrd)->critical_count > 0)) 494 495#define THR_CRITICAL_ENTER(thrd) \ 496 (thrd)->critical_count++ 497 498#define THR_CRITICAL_LEAVE(thrd) \ 499 do { \ 500 (thrd)->critical_count--; \ 501 _thr_ast(thrd); \ 502 } while (0) 503 504#define THR_UMUTEX_TRYLOCK(thrd, lck) \ 505 _thr_umutex_trylock((lck), TID(thrd)) 506 507#define THR_UMUTEX_LOCK(thrd, lck) \ 508 _thr_umutex_lock((lck), TID(thrd)) 509 510#define THR_UMUTEX_TIMEDLOCK(thrd, lck, timo) \ 511 _thr_umutex_timedlock((lck), TID(thrd), (timo)) 512 513#define THR_UMUTEX_UNLOCK(thrd, lck) \ 514 _thr_umutex_unlock((lck), TID(thrd)) 515 516#define THR_LOCK_ACQUIRE(thrd, lck) \ 517do { \ 518 (thrd)->locklevel++; \ 519 _thr_umutex_lock(lck, TID(thrd)); \ 520} while (0) 521
| 529}; 530 531#define THR_SHOULD_GC(thrd) \ 532 ((thrd)->refcount == 0 && (thrd)->state == PS_DEAD && \ 533 ((thrd)->flags & THR_FLAGS_DETACHED) != 0) 534 535#define THR_IN_CRITICAL(thrd) \ 536 (((thrd)->locklevel > 0) || \ 537 ((thrd)->critical_count > 0)) 538 539#define THR_CRITICAL_ENTER(thrd) \ 540 (thrd)->critical_count++ 541 542#define THR_CRITICAL_LEAVE(thrd) \ 543 do { \ 544 (thrd)->critical_count--; \ 545 _thr_ast(thrd); \ 546 } while (0) 547 548#define THR_UMUTEX_TRYLOCK(thrd, lck) \ 549 _thr_umutex_trylock((lck), TID(thrd)) 550 551#define THR_UMUTEX_LOCK(thrd, lck) \ 552 _thr_umutex_lock((lck), TID(thrd)) 553 554#define THR_UMUTEX_TIMEDLOCK(thrd, lck, timo) \ 555 _thr_umutex_timedlock((lck), TID(thrd), (timo)) 556 557#define THR_UMUTEX_UNLOCK(thrd, lck) \ 558 _thr_umutex_unlock((lck), TID(thrd)) 559 560#define THR_LOCK_ACQUIRE(thrd, lck) \ 561do { \ 562 (thrd)->locklevel++; \ 563 _thr_umutex_lock(lck, TID(thrd)); \ 564} while (0) 565
|
| 566#define THR_LOCK_ACQUIRE_SPIN(thrd, lck) \ 567do { \ 568 (thrd)->locklevel++; \ 569 _thr_umutex_lock_spin(lck, TID(thrd)); \ 570} while (0) 571
|
522#ifdef _PTHREADS_INVARIANTS 523#define THR_ASSERT_LOCKLEVEL(thrd) \ 524do { \ 525 if (__predict_false((thrd)->locklevel <= 0)) \ 526 _thr_assert_lock_level(); \ 527} while (0) 528#else 529#define THR_ASSERT_LOCKLEVEL(thrd) 530#endif 531 532#define THR_LOCK_RELEASE(thrd, lck) \ 533do { \ 534 THR_ASSERT_LOCKLEVEL(thrd); \ 535 _thr_umutex_unlock((lck), TID(thrd)); \ 536 (thrd)->locklevel--; \ 537 _thr_ast(thrd); \ 538} while (0) 539 540#define THR_LOCK(curthrd) THR_LOCK_ACQUIRE(curthrd, &(curthrd)->lock) 541#define THR_UNLOCK(curthrd) THR_LOCK_RELEASE(curthrd, &(curthrd)->lock) 542#define THR_THREAD_LOCK(curthrd, thr) THR_LOCK_ACQUIRE(curthrd, &(thr)->lock) 543#define THR_THREAD_UNLOCK(curthrd, thr) THR_LOCK_RELEASE(curthrd, &(thr)->lock) 544 545#define THREAD_LIST_RDLOCK(curthrd) \ 546do { \ 547 (curthrd)->locklevel++; \ 548 _thr_rwl_rdlock(&_thr_list_lock); \ 549} while (0) 550 551#define THREAD_LIST_WRLOCK(curthrd) \ 552do { \ 553 (curthrd)->locklevel++; \ 554 _thr_rwl_wrlock(&_thr_list_lock); \ 555} while (0) 556 557#define THREAD_LIST_UNLOCK(curthrd) \ 558do { \ 559 _thr_rwl_unlock(&_thr_list_lock); \ 560 (curthrd)->locklevel--; \ 561 _thr_ast(curthrd); \ 562} while (0) 563 564/* 565 * Macros to insert/remove threads to the all thread list and 566 * the gc list. 567 */ 568#define THR_LIST_ADD(thrd) do { \ 569 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) == 0) { \ 570 TAILQ_INSERT_HEAD(&_thread_list, thrd, tle); \ 571 _thr_hash_add(thrd); \ 572 (thrd)->tlflags |= TLFLAGS_IN_TDLIST; \ 573 } \ 574} while (0) 575#define THR_LIST_REMOVE(thrd) do { \ 576 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) != 0) { \ 577 TAILQ_REMOVE(&_thread_list, thrd, tle); \ 578 _thr_hash_remove(thrd); \ 579 (thrd)->tlflags &= ~TLFLAGS_IN_TDLIST; \ 580 } \ 581} while (0) 582#define THR_GCLIST_ADD(thrd) do { \ 583 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) == 0) { \ 584 TAILQ_INSERT_HEAD(&_thread_gc_list, thrd, gcle);\ 585 (thrd)->tlflags |= TLFLAGS_IN_GCLIST; \ 586 _gc_count++; \ 587 } \ 588} while (0) 589#define THR_GCLIST_REMOVE(thrd) do { \ 590 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) != 0) { \ 591 TAILQ_REMOVE(&_thread_gc_list, thrd, gcle); \ 592 (thrd)->tlflags &= ~TLFLAGS_IN_GCLIST; \ 593 _gc_count--; \ 594 } \ 595} while (0) 596 597#define THR_REF_ADD(curthread, pthread) { \ 598 THR_CRITICAL_ENTER(curthread); \ 599 pthread->refcount++; \ 600} while (0) 601 602#define THR_REF_DEL(curthread, pthread) { \ 603 pthread->refcount--; \ 604 THR_CRITICAL_LEAVE(curthread); \ 605} while (0) 606 607#define GC_NEEDED() (_gc_count >= 5) 608 609#define SHOULD_REPORT_EVENT(curthr, e) \ 610 (curthr->report_events && \ 611 (((curthr)->event_mask | _thread_event_mask ) & e) != 0) 612 613extern int __isthreaded; 614 615/* 616 * Global variables for the pthread kernel. 617 */ 618 619extern char *_usrstack __hidden; 620extern struct pthread *_thr_initial __hidden; 621 622/* For debugger */ 623extern int _libthr_debug; 624extern int _thread_event_mask; 625extern struct pthread *_thread_last_event; 626 627/* List of all threads: */ 628extern pthreadlist _thread_list; 629 630/* List of threads needing GC: */ 631extern pthreadlist _thread_gc_list __hidden; 632 633extern int _thread_active_threads; 634extern atfork_head _thr_atfork_list __hidden; 635extern struct urwlock _thr_atfork_lock __hidden; 636 637/* Default thread attributes: */ 638extern struct pthread_attr _pthread_attr_default __hidden; 639 640/* Default mutex attributes: */ 641extern struct pthread_mutex_attr _pthread_mutexattr_default __hidden; 642extern struct pthread_mutex_attr _pthread_mutexattr_adaptive_default __hidden; 643 644/* Default condition variable attributes: */ 645extern struct pthread_cond_attr _pthread_condattr_default __hidden; 646 647extern struct pthread_prio _thr_priorities[] __hidden; 648 649extern pid_t _thr_pid __hidden; 650extern int _thr_is_smp __hidden; 651 652extern size_t _thr_guard_default __hidden; 653extern size_t _thr_stack_default __hidden; 654extern size_t _thr_stack_initial __hidden; 655extern int _thr_page_size __hidden; 656extern int _thr_spinloops __hidden; 657extern int _thr_yieldloops __hidden; 658 659/* Garbage thread count. */ 660extern int _gc_count __hidden; 661 662extern struct umutex _mutex_static_lock __hidden; 663extern struct umutex _cond_static_lock __hidden; 664extern struct umutex _rwlock_static_lock __hidden; 665extern struct umutex _keytable_lock __hidden; 666extern struct urwlock _thr_list_lock __hidden; 667extern struct umutex _thr_event_lock __hidden; 668 669/* 670 * Function prototype definitions. 671 */ 672__BEGIN_DECLS 673int _thr_setthreaded(int) __hidden;
| 572#ifdef _PTHREADS_INVARIANTS 573#define THR_ASSERT_LOCKLEVEL(thrd) \ 574do { \ 575 if (__predict_false((thrd)->locklevel <= 0)) \ 576 _thr_assert_lock_level(); \ 577} while (0) 578#else 579#define THR_ASSERT_LOCKLEVEL(thrd) 580#endif 581 582#define THR_LOCK_RELEASE(thrd, lck) \ 583do { \ 584 THR_ASSERT_LOCKLEVEL(thrd); \ 585 _thr_umutex_unlock((lck), TID(thrd)); \ 586 (thrd)->locklevel--; \ 587 _thr_ast(thrd); \ 588} while (0) 589 590#define THR_LOCK(curthrd) THR_LOCK_ACQUIRE(curthrd, &(curthrd)->lock) 591#define THR_UNLOCK(curthrd) THR_LOCK_RELEASE(curthrd, &(curthrd)->lock) 592#define THR_THREAD_LOCK(curthrd, thr) THR_LOCK_ACQUIRE(curthrd, &(thr)->lock) 593#define THR_THREAD_UNLOCK(curthrd, thr) THR_LOCK_RELEASE(curthrd, &(thr)->lock) 594 595#define THREAD_LIST_RDLOCK(curthrd) \ 596do { \ 597 (curthrd)->locklevel++; \ 598 _thr_rwl_rdlock(&_thr_list_lock); \ 599} while (0) 600 601#define THREAD_LIST_WRLOCK(curthrd) \ 602do { \ 603 (curthrd)->locklevel++; \ 604 _thr_rwl_wrlock(&_thr_list_lock); \ 605} while (0) 606 607#define THREAD_LIST_UNLOCK(curthrd) \ 608do { \ 609 _thr_rwl_unlock(&_thr_list_lock); \ 610 (curthrd)->locklevel--; \ 611 _thr_ast(curthrd); \ 612} while (0) 613 614/* 615 * Macros to insert/remove threads to the all thread list and 616 * the gc list. 617 */ 618#define THR_LIST_ADD(thrd) do { \ 619 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) == 0) { \ 620 TAILQ_INSERT_HEAD(&_thread_list, thrd, tle); \ 621 _thr_hash_add(thrd); \ 622 (thrd)->tlflags |= TLFLAGS_IN_TDLIST; \ 623 } \ 624} while (0) 625#define THR_LIST_REMOVE(thrd) do { \ 626 if (((thrd)->tlflags & TLFLAGS_IN_TDLIST) != 0) { \ 627 TAILQ_REMOVE(&_thread_list, thrd, tle); \ 628 _thr_hash_remove(thrd); \ 629 (thrd)->tlflags &= ~TLFLAGS_IN_TDLIST; \ 630 } \ 631} while (0) 632#define THR_GCLIST_ADD(thrd) do { \ 633 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) == 0) { \ 634 TAILQ_INSERT_HEAD(&_thread_gc_list, thrd, gcle);\ 635 (thrd)->tlflags |= TLFLAGS_IN_GCLIST; \ 636 _gc_count++; \ 637 } \ 638} while (0) 639#define THR_GCLIST_REMOVE(thrd) do { \ 640 if (((thrd)->tlflags & TLFLAGS_IN_GCLIST) != 0) { \ 641 TAILQ_REMOVE(&_thread_gc_list, thrd, gcle); \ 642 (thrd)->tlflags &= ~TLFLAGS_IN_GCLIST; \ 643 _gc_count--; \ 644 } \ 645} while (0) 646 647#define THR_REF_ADD(curthread, pthread) { \ 648 THR_CRITICAL_ENTER(curthread); \ 649 pthread->refcount++; \ 650} while (0) 651 652#define THR_REF_DEL(curthread, pthread) { \ 653 pthread->refcount--; \ 654 THR_CRITICAL_LEAVE(curthread); \ 655} while (0) 656 657#define GC_NEEDED() (_gc_count >= 5) 658 659#define SHOULD_REPORT_EVENT(curthr, e) \ 660 (curthr->report_events && \ 661 (((curthr)->event_mask | _thread_event_mask ) & e) != 0) 662 663extern int __isthreaded; 664 665/* 666 * Global variables for the pthread kernel. 667 */ 668 669extern char *_usrstack __hidden; 670extern struct pthread *_thr_initial __hidden; 671 672/* For debugger */ 673extern int _libthr_debug; 674extern int _thread_event_mask; 675extern struct pthread *_thread_last_event; 676 677/* List of all threads: */ 678extern pthreadlist _thread_list; 679 680/* List of threads needing GC: */ 681extern pthreadlist _thread_gc_list __hidden; 682 683extern int _thread_active_threads; 684extern atfork_head _thr_atfork_list __hidden; 685extern struct urwlock _thr_atfork_lock __hidden; 686 687/* Default thread attributes: */ 688extern struct pthread_attr _pthread_attr_default __hidden; 689 690/* Default mutex attributes: */ 691extern struct pthread_mutex_attr _pthread_mutexattr_default __hidden; 692extern struct pthread_mutex_attr _pthread_mutexattr_adaptive_default __hidden; 693 694/* Default condition variable attributes: */ 695extern struct pthread_cond_attr _pthread_condattr_default __hidden; 696 697extern struct pthread_prio _thr_priorities[] __hidden; 698 699extern pid_t _thr_pid __hidden; 700extern int _thr_is_smp __hidden; 701 702extern size_t _thr_guard_default __hidden; 703extern size_t _thr_stack_default __hidden; 704extern size_t _thr_stack_initial __hidden; 705extern int _thr_page_size __hidden; 706extern int _thr_spinloops __hidden; 707extern int _thr_yieldloops __hidden; 708 709/* Garbage thread count. */ 710extern int _gc_count __hidden; 711 712extern struct umutex _mutex_static_lock __hidden; 713extern struct umutex _cond_static_lock __hidden; 714extern struct umutex _rwlock_static_lock __hidden; 715extern struct umutex _keytable_lock __hidden; 716extern struct urwlock _thr_list_lock __hidden; 717extern struct umutex _thr_event_lock __hidden; 718 719/* 720 * Function prototype definitions. 721 */ 722__BEGIN_DECLS 723int _thr_setthreaded(int) __hidden;
|
674int _mutex_cv_lock(pthread_mutex_t *, int count) __hidden; 675int _mutex_cv_unlock(pthread_mutex_t *, int *count) __hidden;
| 724int _mutex_cv_lock(struct pthread_mutex *, int count) __hidden; 725int _mutex_cv_unlock(struct pthread_mutex *, int *count) __hidden; 726int _mutex_cv_attach(struct pthread_mutex *, int count) __hidden; 727int _mutex_cv_detach(struct pthread_mutex *, int *count) __hidden; 728int _mutex_owned(struct pthread *, const struct pthread_mutex *) __hidden;
|
676int _mutex_reinit(pthread_mutex_t *) __hidden; 677void _mutex_fork(struct pthread *curthread) __hidden; 678void _libpthread_init(struct pthread *) __hidden; 679struct pthread *_thr_alloc(struct pthread *) __hidden; 680void _thread_exit(const char *, int, const char *) __hidden __dead2; 681int _thr_ref_add(struct pthread *, struct pthread *, int) __hidden; 682void _thr_ref_delete(struct pthread *, struct pthread *) __hidden; 683void _thr_ref_delete_unlocked(struct pthread *, struct pthread *) __hidden; 684int _thr_find_thread(struct pthread *, struct pthread *, int) __hidden; 685void _thr_rtld_init(void) __hidden; 686void _thr_rtld_fini(void) __hidden; 687void _thr_rtld_postfork_child(void) __hidden; 688int _thr_stack_alloc(struct pthread_attr *) __hidden; 689void _thr_stack_free(struct pthread_attr *) __hidden; 690void _thr_free(struct pthread *, struct pthread *) __hidden; 691void _thr_gc(struct pthread *) __hidden; 692void _thread_cleanupspecific(void) __hidden; 693void _thread_printf(int, const char *, ...) __hidden; 694void _thr_spinlock_init(void) __hidden; 695void _thr_cancel_enter(struct pthread *) __hidden; 696void _thr_cancel_enter2(struct pthread *, int) __hidden; 697void _thr_cancel_leave(struct pthread *, int) __hidden; 698void _thr_testcancel(struct pthread *) __hidden; 699void _thr_signal_block(struct pthread *) __hidden; 700void _thr_signal_unblock(struct pthread *) __hidden; 701void _thr_signal_init(void) __hidden; 702void _thr_signal_deinit(void) __hidden; 703int _thr_send_sig(struct pthread *, int sig) __hidden; 704void _thr_list_init(void) __hidden; 705void _thr_hash_add(struct pthread *) __hidden; 706void _thr_hash_remove(struct pthread *) __hidden; 707struct pthread *_thr_hash_find(struct pthread *) __hidden; 708void _thr_link(struct pthread *, struct pthread *) __hidden; 709void _thr_unlink(struct pthread *, struct pthread *) __hidden; 710void _thr_assert_lock_level(void) __hidden __dead2; 711void _thr_ast(struct pthread *) __hidden; 712void _thr_once_init(void) __hidden; 713void _thr_report_creation(struct pthread *curthread, 714 struct pthread *newthread) __hidden; 715void _thr_report_death(struct pthread *curthread) __hidden; 716int _thr_getscheduler(lwpid_t, int *, struct sched_param *) __hidden; 717int _thr_setscheduler(lwpid_t, int, const struct sched_param *) __hidden; 718void _thr_signal_prefork(void) __hidden; 719void _thr_signal_postfork(void) __hidden; 720void _thr_signal_postfork_child(void) __hidden; 721void _thr_try_gc(struct pthread *, struct pthread *) __hidden; 722int _rtp_to_schedparam(const struct rtprio *rtp, int *policy, 723 struct sched_param *param) __hidden; 724int _schedparam_to_rtp(int policy, const struct sched_param *param, 725 struct rtprio *rtp) __hidden; 726void _thread_bp_create(void); 727void _thread_bp_death(void); 728int _sched_yield(void); 729 730void _pthread_cleanup_push(void (*)(void *), void *); 731void _pthread_cleanup_pop(int); 732void _pthread_exit_mask(void *status, sigset_t *mask) __dead2 __hidden; 733void _pthread_cancel_enter(int maycancel); 734void _pthread_cancel_leave(int maycancel); 735 736/* #include <fcntl.h> */ 737#ifdef _SYS_FCNTL_H_ 738int __sys_fcntl(int, int, ...); 739int __sys_open(const char *, int, ...); 740int __sys_openat(int, const char *, int, ...); 741#endif 742 743/* #include <signal.h> */ 744#ifdef _SIGNAL_H_ 745int __sys_kill(pid_t, int); 746int __sys_sigaction(int, const struct sigaction *, struct sigaction *); 747int __sys_sigpending(sigset_t *); 748int __sys_sigprocmask(int, const sigset_t *, sigset_t *); 749int __sys_sigsuspend(const sigset_t *); 750int __sys_sigreturn(const ucontext_t *); 751int __sys_sigaltstack(const struct sigaltstack *, struct sigaltstack *); 752int __sys_sigwait(const sigset_t *, int *); 753int __sys_sigtimedwait(const sigset_t *, siginfo_t *, 754 const struct timespec *); 755int __sys_sigwaitinfo(const sigset_t *set, siginfo_t *info); 756#endif 757 758/* #include <time.h> */ 759#ifdef _TIME_H_ 760int __sys_nanosleep(const struct timespec *, struct timespec *); 761#endif 762 763/* #include <sys/ucontext.h> */ 764#ifdef _SYS_UCONTEXT_H_ 765int __sys_setcontext(const ucontext_t *ucp); 766int __sys_swapcontext(ucontext_t *oucp, const ucontext_t *ucp); 767#endif 768 769/* #include <unistd.h> */ 770#ifdef _UNISTD_H_ 771int __sys_close(int); 772int __sys_fork(void); 773pid_t __sys_getpid(void); 774ssize_t __sys_read(int, void *, size_t); 775ssize_t __sys_write(int, const void *, size_t); 776void __sys_exit(int); 777#endif 778 779int _umtx_op_err(void *, int op, u_long, void *, void *) __hidden; 780 781static inline int 782_thr_isthreaded(void) 783{ 784 return (__isthreaded != 0); 785} 786 787static inline int 788_thr_is_inited(void) 789{ 790 return (_thr_initial != NULL); 791} 792 793static inline void 794_thr_check_init(void) 795{ 796 if (_thr_initial == NULL) 797 _libpthread_init(NULL); 798} 799
| 729int _mutex_reinit(pthread_mutex_t *) __hidden; 730void _mutex_fork(struct pthread *curthread) __hidden; 731void _libpthread_init(struct pthread *) __hidden; 732struct pthread *_thr_alloc(struct pthread *) __hidden; 733void _thread_exit(const char *, int, const char *) __hidden __dead2; 734int _thr_ref_add(struct pthread *, struct pthread *, int) __hidden; 735void _thr_ref_delete(struct pthread *, struct pthread *) __hidden; 736void _thr_ref_delete_unlocked(struct pthread *, struct pthread *) __hidden; 737int _thr_find_thread(struct pthread *, struct pthread *, int) __hidden; 738void _thr_rtld_init(void) __hidden; 739void _thr_rtld_fini(void) __hidden; 740void _thr_rtld_postfork_child(void) __hidden; 741int _thr_stack_alloc(struct pthread_attr *) __hidden; 742void _thr_stack_free(struct pthread_attr *) __hidden; 743void _thr_free(struct pthread *, struct pthread *) __hidden; 744void _thr_gc(struct pthread *) __hidden; 745void _thread_cleanupspecific(void) __hidden; 746void _thread_printf(int, const char *, ...) __hidden; 747void _thr_spinlock_init(void) __hidden; 748void _thr_cancel_enter(struct pthread *) __hidden; 749void _thr_cancel_enter2(struct pthread *, int) __hidden; 750void _thr_cancel_leave(struct pthread *, int) __hidden; 751void _thr_testcancel(struct pthread *) __hidden; 752void _thr_signal_block(struct pthread *) __hidden; 753void _thr_signal_unblock(struct pthread *) __hidden; 754void _thr_signal_init(void) __hidden; 755void _thr_signal_deinit(void) __hidden; 756int _thr_send_sig(struct pthread *, int sig) __hidden; 757void _thr_list_init(void) __hidden; 758void _thr_hash_add(struct pthread *) __hidden; 759void _thr_hash_remove(struct pthread *) __hidden; 760struct pthread *_thr_hash_find(struct pthread *) __hidden; 761void _thr_link(struct pthread *, struct pthread *) __hidden; 762void _thr_unlink(struct pthread *, struct pthread *) __hidden; 763void _thr_assert_lock_level(void) __hidden __dead2; 764void _thr_ast(struct pthread *) __hidden; 765void _thr_once_init(void) __hidden; 766void _thr_report_creation(struct pthread *curthread, 767 struct pthread *newthread) __hidden; 768void _thr_report_death(struct pthread *curthread) __hidden; 769int _thr_getscheduler(lwpid_t, int *, struct sched_param *) __hidden; 770int _thr_setscheduler(lwpid_t, int, const struct sched_param *) __hidden; 771void _thr_signal_prefork(void) __hidden; 772void _thr_signal_postfork(void) __hidden; 773void _thr_signal_postfork_child(void) __hidden; 774void _thr_try_gc(struct pthread *, struct pthread *) __hidden; 775int _rtp_to_schedparam(const struct rtprio *rtp, int *policy, 776 struct sched_param *param) __hidden; 777int _schedparam_to_rtp(int policy, const struct sched_param *param, 778 struct rtprio *rtp) __hidden; 779void _thread_bp_create(void); 780void _thread_bp_death(void); 781int _sched_yield(void); 782 783void _pthread_cleanup_push(void (*)(void *), void *); 784void _pthread_cleanup_pop(int); 785void _pthread_exit_mask(void *status, sigset_t *mask) __dead2 __hidden; 786void _pthread_cancel_enter(int maycancel); 787void _pthread_cancel_leave(int maycancel); 788 789/* #include <fcntl.h> */ 790#ifdef _SYS_FCNTL_H_ 791int __sys_fcntl(int, int, ...); 792int __sys_open(const char *, int, ...); 793int __sys_openat(int, const char *, int, ...); 794#endif 795 796/* #include <signal.h> */ 797#ifdef _SIGNAL_H_ 798int __sys_kill(pid_t, int); 799int __sys_sigaction(int, const struct sigaction *, struct sigaction *); 800int __sys_sigpending(sigset_t *); 801int __sys_sigprocmask(int, const sigset_t *, sigset_t *); 802int __sys_sigsuspend(const sigset_t *); 803int __sys_sigreturn(const ucontext_t *); 804int __sys_sigaltstack(const struct sigaltstack *, struct sigaltstack *); 805int __sys_sigwait(const sigset_t *, int *); 806int __sys_sigtimedwait(const sigset_t *, siginfo_t *, 807 const struct timespec *); 808int __sys_sigwaitinfo(const sigset_t *set, siginfo_t *info); 809#endif 810 811/* #include <time.h> */ 812#ifdef _TIME_H_ 813int __sys_nanosleep(const struct timespec *, struct timespec *); 814#endif 815 816/* #include <sys/ucontext.h> */ 817#ifdef _SYS_UCONTEXT_H_ 818int __sys_setcontext(const ucontext_t *ucp); 819int __sys_swapcontext(ucontext_t *oucp, const ucontext_t *ucp); 820#endif 821 822/* #include <unistd.h> */ 823#ifdef _UNISTD_H_ 824int __sys_close(int); 825int __sys_fork(void); 826pid_t __sys_getpid(void); 827ssize_t __sys_read(int, void *, size_t); 828ssize_t __sys_write(int, const void *, size_t); 829void __sys_exit(int); 830#endif 831 832int _umtx_op_err(void *, int op, u_long, void *, void *) __hidden; 833 834static inline int 835_thr_isthreaded(void) 836{ 837 return (__isthreaded != 0); 838} 839 840static inline int 841_thr_is_inited(void) 842{ 843 return (_thr_initial != NULL); 844} 845 846static inline void 847_thr_check_init(void) 848{ 849 if (_thr_initial == NULL) 850 _libpthread_init(NULL); 851} 852
|
| 853struct wake_addr *_thr_alloc_wake_addr(void); 854void _thr_release_wake_addr(struct wake_addr *); 855int _thr_sleep(struct pthread *, int, const struct timespec *); 856 857void _thr_wake_addr_init(void) __hidden; 858 859static inline void 860_thr_clear_wake(struct pthread *td) 861{ 862 td->wake_addr->value = 0; 863} 864 865static inline int 866_thr_is_woken(struct pthread *td) 867{ 868 return td->wake_addr->value != 0; 869} 870 871static inline void 872_thr_set_wake(unsigned int *waddr) 873{ 874 *waddr = 1; 875 _thr_umtx_wake(waddr, INT_MAX, 0); 876} 877 878void _thr_wake_all(unsigned int *waddrs[], int) __hidden; 879 880static inline struct pthread * 881_sleepq_first(struct sleepqueue *sq) 882{ 883 return TAILQ_FIRST(&sq->sq_blocked); 884} 885 886void _sleepq_init(void) __hidden; 887struct sleepqueue *_sleepq_alloc(void) __hidden; 888void _sleepq_free(struct sleepqueue *) __hidden; 889void _sleepq_lock(void *) __hidden; 890void _sleepq_unlock(void *) __hidden; 891struct sleepqueue *_sleepq_lookup(void *) __hidden; 892void _sleepq_add(void *, struct pthread *) __hidden; 893int _sleepq_remove(struct sleepqueue *, struct pthread *) __hidden; 894void _sleepq_drop(struct sleepqueue *, 895 void (*cb)(struct pthread *, void *arg), void *) __hidden; 896
|
800struct dl_phdr_info; 801void __pthread_cxa_finalize(struct dl_phdr_info *phdr_info); 802void _thr_tsd_unload(struct dl_phdr_info *phdr_info) __hidden; 803void _thr_sigact_unload(struct dl_phdr_info *phdr_info) __hidden; 804 805__END_DECLS 806 807#endif /* !_THR_PRIVATE_H */
| 897struct dl_phdr_info; 898void __pthread_cxa_finalize(struct dl_phdr_info *phdr_info); 899void _thr_tsd_unload(struct dl_phdr_info *phdr_info) __hidden; 900void _thr_sigact_unload(struct dl_phdr_info *phdr_info) __hidden; 901 902__END_DECLS 903 904#endif /* !_THR_PRIVATE_H */
|