|
@@ -1,1165 +0,0 @@
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-#include <errno.h>
|
|
|
-#include <netdb.h> /* for h_errno */
|
|
|
-#include <stddef.h>
|
|
|
-#include <stdio.h>
|
|
|
-#include <stdlib.h>
|
|
|
-#include <string.h>
|
|
|
-#include <unistd.h>
|
|
|
-#include <fcntl.h>
|
|
|
-#include <sys/wait.h>
|
|
|
-#include <sys/resource.h>
|
|
|
-#include "pthread.h"
|
|
|
-#include "internals.h"
|
|
|
-#include "spinlock.h"
|
|
|
-#include "restart.h"
|
|
|
-#include "debug.h"
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-#include <signal.h>
|
|
|
-#include <sys/types.h>
|
|
|
-#include <sys/syscall.h>
|
|
|
-
|
|
|
-libpthread_hidden_proto(waitpid)
|
|
|
-libpthread_hidden_proto(raise)
|
|
|
-
|
|
|
-
|
|
|
-extern int _errno;
|
|
|
-extern int _h_errno;
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-struct _pthread_descr_struct __pthread_initial_thread = {
|
|
|
- &__pthread_initial_thread,
|
|
|
- &__pthread_initial_thread,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- PTHREAD_THREADS_MAX,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- &__pthread_handles[0].h_lock,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- &_errno,
|
|
|
- 0,
|
|
|
- &_h_errno,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- PTHREAD_START_ARGS_INITIALIZER,
|
|
|
- {NULL},
|
|
|
- {NULL},
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- &__pthread_initial_thread,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- {{{0, }}, 0, NULL},
|
|
|
- __ATOMIC_INITIALIZER,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- 0
|
|
|
-#ifdef __UCLIBC_HAS_XLOCALE__
|
|
|
- ,
|
|
|
- &__global_locale_data,
|
|
|
-#endif
|
|
|
-};
|
|
|
-
|
|
|
-
|
|
|
- variables, the p_pid and p_priority fields,
|
|
|
- and the address for identification. */
|
|
|
-#define manager_thread (&__pthread_manager_thread)
|
|
|
-struct _pthread_descr_struct __pthread_manager_thread = {
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- &__pthread_handles[1].h_lock,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- &__pthread_manager_thread.p_errno,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- PTHREAD_START_ARGS_INITIALIZER,
|
|
|
- {NULL},
|
|
|
- {NULL},
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- 0,
|
|
|
- &__pthread_manager_thread,
|
|
|
- 1,
|
|
|
- 0,
|
|
|
- {{{0, }}, 0, NULL},
|
|
|
- __ATOMIC_INITIALIZER,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- 0,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- NULL,
|
|
|
- 0
|
|
|
-#ifdef __UCLIBC_HAS_XLOCALE__
|
|
|
- ,
|
|
|
- &__global_locale_data,
|
|
|
-#endif
|
|
|
-};
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-pthread_descr __pthread_main_thread = &__pthread_initial_thread;
|
|
|
-
|
|
|
-
|
|
|
- stacks of other threads (below). Aligned on a STACK_SIZE boundary. */
|
|
|
-
|
|
|
-char *__pthread_initial_thread_bos = NULL;
|
|
|
-
|
|
|
-#ifndef __ARCH_USE_MMU__
|
|
|
-
|
|
|
-char *__pthread_initial_thread_tos = NULL;
|
|
|
-char *__pthread_initial_thread_mid = NULL;
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-int __pthread_manager_request = -1;
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-int __pthread_manager_reader;
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-char *__pthread_manager_thread_bos = NULL;
|
|
|
-char *__pthread_manager_thread_tos = NULL;
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-int __pthread_exit_requested = 0;
|
|
|
-int __pthread_exit_code = 0;
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-const int __pthread_threads_max = PTHREAD_THREADS_MAX;
|
|
|
-const int __pthread_sizeof_handle = sizeof(struct pthread_handle_struct);
|
|
|
-const int __pthread_offsetof_descr = offsetof(struct pthread_handle_struct, h_descr);
|
|
|
-const int __pthread_offsetof_pid = offsetof(struct _pthread_descr_struct,
|
|
|
- p_pid);
|
|
|
-const int __linuxthreads_pthread_sizeof_descr
|
|
|
- = sizeof(struct _pthread_descr_struct);
|
|
|
-
|
|
|
-const int __linuxthreads_initial_report_events;
|
|
|
-
|
|
|
-const char __linuxthreads_version[] = VERSION;
|
|
|
-
|
|
|
-
|
|
|
-static void pthread_onexit_process(int retcode, void *arg);
|
|
|
-static void pthread_handle_sigcancel(int sig);
|
|
|
-static void pthread_handle_sigrestart(int sig);
|
|
|
-static void pthread_handle_sigdebug(int sig);
|
|
|
-int __pthread_timedsuspend_new(pthread_descr self, const struct timespec *abstime);
|
|
|
-
|
|
|
-
|
|
|
- In these variables we keep track of the used variables. If the
|
|
|
- platform does not support any real-time signals we will define the
|
|
|
- values to some unreasonable value which will signal failing of all
|
|
|
- the functions below. */
|
|
|
-#ifndef __NR_rt_sigaction
|
|
|
-static int current_rtmin = -1;
|
|
|
-static int current_rtmax = -1;
|
|
|
-int __pthread_sig_restart = SIGUSR1;
|
|
|
-int __pthread_sig_cancel = SIGUSR2;
|
|
|
-int __pthread_sig_debug;
|
|
|
-#else
|
|
|
-
|
|
|
-#if __SIGRTMAX - __SIGRTMIN >= 3
|
|
|
-static int current_rtmin = __SIGRTMIN + 3;
|
|
|
-static int current_rtmax = __SIGRTMAX;
|
|
|
-int __pthread_sig_restart = __SIGRTMIN;
|
|
|
-int __pthread_sig_cancel = __SIGRTMIN + 1;
|
|
|
-int __pthread_sig_debug = __SIGRTMIN + 2;
|
|
|
-void (*__pthread_restart)(pthread_descr) = __pthread_restart_new;
|
|
|
-void (*__pthread_suspend)(pthread_descr) = __pthread_wait_for_restart_signal;
|
|
|
-int (*__pthread_timedsuspend)(pthread_descr, const struct timespec *) = __pthread_timedsuspend_new;
|
|
|
-#else
|
|
|
-static int current_rtmin = __SIGRTMIN;
|
|
|
-static int current_rtmax = __SIGRTMAX;
|
|
|
-int __pthread_sig_restart = SIGUSR1;
|
|
|
-int __pthread_sig_cancel = SIGUSR2;
|
|
|
-int __pthread_sig_debug;
|
|
|
-void (*__pthread_restart)(pthread_descr) = __pthread_restart_old;
|
|
|
-void (*__pthread_suspend)(pthread_descr) = __pthread_suspend_old;
|
|
|
-int (*__pthread_timedsuspend)(pthread_descr, const struct timespec *) = __pthread_timedsuspend_old;
|
|
|
-
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
-int __libc_current_sigrtmin (void)
|
|
|
-{
|
|
|
- return current_rtmin;
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
-int __libc_current_sigrtmax (void)
|
|
|
-{
|
|
|
- return current_rtmax;
|
|
|
-}
|
|
|
-
|
|
|
-#if 0
|
|
|
-
|
|
|
- priority. Please note that we don't use a lock since we assume
|
|
|
- this function to be called at program start. */
|
|
|
-int __libc_allocate_rtsig (int high);
|
|
|
-int __libc_allocate_rtsig (int high)
|
|
|
-{
|
|
|
- if (current_rtmin == -1 || current_rtmin > current_rtmax)
|
|
|
-
|
|
|
- return -1;
|
|
|
- return high ? current_rtmin++ : current_rtmax--;
|
|
|
-}
|
|
|
-#endif
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
- Initialization is split in two functions:
|
|
|
- - a constructor function that blocks the __pthread_sig_restart signal
|
|
|
- (must do this very early, since the program could capture the signal
|
|
|
- mask with e.g. sigsetjmp before creating the first thread);
|
|
|
- - a regular function called from pthread_create when needed. */
|
|
|
-
|
|
|
-static void pthread_initialize(void) __attribute__((constructor));
|
|
|
-
|
|
|
-libpthread_hidden_proto(pthread_attr_destroy)
|
|
|
-libpthread_hidden_proto(pthread_attr_init)
|
|
|
-libpthread_hidden_proto(pthread_attr_getdetachstate)
|
|
|
-libpthread_hidden_proto(pthread_attr_setdetachstate)
|
|
|
-libpthread_hidden_proto(pthread_attr_getinheritsched)
|
|
|
-libpthread_hidden_proto(pthread_attr_setinheritsched)
|
|
|
-libpthread_hidden_proto(pthread_attr_setschedparam)
|
|
|
-libpthread_hidden_proto(pthread_attr_getschedparam)
|
|
|
-libpthread_hidden_proto(pthread_attr_getschedpolicy)
|
|
|
-libpthread_hidden_proto(pthread_attr_setschedpolicy)
|
|
|
-libpthread_hidden_proto(pthread_attr_getscope)
|
|
|
-libpthread_hidden_proto(pthread_attr_setscope)
|
|
|
-
|
|
|
-libpthread_hidden_proto(pthread_exit)
|
|
|
-
|
|
|
-libpthread_hidden_proto(pthread_equal)
|
|
|
-libpthread_hidden_proto(pthread_self)
|
|
|
-libpthread_hidden_proto(pthread_getschedparam)
|
|
|
-libpthread_hidden_proto(pthread_setschedparam)
|
|
|
-
|
|
|
-libpthread_hidden_proto(pthread_setcancelstate)
|
|
|
-libpthread_hidden_proto(pthread_setcanceltype)
|
|
|
-libpthread_hidden_proto(_pthread_cleanup_push_defer)
|
|
|
-libpthread_hidden_proto(_pthread_cleanup_pop_restore)
|
|
|
-
|
|
|
-libpthread_hidden_proto(pthread_cond_broadcast)
|
|
|
-libpthread_hidden_proto(pthread_cond_destroy)
|
|
|
-libpthread_hidden_proto(pthread_cond_init)
|
|
|
-libpthread_hidden_proto(pthread_cond_signal)
|
|
|
-libpthread_hidden_proto(pthread_cond_wait)
|
|
|
-libpthread_hidden_proto(pthread_cond_timedwait)
|
|
|
-
|
|
|
-libpthread_hidden_proto(pthread_condattr_destroy)
|
|
|
-libpthread_hidden_proto(pthread_condattr_init)
|
|
|
-
|
|
|
-struct pthread_functions __pthread_functions =
|
|
|
- {
|
|
|
-#if !defined __UCLIBC_HAS_TLS__ && defined __UCLIBC_HAS_RPC__
|
|
|
- .ptr_pthread_internal_tsd_set = __pthread_internal_tsd_set,
|
|
|
- .ptr_pthread_internal_tsd_get = __pthread_internal_tsd_get,
|
|
|
- .ptr_pthread_internal_tsd_address = __pthread_internal_tsd_address,
|
|
|
-#endif
|
|
|
-
|
|
|
- .ptr_pthread_fork = __pthread_fork,
|
|
|
-*/
|
|
|
- .ptr_pthread_attr_destroy = pthread_attr_destroy,
|
|
|
- .ptr_pthread_attr_init = pthread_attr_init,
|
|
|
- .ptr_pthread_attr_getdetachstate = pthread_attr_getdetachstate,
|
|
|
- .ptr_pthread_attr_setdetachstate = pthread_attr_setdetachstate,
|
|
|
- .ptr_pthread_attr_getinheritsched = pthread_attr_getinheritsched,
|
|
|
- .ptr_pthread_attr_setinheritsched = pthread_attr_setinheritsched,
|
|
|
- .ptr_pthread_attr_getschedparam = pthread_attr_getschedparam,
|
|
|
- .ptr_pthread_attr_setschedparam = pthread_attr_setschedparam,
|
|
|
- .ptr_pthread_attr_getschedpolicy = pthread_attr_getschedpolicy,
|
|
|
- .ptr_pthread_attr_setschedpolicy = pthread_attr_setschedpolicy,
|
|
|
- .ptr_pthread_attr_getscope = pthread_attr_getscope,
|
|
|
- .ptr_pthread_attr_setscope = pthread_attr_setscope,
|
|
|
- .ptr_pthread_condattr_destroy = pthread_condattr_destroy,
|
|
|
- .ptr_pthread_condattr_init = pthread_condattr_init,
|
|
|
- .ptr_pthread_cond_broadcast = pthread_cond_broadcast,
|
|
|
- .ptr_pthread_cond_destroy = pthread_cond_destroy,
|
|
|
- .ptr_pthread_cond_init = pthread_cond_init,
|
|
|
- .ptr_pthread_cond_signal = pthread_cond_signal,
|
|
|
- .ptr_pthread_cond_wait = pthread_cond_wait,
|
|
|
- .ptr_pthread_cond_timedwait = pthread_cond_timedwait,
|
|
|
- .ptr_pthread_equal = pthread_equal,
|
|
|
- .ptr___pthread_exit = pthread_exit,
|
|
|
- .ptr_pthread_getschedparam = pthread_getschedparam,
|
|
|
- .ptr_pthread_setschedparam = pthread_setschedparam,
|
|
|
- .ptr_pthread_mutex_destroy = __pthread_mutex_destroy,
|
|
|
- .ptr_pthread_mutex_init = __pthread_mutex_init,
|
|
|
- .ptr_pthread_mutex_lock = __pthread_mutex_lock,
|
|
|
- .ptr_pthread_mutex_trylock = __pthread_mutex_trylock,
|
|
|
- .ptr_pthread_mutex_unlock = __pthread_mutex_unlock,
|
|
|
- .ptr_pthread_self = pthread_self,
|
|
|
- .ptr_pthread_setcancelstate = pthread_setcancelstate,
|
|
|
- .ptr_pthread_setcanceltype = pthread_setcanceltype,
|
|
|
-
|
|
|
- .ptr_pthread_do_exit = pthread_do_exit,
|
|
|
- .ptr_pthread_thread_self = pthread_thread_self,
|
|
|
- .ptr_pthread_cleanup_upto = pthread_cleanup_upto,
|
|
|
- .ptr_pthread_sigaction = pthread_sigaction,
|
|
|
- .ptr_pthread_sigwait = pthread_sigwait,
|
|
|
- .ptr_pthread_raise = pthread_raise,
|
|
|
- .ptr__pthread_cleanup_push = _pthread_cleanup_push,
|
|
|
- .ptr__pthread_cleanup_pop = _pthread_cleanup_pop,
|
|
|
-*/
|
|
|
- .ptr__pthread_cleanup_push_defer = __pthread_cleanup_push_defer,
|
|
|
- .ptr__pthread_cleanup_pop_restore = __pthread_cleanup_pop_restore
|
|
|
- };
|
|
|
-#ifdef SHARED
|
|
|
-# define ptr_pthread_functions &__pthread_functions
|
|
|
-#else
|
|
|
-# define ptr_pthread_functions NULL
|
|
|
-#endif
|
|
|
-
|
|
|
-static int *__libc_multiple_threads_ptr;
|
|
|
-
|
|
|
-
|
|
|
- startup of the C library. */
|
|
|
-void __pthread_initialize_minimal(void)
|
|
|
-{
|
|
|
-
|
|
|
- * that for the main thread now. */
|
|
|
-#ifdef INIT_THREAD_SELF
|
|
|
- INIT_THREAD_SELF(&__pthread_initial_thread, 0);
|
|
|
-#endif
|
|
|
-
|
|
|
- __libc_multiple_threads_ptr = __libc_pthread_init (ptr_pthread_functions);
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
-static void pthread_initialize(void)
|
|
|
-{
|
|
|
- struct sigaction sa;
|
|
|
- sigset_t mask;
|
|
|
-#ifdef __ARCH_USE_MMU__
|
|
|
- struct rlimit limit;
|
|
|
- rlim_t max_stack;
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
- if (__pthread_initial_thread_bos != NULL) return;
|
|
|
-#ifdef TEST_FOR_COMPARE_AND_SWAP
|
|
|
-
|
|
|
- __pthread_has_cas = compare_and_swap_is_available();
|
|
|
-#endif
|
|
|
-
|
|
|
- below the current stack address, and align that on a
|
|
|
- STACK_SIZE boundary. */
|
|
|
- __pthread_initial_thread_bos =
|
|
|
- (char *)(((long)CURRENT_STACK_FRAME - 2 * STACK_SIZE) & ~(STACK_SIZE - 1));
|
|
|
-
|
|
|
- __pthread_initial_thread.p_pid = getpid();
|
|
|
-
|
|
|
- main thread now. */
|
|
|
-#ifdef INIT_THREAD_SELF
|
|
|
- INIT_THREAD_SELF(&__pthread_initial_thread, 0);
|
|
|
-#endif
|
|
|
-
|
|
|
- __pthread_initial_thread.p_errnop = &_errno;
|
|
|
- __pthread_initial_thread.p_h_errnop = &_h_errno;
|
|
|
-
|
|
|
-#ifdef __UCLIBC_HAS_XLOCALE__
|
|
|
-
|
|
|
- __pthread_initial_thread.locale = __curlocale_var;
|
|
|
-#endif
|
|
|
-
|
|
|
- {
|
|
|
- FILE *fp;
|
|
|
-
|
|
|
- _stdio_user_locking = 0;
|
|
|
- for (fp = _stdio_openlist; fp != NULL; fp = fp->__nextopen) {
|
|
|
- if (fp->__user_locking != 1) {
|
|
|
- fp->__user_locking = 0;
|
|
|
- }
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
- beyond STACK_SIZE minus two pages (one page for the thread descriptor
|
|
|
- immediately beyond, and one page to act as a guard page). */
|
|
|
-
|
|
|
-#ifdef __ARCH_USE_MMU__
|
|
|
-
|
|
|
- * on a non-MMU system. Thus, we don't need the rlimit either. -StS */
|
|
|
- getrlimit(RLIMIT_STACK, &limit);
|
|
|
- max_stack = STACK_SIZE - 2 * getpagesize();
|
|
|
- if (limit.rlim_cur > max_stack) {
|
|
|
- limit.rlim_cur = max_stack;
|
|
|
- setrlimit(RLIMIT_STACK, &limit);
|
|
|
- }
|
|
|
-#else
|
|
|
-
|
|
|
- * We don't have a way of knowing where the kernel started things -- top
|
|
|
- * or bottom (well, that isn't exactly true, but the solution is fairly
|
|
|
- * complex and error prone). All we can determine here is an address
|
|
|
- * that lies within that stack. Save that address as a reference so that
|
|
|
- * as other thread stacks are created, we can adjust the estimated bounds
|
|
|
- * of the initial thread's stack appropriately.
|
|
|
- *
|
|
|
- * This checking is handled in NOMMU_INITIAL_THREAD_BOUNDS(), so see that
|
|
|
- * for a few more details.
|
|
|
- */
|
|
|
- __pthread_initial_thread_mid = CURRENT_STACK_FRAME;
|
|
|
- __pthread_initial_thread_tos = (char *) -1;
|
|
|
- __pthread_initial_thread_bos = (char *) 1;
|
|
|
- PDEBUG("initial thread stack bounds: bos=%p, tos=%p\n",
|
|
|
- __pthread_initial_thread_bos, __pthread_initial_thread_tos);
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
- Since signal handlers are shared between threads, these settings
|
|
|
- will be inherited by all other threads. */
|
|
|
- memset(&sa, 0, sizeof(sa));
|
|
|
- sa.sa_handler = pthread_handle_sigrestart;
|
|
|
- __libc_sigaction(__pthread_sig_restart, &sa, NULL);
|
|
|
- sa.sa_handler = pthread_handle_sigcancel;
|
|
|
- sigaddset(&sa.sa_mask, __pthread_sig_restart);
|
|
|
- __libc_sigaction(__pthread_sig_cancel, &sa, NULL);
|
|
|
- if (__pthread_sig_debug > 0) {
|
|
|
- sa.sa_handler = pthread_handle_sigdebug;
|
|
|
- __sigemptyset(&sa.sa_mask);
|
|
|
- __libc_sigaction(__pthread_sig_debug, &sa, NULL);
|
|
|
- }
|
|
|
-
|
|
|
- __sigemptyset(&mask);
|
|
|
- sigaddset(&mask, __pthread_sig_restart);
|
|
|
- sigprocmask(SIG_BLOCK, &mask, NULL);
|
|
|
-
|
|
|
- sigdelset(&mask, __pthread_sig_restart);
|
|
|
- sigaddset(&mask, __pthread_sig_cancel);
|
|
|
- sigprocmask(SIG_UNBLOCK, &mask, NULL);
|
|
|
-
|
|
|
-
|
|
|
- before pthread_onexit_process. */
|
|
|
- on_exit(pthread_onexit_process, NULL);
|
|
|
-}
|
|
|
-
|
|
|
-void __pthread_initialize(void);
|
|
|
-void __pthread_initialize(void)
|
|
|
-{
|
|
|
- pthread_initialize();
|
|
|
-}
|
|
|
-
|
|
|
-int __pthread_initialize_manager(void)
|
|
|
-{
|
|
|
- int manager_pipe[2];
|
|
|
- int pid;
|
|
|
- int report_events;
|
|
|
- struct pthread_request request;
|
|
|
-
|
|
|
- *__libc_multiple_threads_ptr = 1;
|
|
|
-
|
|
|
-
|
|
|
- constructor run before our constructor), do it now */
|
|
|
- if (__pthread_initial_thread_bos == NULL) pthread_initialize();
|
|
|
-
|
|
|
- __pthread_manager_thread_bos = malloc(THREAD_MANAGER_STACK_SIZE);
|
|
|
- if (__pthread_manager_thread_bos == NULL) return -1;
|
|
|
- __pthread_manager_thread_tos =
|
|
|
- __pthread_manager_thread_bos + THREAD_MANAGER_STACK_SIZE;
|
|
|
-
|
|
|
-
|
|
|
- * with the manager stack frame */
|
|
|
- NOMMU_INITIAL_THREAD_BOUNDS(__pthread_manager_thread_tos,__pthread_manager_thread_bos);
|
|
|
- PDEBUG("manager stack: size=%ld, bos=%p, tos=%p\n", THREAD_MANAGER_STACK_SIZE,
|
|
|
- __pthread_manager_thread_bos, __pthread_manager_thread_tos);
|
|
|
-#if 0
|
|
|
- PDEBUG("initial stack: estimate bos=%p, tos=%p\n",
|
|
|
- __pthread_initial_thread_bos, __pthread_initial_thread_tos);
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
- if (pipe(manager_pipe) == -1) {
|
|
|
- free(__pthread_manager_thread_bos);
|
|
|
- return -1;
|
|
|
- }
|
|
|
-
|
|
|
- pid = 0;
|
|
|
-#if defined(USE_TLS) && USE_TLS
|
|
|
- if (__linuxthreads_initial_report_events != 0)
|
|
|
- THREAD_SETMEM (((pthread_descr) NULL), p_report_events,
|
|
|
- __linuxthreads_initial_report_events);
|
|
|
- report_events = THREAD_GETMEM (((pthread_descr) NULL), p_report_events);
|
|
|
-#else
|
|
|
- if (__linuxthreads_initial_report_events != 0)
|
|
|
- __pthread_initial_thread.p_report_events
|
|
|
- = __linuxthreads_initial_report_events;
|
|
|
- report_events = __pthread_initial_thread.p_report_events;
|
|
|
-#endif
|
|
|
- if (__builtin_expect (report_events, 0))
|
|
|
- {
|
|
|
-
|
|
|
- the manager thread. */
|
|
|
- int idx = __td_eventword (TD_CREATE);
|
|
|
- uint32_t mask = __td_eventmask (TD_CREATE);
|
|
|
-
|
|
|
- if ((mask & (__pthread_threads_events.event_bits[idx]
|
|
|
- | __pthread_initial_thread.p_eventbuf.eventmask.event_bits[idx]))
|
|
|
- != 0)
|
|
|
- {
|
|
|
-
|
|
|
- __pthread_lock(__pthread_manager_thread.p_lock, NULL);
|
|
|
-
|
|
|
-#ifdef __ia64__
|
|
|
- pid = __clone2(__pthread_manager_event,
|
|
|
- (void **) __pthread_manager_thread_tos,
|
|
|
- THREAD_MANAGER_STACK_SIZE,
|
|
|
- CLONE_VM | CLONE_FS | CLONE_FILES | CLONE_SIGHAND,
|
|
|
- (void *)(long)manager_pipe[0]);
|
|
|
-#else
|
|
|
- pid = clone(__pthread_manager_event,
|
|
|
- (void **) __pthread_manager_thread_tos,
|
|
|
- CLONE_VM | CLONE_FS | CLONE_FILES | CLONE_SIGHAND,
|
|
|
- (void *)(long)manager_pipe[0]);
|
|
|
-#endif
|
|
|
-
|
|
|
- if (pid != -1)
|
|
|
- {
|
|
|
-
|
|
|
- the newly created thread's data structure. We cannot let
|
|
|
- the new thread do this since we don't know whether it was
|
|
|
- already scheduled when we send the event. */
|
|
|
- __pthread_manager_thread.p_eventbuf.eventdata =
|
|
|
- &__pthread_manager_thread;
|
|
|
- __pthread_manager_thread.p_eventbuf.eventnum = TD_CREATE;
|
|
|
- __pthread_last_event = &__pthread_manager_thread;
|
|
|
- __pthread_manager_thread.p_tid = 2* PTHREAD_THREADS_MAX + 1;
|
|
|
- __pthread_manager_thread.p_pid = pid;
|
|
|
-
|
|
|
-
|
|
|
- __linuxthreads_create_event ();
|
|
|
- }
|
|
|
-
|
|
|
- __pthread_unlock(__pthread_manager_thread.p_lock);
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- if (pid == 0) {
|
|
|
-#ifdef __ia64__
|
|
|
- pid = __clone2(__pthread_manager, (void **) __pthread_manager_thread_tos,
|
|
|
- THREAD_MANAGER_STACK_SIZE,
|
|
|
- CLONE_VM | CLONE_FS | CLONE_FILES | CLONE_SIGHAND,
|
|
|
- (void *)(long)manager_pipe[0]);
|
|
|
-#else
|
|
|
- pid = clone(__pthread_manager, (void **) __pthread_manager_thread_tos,
|
|
|
- CLONE_VM | CLONE_FS | CLONE_FILES | CLONE_SIGHAND,
|
|
|
- (void *)(long)manager_pipe[0]);
|
|
|
-#endif
|
|
|
- }
|
|
|
- if (pid == -1) {
|
|
|
- free(__pthread_manager_thread_bos);
|
|
|
- close(manager_pipe[0]);
|
|
|
- close(manager_pipe[1]);
|
|
|
- return -1;
|
|
|
- }
|
|
|
- __pthread_manager_request = manager_pipe[1];
|
|
|
- __pthread_manager_reader = manager_pipe[0];
|
|
|
- __pthread_manager_thread.p_tid = 2* PTHREAD_THREADS_MAX + 1;
|
|
|
- __pthread_manager_thread.p_pid = pid;
|
|
|
-
|
|
|
-
|
|
|
- if (__pthread_threads_debug && __pthread_sig_debug > 0)
|
|
|
- {
|
|
|
- raise(__pthread_sig_debug);
|
|
|
-
|
|
|
- ready to handle us. */
|
|
|
- __pthread_wait_for_restart_signal(thread_self());
|
|
|
- }
|
|
|
-
|
|
|
- PDEBUG("send REQ_DEBUG to manager thread\n");
|
|
|
- request.req_kind = REQ_DEBUG;
|
|
|
- TEMP_FAILURE_RETRY(write(__pthread_manager_request,
|
|
|
- (char *) &request, sizeof(request)));
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-int pthread_create(pthread_t *thread, const pthread_attr_t *attr,
|
|
|
- void * (*start_routine)(void *), void *arg)
|
|
|
-{
|
|
|
- pthread_descr self = thread_self();
|
|
|
- struct pthread_request request;
|
|
|
- if (__pthread_manager_request < 0) {
|
|
|
- if (__pthread_initialize_manager() < 0) return EAGAIN;
|
|
|
- }
|
|
|
- request.req_thread = self;
|
|
|
- request.req_kind = REQ_CREATE;
|
|
|
- request.req_args.create.attr = attr;
|
|
|
- request.req_args.create.fn = start_routine;
|
|
|
- request.req_args.create.arg = arg;
|
|
|
- sigprocmask(SIG_SETMASK, NULL, &request.req_args.create.mask);
|
|
|
- PDEBUG("write REQ_CREATE to manager thread\n");
|
|
|
- TEMP_FAILURE_RETRY(write(__pthread_manager_request,
|
|
|
- (char *) &request, sizeof(request)));
|
|
|
- PDEBUG("before suspend(self)\n");
|
|
|
- suspend(self);
|
|
|
- PDEBUG("after suspend(self)\n");
|
|
|
- if (THREAD_GETMEM(self, p_retcode) == 0)
|
|
|
- *thread = (pthread_t) THREAD_GETMEM(self, p_retval);
|
|
|
- return THREAD_GETMEM(self, p_retcode);
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-pthread_t pthread_self(void)
|
|
|
-{
|
|
|
- pthread_descr self = thread_self();
|
|
|
- return THREAD_GETMEM(self, p_tid);
|
|
|
-}
|
|
|
-libpthread_hidden_def (pthread_self)
|
|
|
-
|
|
|
-int pthread_equal(pthread_t thread1, pthread_t thread2)
|
|
|
-{
|
|
|
- return thread1 == thread2;
|
|
|
-}
|
|
|
-libpthread_hidden_def (pthread_equal)
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-#ifndef THREAD_SELF
|
|
|
-
|
|
|
-pthread_descr __pthread_find_self(void)
|
|
|
-{
|
|
|
- char * sp = CURRENT_STACK_FRAME;
|
|
|
- pthread_handle h;
|
|
|
-
|
|
|
-
|
|
|
- the manager threads handled specially in thread_self(), so start at 2 */
|
|
|
- h = __pthread_handles + 2;
|
|
|
- while (! (sp <= (char *) h->h_descr && sp >= h->h_bottom)) h++;
|
|
|
-
|
|
|
-#ifdef DEBUG_PT
|
|
|
- if (h->h_descr == NULL) {
|
|
|
- printf("*** %s ERROR descriptor is NULL!!!!! ***\n\n", __FUNCTION__);
|
|
|
- _exit(1);
|
|
|
- }
|
|
|
-#endif
|
|
|
-
|
|
|
- return h->h_descr;
|
|
|
-}
|
|
|
-#else
|
|
|
-
|
|
|
-static pthread_descr thread_self_stack(void)
|
|
|
-{
|
|
|
- char *sp = CURRENT_STACK_FRAME;
|
|
|
- pthread_handle h;
|
|
|
-
|
|
|
- if (sp >= __pthread_manager_thread_bos && sp < __pthread_manager_thread_tos)
|
|
|
- return manager_thread;
|
|
|
- h = __pthread_handles + 2;
|
|
|
-# if defined(USE_TLS) && USE_TLS
|
|
|
- while (h->h_descr == NULL
|
|
|
- || ! (sp <= (char *) h->h_descr->p_stackaddr && sp >= h->h_bottom))
|
|
|
- h++;
|
|
|
-# else
|
|
|
- while (! (sp <= (char *) h->h_descr && sp >= h->h_bottom))
|
|
|
- h++;
|
|
|
-# endif
|
|
|
- return h->h_descr;
|
|
|
-}
|
|
|
-
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-int pthread_setschedparam(pthread_t thread, int policy,
|
|
|
- const struct sched_param *param)
|
|
|
-{
|
|
|
- pthread_handle handle = thread_handle(thread);
|
|
|
- pthread_descr th;
|
|
|
-
|
|
|
- __pthread_lock(&handle->h_lock, NULL);
|
|
|
- if (invalid_handle(handle, thread)) {
|
|
|
- __pthread_unlock(&handle->h_lock);
|
|
|
- return ESRCH;
|
|
|
- }
|
|
|
- th = handle->h_descr;
|
|
|
- if (sched_setscheduler(th->p_pid, policy, param) == -1) {
|
|
|
- __pthread_unlock(&handle->h_lock);
|
|
|
- return errno;
|
|
|
- }
|
|
|
- th->p_priority = policy == SCHED_OTHER ? 0 : param->sched_priority;
|
|
|
- __pthread_unlock(&handle->h_lock);
|
|
|
- if (__pthread_manager_request >= 0)
|
|
|
- __pthread_manager_adjust_prio(th->p_priority);
|
|
|
- return 0;
|
|
|
-}
|
|
|
-libpthread_hidden_def(pthread_setschedparam)
|
|
|
-
|
|
|
-int pthread_getschedparam(pthread_t thread, int *policy,
|
|
|
- struct sched_param *param)
|
|
|
-{
|
|
|
- pthread_handle handle = thread_handle(thread);
|
|
|
- int pid, pol;
|
|
|
-
|
|
|
- __pthread_lock(&handle->h_lock, NULL);
|
|
|
- if (invalid_handle(handle, thread)) {
|
|
|
- __pthread_unlock(&handle->h_lock);
|
|
|
- return ESRCH;
|
|
|
- }
|
|
|
- pid = handle->h_descr->p_pid;
|
|
|
- __pthread_unlock(&handle->h_lock);
|
|
|
- pol = sched_getscheduler(pid);
|
|
|
- if (pol == -1) return errno;
|
|
|
- if (sched_getparam(pid, param) == -1) return errno;
|
|
|
- *policy = pol;
|
|
|
- return 0;
|
|
|
-}
|
|
|
-libpthread_hidden_def(pthread_getschedparam)
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-static void pthread_onexit_process(int retcode, void *arg attribute_unused)
|
|
|
-{
|
|
|
- struct pthread_request request;
|
|
|
- pthread_descr self = thread_self();
|
|
|
-
|
|
|
- if (__pthread_manager_request >= 0) {
|
|
|
- request.req_thread = self;
|
|
|
- request.req_kind = REQ_PROCESS_EXIT;
|
|
|
- request.req_args.exit.code = retcode;
|
|
|
- TEMP_FAILURE_RETRY(write(__pthread_manager_request,
|
|
|
- (char *) &request, sizeof(request)));
|
|
|
- suspend(self);
|
|
|
-
|
|
|
- children, so that timings for main thread account for all threads. */
|
|
|
- if (self == __pthread_main_thread) {
|
|
|
- waitpid(__pthread_manager_thread.p_pid, NULL, __WCLONE);
|
|
|
-
|
|
|
- * (possibly holding locks), free cannot be used any more. */
|
|
|
- __pthread_manager_thread_bos = __pthread_manager_thread_tos = NULL;
|
|
|
- }
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
- in the thread descriptor, and optionally performs a siglongjmp
|
|
|
- (for pthread_cond_timedwait). */
|
|
|
-
|
|
|
-static void pthread_handle_sigrestart(int sig)
|
|
|
-{
|
|
|
- pthread_descr self = thread_self();
|
|
|
- THREAD_SETMEM(self, p_signal, sig);
|
|
|
- if (THREAD_GETMEM(self, p_signal_jmp) != NULL)
|
|
|
- siglongjmp(*THREAD_GETMEM(self, p_signal_jmp), 1);
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
- (in asynchronous mode), for process-wide exit and exec requests.
|
|
|
- For the thread manager thread, redirect the signal to
|
|
|
- __pthread_manager_sighandler. */
|
|
|
-
|
|
|
-static void pthread_handle_sigcancel(int sig)
|
|
|
-{
|
|
|
- pthread_descr self = thread_self();
|
|
|
- sigjmp_buf * jmpbuf;
|
|
|
-
|
|
|
-
|
|
|
- if (self == &__pthread_manager_thread)
|
|
|
- {
|
|
|
-#ifdef THREAD_SELF
|
|
|
-
|
|
|
- initialized, so that the thread register might still point to the
|
|
|
- manager thread. Double check that this is really the manager
|
|
|
- thread. */
|
|
|
- pthread_descr real_self = thread_self_stack();
|
|
|
- if (real_self == &__pthread_manager_thread)
|
|
|
- {
|
|
|
- __pthread_manager_sighandler(sig);
|
|
|
- return;
|
|
|
- }
|
|
|
-
|
|
|
- self = real_self;
|
|
|
-# ifdef INIT_THREAD_SELF
|
|
|
- INIT_THREAD_SELF(self, self->p_nr);
|
|
|
-# endif
|
|
|
-#else
|
|
|
- __pthread_manager_sighandler(sig);
|
|
|
- return;
|
|
|
-#endif
|
|
|
- }
|
|
|
- if (__builtin_expect (__pthread_exit_requested, 0)) {
|
|
|
-
|
|
|
- children, so that timings for main thread account for all threads. */
|
|
|
- if (self == __pthread_main_thread) {
|
|
|
-#if defined(USE_TLS) && USE_TLS
|
|
|
- waitpid(__pthread_manager_thread->p_pid, NULL, __WCLONE);
|
|
|
-#else
|
|
|
- waitpid(__pthread_manager_thread.p_pid, NULL, __WCLONE);
|
|
|
-#endif
|
|
|
- }
|
|
|
- _exit(__pthread_exit_code);
|
|
|
- }
|
|
|
- if (__builtin_expect (THREAD_GETMEM(self, p_canceled), 0)
|
|
|
- && THREAD_GETMEM(self, p_cancelstate) == PTHREAD_CANCEL_ENABLE) {
|
|
|
- if (THREAD_GETMEM(self, p_canceltype) == PTHREAD_CANCEL_ASYNCHRONOUS)
|
|
|
- __pthread_do_exit(PTHREAD_CANCELED, CURRENT_STACK_FRAME);
|
|
|
- jmpbuf = THREAD_GETMEM(self, p_cancel_jmp);
|
|
|
- if (jmpbuf != NULL) {
|
|
|
- THREAD_SETMEM(self, p_cancel_jmp, NULL);
|
|
|
- siglongjmp(*jmpbuf, 1);
|
|
|
- }
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
- The debugging strategy is as follows:
|
|
|
- On reception of a REQ_DEBUG request (sent by new threads created to
|
|
|
- the thread manager under debugging mode), the thread manager throws
|
|
|
- __pthread_sig_debug to itself. The debugger (if active) intercepts
|
|
|
- this signal, takes into account new threads and continue execution
|
|
|
- of the thread manager by propagating the signal because it doesn't
|
|
|
- know what it is specifically done for. In the current implementation,
|
|
|
- the thread manager simply discards it. */
|
|
|
-
|
|
|
-static void pthread_handle_sigdebug(int sig attribute_unused)
|
|
|
-{
|
|
|
-
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
- Close the pipe used for requests and set the main thread to the forked
|
|
|
- thread.
|
|
|
- Notice that we can't free the stack segments, as the forked thread
|
|
|
- may hold pointers into them. */
|
|
|
-
|
|
|
-void __pthread_reset_main_thread(void)
|
|
|
-{
|
|
|
- pthread_descr self = thread_self();
|
|
|
-
|
|
|
- if (__pthread_manager_request != -1) {
|
|
|
-
|
|
|
- free(__pthread_manager_thread_bos);
|
|
|
- __pthread_manager_thread_bos = __pthread_manager_thread_tos = NULL;
|
|
|
-
|
|
|
- close(__pthread_manager_request);
|
|
|
- close(__pthread_manager_reader);
|
|
|
- __pthread_manager_request = __pthread_manager_reader = -1;
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
- THREAD_SETMEM(self, p_pid, getpid());
|
|
|
-
|
|
|
- __pthread_main_thread = self;
|
|
|
- THREAD_SETMEM(self, p_nextlive, self);
|
|
|
- THREAD_SETMEM(self, p_prevlive, self);
|
|
|
-
|
|
|
- THREAD_SETMEM(self, p_errnop, &_errno);
|
|
|
- THREAD_SETMEM(self, p_h_errnop, &_h_errno);
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-void __pthread_kill_other_threads_np(void)
|
|
|
-{
|
|
|
- struct sigaction sa;
|
|
|
-
|
|
|
- pthread_onexit_process(0, NULL);
|
|
|
-
|
|
|
- changes its mind, does not exec(), and creates new threads instead. */
|
|
|
- __pthread_reset_main_thread();
|
|
|
-
|
|
|
- implementation uses since this would be passed to the new
|
|
|
- process. */
|
|
|
- memset(&sa, 0, sizeof(sa));
|
|
|
- if (SIG_DFL)
|
|
|
- sa.sa_handler = SIG_DFL;
|
|
|
- __libc_sigaction(__pthread_sig_restart, &sa, NULL);
|
|
|
- __libc_sigaction(__pthread_sig_cancel, &sa, NULL);
|
|
|
- if (__pthread_sig_debug > 0)
|
|
|
- __libc_sigaction(__pthread_sig_debug, &sa, NULL);
|
|
|
-}
|
|
|
-weak_alias (__pthread_kill_other_threads_np, pthread_kill_other_threads_np)
|
|
|
-
|
|
|
-
|
|
|
-static int current_level;
|
|
|
-
|
|
|
-int __pthread_setconcurrency(int level)
|
|
|
-{
|
|
|
-
|
|
|
- current_level = level;
|
|
|
- return 0;
|
|
|
-}
|
|
|
-weak_alias (__pthread_setconcurrency, pthread_setconcurrency)
|
|
|
-
|
|
|
-int __pthread_getconcurrency(void)
|
|
|
-{
|
|
|
- return current_level;
|
|
|
-}
|
|
|
-weak_alias (__pthread_getconcurrency, pthread_getconcurrency)
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-void __pthread_wait_for_restart_signal(pthread_descr self)
|
|
|
-{
|
|
|
- sigset_t mask;
|
|
|
-
|
|
|
- sigprocmask(SIG_SETMASK, NULL, &mask);
|
|
|
- sigdelset(&mask, __pthread_sig_restart);
|
|
|
- THREAD_SETMEM(self, p_signal, 0);
|
|
|
- do {
|
|
|
- sigsuspend(&mask);
|
|
|
- } while (THREAD_GETMEM(self, p_signal) !=__pthread_sig_restart);
|
|
|
-
|
|
|
- READ_MEMORY_BARRIER();
|
|
|
-}
|
|
|
-
|
|
|
-#ifndef __NR_rt_sigaction
|
|
|
-
|
|
|
- signals.
|
|
|
- On these kernels, we use SIGUSR1 and SIGUSR2 for restart and cancellation.
|
|
|
- Since the restart signal does not queue, we use an atomic counter to create
|
|
|
- queuing semantics. This is needed to resolve a rare race condition in
|
|
|
- pthread_cond_timedwait_relative. */
|
|
|
-
|
|
|
-void __pthread_restart_old(pthread_descr th)
|
|
|
-{
|
|
|
- if (atomic_increment(&th->p_resume_count) == -1)
|
|
|
- kill(th->p_pid, __pthread_sig_restart);
|
|
|
-}
|
|
|
-
|
|
|
-void __pthread_suspend_old(pthread_descr self)
|
|
|
-{
|
|
|
- if (atomic_decrement(&self->p_resume_count) <= 0)
|
|
|
- __pthread_wait_for_restart_signal(self);
|
|
|
-}
|
|
|
-
|
|
|
-int
|
|
|
-__pthread_timedsuspend_old(pthread_descr self, const struct timespec *abstime)
|
|
|
-{
|
|
|
- sigset_t unblock, initial_mask;
|
|
|
- int was_signalled = 0;
|
|
|
- sigjmp_buf jmpbuf;
|
|
|
-
|
|
|
- if (atomic_decrement(&self->p_resume_count) == 0) {
|
|
|
-
|
|
|
- the signal and sleep. */
|
|
|
-
|
|
|
- if (sigsetjmp(jmpbuf, 1) == 0) {
|
|
|
- THREAD_SETMEM(self, p_signal_jmp, &jmpbuf);
|
|
|
- THREAD_SETMEM(self, p_signal, 0);
|
|
|
-
|
|
|
- __sigemptyset(&unblock);
|
|
|
- sigaddset(&unblock, __pthread_sig_restart);
|
|
|
- sigprocmask(SIG_UNBLOCK, &unblock, &initial_mask);
|
|
|
-
|
|
|
- while (1) {
|
|
|
- struct timeval now;
|
|
|
- struct timespec reltime;
|
|
|
-
|
|
|
-
|
|
|
- gettimeofday (&now, NULL);
|
|
|
- reltime.tv_nsec = abstime->tv_nsec - now.tv_usec * 1000;
|
|
|
- reltime.tv_sec = abstime->tv_sec - now.tv_sec;
|
|
|
- if (reltime.tv_nsec < 0) {
|
|
|
- reltime.tv_nsec += 1000000000;
|
|
|
- reltime.tv_sec -= 1;
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
- resume waiting as required by Single Unix Specification. */
|
|
|
- if (reltime.tv_sec < 0 || nanosleep(&reltime, NULL) == 0)
|
|
|
- break;
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
- sigprocmask(SIG_SETMASK, &initial_mask, NULL);
|
|
|
- was_signalled = 0;
|
|
|
- } else {
|
|
|
- was_signalled = 1;
|
|
|
- }
|
|
|
- THREAD_SETMEM(self, p_signal_jmp, NULL);
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
- due to the delivery of a restart signal. In that case,
|
|
|
- we know we have been dequeued and resumed and that the
|
|
|
- resume count is balanced. Otherwise, there are some
|
|
|
- cases to consider. First, try to bump up the resume count
|
|
|
- back to zero. If it goes to 1, it means restart() was
|
|
|
- invoked on this thread. The signal must be consumed
|
|
|
- and the count bumped down and everything is cool. We
|
|
|
- can return a 1 to the caller.
|
|
|
- Otherwise, no restart was delivered yet, so a potential
|
|
|
- race exists; we return a 0 to the caller which must deal
|
|
|
- with this race in an appropriate way; for example by
|
|
|
- atomically removing the thread from consideration for a
|
|
|
- wakeup---if such a thing fails, it means a restart is
|
|
|
- being delivered. */
|
|
|
-
|
|
|
- if (!was_signalled) {
|
|
|
- if (atomic_increment(&self->p_resume_count) != -1) {
|
|
|
- __pthread_wait_for_restart_signal(self);
|
|
|
- atomic_decrement(&self->p_resume_count);
|
|
|
-
|
|
|
- return 1;
|
|
|
- }
|
|
|
-
|
|
|
- return 0;
|
|
|
- }
|
|
|
-
|
|
|
- return 1;
|
|
|
-}
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
-#ifdef __NR_rt_sigaction
|
|
|
-void __pthread_restart_new(pthread_descr th)
|
|
|
-{
|
|
|
-
|
|
|
- our assumptions. The intent is to commit previous writes to shared
|
|
|
- memory so the woken thread will have a consistent view. Complementary
|
|
|
- read barriers are present to the suspend functions. */
|
|
|
- WRITE_MEMORY_BARRIER();
|
|
|
- kill(th->p_pid, __pthread_sig_restart);
|
|
|
-}
|
|
|
-
|
|
|
-int __pthread_timedsuspend_new(pthread_descr self, const struct timespec *abstime)
|
|
|
-{
|
|
|
- sigset_t unblock, initial_mask;
|
|
|
- int was_signalled = 0;
|
|
|
- sigjmp_buf jmpbuf;
|
|
|
-
|
|
|
- if (sigsetjmp(jmpbuf, 1) == 0) {
|
|
|
- THREAD_SETMEM(self, p_signal_jmp, &jmpbuf);
|
|
|
- THREAD_SETMEM(self, p_signal, 0);
|
|
|
-
|
|
|
- __sigemptyset(&unblock);
|
|
|
- sigaddset(&unblock, __pthread_sig_restart);
|
|
|
- sigprocmask(SIG_UNBLOCK, &unblock, &initial_mask);
|
|
|
-
|
|
|
- while (1) {
|
|
|
- struct timeval now;
|
|
|
- struct timespec reltime;
|
|
|
-
|
|
|
-
|
|
|
- gettimeofday (&now, NULL);
|
|
|
- reltime.tv_nsec = abstime->tv_nsec - now.tv_usec * 1000;
|
|
|
- reltime.tv_sec = abstime->tv_sec - now.tv_sec;
|
|
|
- if (reltime.tv_nsec < 0) {
|
|
|
- reltime.tv_nsec += 1000000000;
|
|
|
- reltime.tv_sec -= 1;
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
- resume waiting as required by Single Unix Specification. */
|
|
|
- if (reltime.tv_sec < 0 || nanosleep(&reltime, NULL) == 0)
|
|
|
- break;
|
|
|
- }
|
|
|
-
|
|
|
-
|
|
|
- sigprocmask(SIG_SETMASK, &initial_mask, NULL);
|
|
|
- was_signalled = 0;
|
|
|
- } else {
|
|
|
- was_signalled = 1;
|
|
|
- }
|
|
|
- THREAD_SETMEM(self, p_signal_jmp, NULL);
|
|
|
-
|
|
|
-
|
|
|
- due to the delivery of a restart signal. In that case,
|
|
|
- everything is cool. We have been removed from whatever
|
|
|
- we were waiting on by the other thread, and consumed its signal.
|
|
|
-
|
|
|
- Otherwise we this thread woke up spontaneously, or due to a signal other
|
|
|
- than restart. This is an ambiguous case that must be resolved by
|
|
|
- the caller; the thread is still eligible for a restart wakeup
|
|
|
- so there is a race. */
|
|
|
-
|
|
|
- READ_MEMORY_BARRIER();
|
|
|
- return was_signalled;
|
|
|
-}
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
-
|
|
|
-#ifdef DEBUG_PT
|
|
|
-#include <stdarg.h>
|
|
|
-
|
|
|
-void __pthread_message(char * fmt, ...)
|
|
|
-{
|
|
|
- char buffer[1024];
|
|
|
- va_list args;
|
|
|
- sprintf(buffer, "%05d : ", getpid());
|
|
|
- va_start(args, fmt);
|
|
|
- vsnprintf(buffer + 8, sizeof(buffer) - 8, fmt, args);
|
|
|
- va_end(args);
|
|
|
- TEMP_FAILURE_RETRY(write(2, buffer, strlen(buffer)));
|
|
|
-}
|
|
|
-
|
|
|
-#endif
|
|
|
-
|
|
|
-
|
|
|
-#ifndef __PIC__
|
|
|
-
|
|
|
- static libpthread is used. */
|
|
|
-extern const char __pthread_provide_wrappers;
|
|
|
-static const char *const __pthread_require_wrappers =
|
|
|
- &__pthread_provide_wrappers;
|
|
|
-#endif
|