X-Git-Url: http://nsz.repo.hu/git/?p=musl;a=blobdiff_plain;f=src%2Fthread%2Fpthread_create.c;h=5d5d3a6f939a84548faab354f970556109bec107;hp=4c1decaa7ce5f660ace954a032fcd88751e603d6;hb=d0ba09837bc033e236e921304b213daea0fe2653;hpb=d5142642b8e6c45449158efdb8f8e87af4dafde8 diff --git a/src/thread/pthread_create.c b/src/thread/pthread_create.c index 4c1decaa..5d5d3a6f 100644 --- a/src/thread/pthread_create.c +++ b/src/thread/pthread_create.c @@ -12,7 +12,7 @@ weak_alias(dummy_0, __pthread_tsd_run_dtors); _Noreturn void pthread_exit(void *result) { pthread_t self = pthread_self(); - int n; + sigset_t set; self->result = result; @@ -32,15 +32,35 @@ _Noreturn void pthread_exit(void *result) self->dead = 1; __unlock(self->killlock); - do n = libc.threads_minus_1; - while (n && a_cas(&libc.threads_minus_1, n, n-1)!=n); - if (!n) exit(0); + /* Block all signals before decrementing the live thread count. + * This is important to ensure that dynamically allocated TLS + * is not under-allocated/over-committed, and possibly for other + * reasons as well. */ + __syscall(SYS_rt_sigprocmask, SIG_BLOCK, SIGALL_SET, &set, _NSIG/8); + + /* It's impossible to determine whether this is "the last thread" + * until performing the atomic decrement, since multiple threads + * could exit at the same time. For the last thread, revert the + * decrement and unblock signals to give the atexit handlers and + * stdio cleanup code a consistent state. */ + if (a_fetch_add(&libc.threads_minus_1, -1)==0) { + libc.threads_minus_1 = 0; + __syscall(SYS_rt_sigprocmask, SIG_SETMASK, &set, 0, _NSIG/8); + exit(0); + } if (self->detached && self->map_base) { - if (self->detached == 2) - __syscall(SYS_set_tid_address, 0); - __syscall(SYS_rt_sigprocmask, SIG_BLOCK, - SIGALL_SET, 0, __SYSCALL_SSLEN); + /* Detached threads must avoid the kernel clear_child_tid + * feature, since the virtual address will have been + * unmapped and possibly already reused by a new mapping + * at the time the kernel would perform the write. In + * the case of threads that started out detached, the + * initial clone flags are correct, but if the thread was + * detached later (== 2), we need to clear it here. */ + if (self->detached == 2) __syscall(SYS_set_tid_address, 0); + + /* The following call unmaps the thread's stack mapping + * and then exits without touching the stack. */ __unmapself(self->map_base, self->map_size); } @@ -69,11 +89,11 @@ static int start(void *p) pthread_exit(0); } __syscall(SYS_rt_sigprocmask, SIG_SETMASK, - self->sigmask, 0, __SYSCALL_SSLEN); + self->sigmask, 0, _NSIG/8); } if (self->unblock_cancel) __syscall(SYS_rt_sigprocmask, SIG_UNBLOCK, - SIGPT_SET, 0, __SYSCALL_SSLEN); + SIGPT_SET, 0, _NSIG/8); pthread_exit(self->start(self->start_arg)); return 0; } @@ -101,7 +121,7 @@ int pthread_create(pthread_t *restrict res, const pthread_attr_t *restrict attrp int ret; size_t size, guard; struct pthread *self = pthread_self(), *new; - unsigned char *map = 0, *stack = 0, *tsd = 0; + unsigned char *map = 0, *stack = 0, *tsd = 0, *stack_limit; unsigned flags = 0x7d8f00; int do_sched = 0; pthread_attr_t attr = {0}; @@ -123,6 +143,7 @@ int pthread_create(pthread_t *restrict res, const pthread_attr_t *restrict attrp size_t need = libc.tls_size + __pthread_tsd_size; size = attr._a_stacksize + DEFAULT_STACK_SIZE; stack = (void *)(attr._a_stackaddr & -16); + stack_limit = (void *)(attr._a_stackaddr - size); /* Use application-provided stack for TLS only when * it does not take more than ~12% or 2k of the * application's stack space. */ @@ -142,22 +163,27 @@ int pthread_create(pthread_t *restrict res, const pthread_attr_t *restrict attrp if (!tsd) { if (guard) { map = mmap(0, size, PROT_NONE, MAP_PRIVATE|MAP_ANON, -1, 0); - if (map == MAP_FAILED) return EAGAIN; + if (map == MAP_FAILED) goto fail; if (mprotect(map+guard, size-guard, PROT_READ|PROT_WRITE)) { munmap(map, size); - return EAGAIN; + goto fail; } } else { map = mmap(0, size, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANON, -1, 0); - if (map == MAP_FAILED) return EAGAIN; + if (map == MAP_FAILED) goto fail; } tsd = map + size - __pthread_tsd_size; - if (!stack) stack = tsd - libc.tls_size; + if (!stack) { + stack = tsd - libc.tls_size; + stack_limit = map + guard; + } } new = __copy_tls(tsd - libc.tls_size); new->map_base = map; new->map_size = size; + new->stack = stack; + new->stack_size = stack - stack_limit; new->pid = self->pid; new->errno_ptr = &new->errno_val; new->start = entry; @@ -171,7 +197,7 @@ int pthread_create(pthread_t *restrict res, const pthread_attr_t *restrict attrp if (attr._a_sched) { do_sched = new->startlock[0] = 1; __syscall(SYS_rt_sigprocmask, SIG_BLOCK, - SIGALL_SET, self->sigmask, __SYSCALL_SSLEN); + SIGALL_SET, self->sigmask, _NSIG/8); } new->unblock_cancel = self->cancel; new->canary = self->canary; @@ -183,12 +209,12 @@ int pthread_create(pthread_t *restrict res, const pthread_attr_t *restrict attrp if (do_sched) { __syscall(SYS_rt_sigprocmask, SIG_SETMASK, - new->sigmask, 0, __SYSCALL_SSLEN); + new->sigmask, 0, _NSIG/8); } if (ret < 0) { a_dec(&libc.threads_minus_1); - munmap(map, size); + if (map) munmap(map, size); return EAGAIN; } @@ -202,4 +228,7 @@ int pthread_create(pthread_t *restrict res, const pthread_attr_t *restrict attrp *res = new; return 0; +fail: + __release_ptc(); + return EAGAIN; }