#include "pthread_impl.h"
+#ifdef __pthread_unwind_next
+#undef __pthread_unwind_next
+#define __pthread_unwind_next __pthread_unwind_next_3
+#endif
+
void __pthread_unwind_next(struct __ptcb *cb)
{
int i, j, not_finished;
exit(0);
if (self->detached && self->map_base) {
- syscall4(__NR_rt_sigprocmask, SIG_BLOCK, (long)(uint64_t[1]){-1},0,8);
+ syscall(__NR_rt_sigprocmask, SIG_BLOCK, (long)(uint64_t[1]){-1},0,8);
__unmapself(self->map_base, self->map_size);
}
- __syscall_exit(0);
+ syscall(SYS_exit, 0);
}
static void docancel(struct pthread *self)
{
struct __ptcb cb = { .__next = self->cancelbuf };
+ sigset_t set;
+ self->canceldisable = 1;
+ self->cancelasync = 0;
+ sigemptyset(&set);
+ sigaddset(&set, SIGCANCEL);
+ __libc_sigprocmask(SIG_UNBLOCK, &set, 0);
__pthread_unwind_next(&cb);
}
struct pthread *self = __pthread_self();
if (si->si_code > 0 || si->si_pid != self->pid) return;
self->cancel = 1;
- if (self->canceldisable || (!self->cancelasync && !self->cancelpoint))
- return;
- docancel(self);
+ if (self->canceldisable) return;
+ if (self->cancelasync || (self->cancelpoint==1 && PC_AT_SYS(ctx)))
+ docancel(self);
}
static void cancelpt(int x)
{
struct pthread *self = __pthread_self();
if (self->canceldisable) return;
- self->cancelpoint = x;
- if (self->cancel) docancel(self);
+ if ((self->cancelpoint+=x)==1 && x>=0 && self->cancel)
+ docancel(self);
}
/* "rsyscall" is a mechanism by which a thread can synchronously force all
return;
}
- if (syscall6(rs.nr, rs.arg[0], rs.arg[1], rs.arg[2],
+ if (syscall(rs.nr, rs.arg[0], rs.arg[1], rs.arg[2],
rs.arg[3], rs.arg[4], rs.arg[5]) < 0 && !rs.err) rs.err=errno;
a_inc(&rs.cnt);
while((i=rs.cnt)) __wait(&rs.cnt, 0, i, 1);
if (rs.err) errno = rs.err, ret = -1;
- else ret = syscall6(nr, a, b, c, d, e, f);
+ else ret = syscall(nr, a, b, c, d, e, f);
UNLOCK(&rs.lock);
return ret;
{
struct sigaction sa = { .sa_flags = SA_SIGINFO | SA_RESTART };
libc.lock = __lock;
+ libc.lockfile = __lockfile;
libc.cancelpt = cancelpt;
libc.rsyscall = rsyscall;
sa.sa_sigaction = cancel_handler;
size += __pthread_tsd_size;
map = mmap(0, size, PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE|MAP_ANON, -1, 0);
if (!map) return EAGAIN;
- mprotect(map, guard, PROT_NONE);
+ if (guard) mprotect(map, guard, PROT_NONE);
tsd = map + size - __pthread_tsd_size;
new = (void *)(tsd - sizeof *new - PAGE_SIZE%sizeof *new);