X-Git-Url: http://nsz.repo.hu/git/?a=blobdiff_plain;ds=sidebyside;f=src%2Fthread%2Fpthread_key_create.c;h=e26f199c3c9cce5b7050bf576501533060d9e06b;hb=11ce1b133d594b6a454d3e8d5941e7a6a432c42b;hp=a78e507a6fc9a9708140e96929a6a572a4e27601;hpb=19a1fe670acb3ab9ead0fe31859ca7d4fe40dd54;p=musl diff --git a/src/thread/pthread_key_create.c b/src/thread/pthread_key_create.c index a78e507a..e26f199c 100644 --- a/src/thread/pthread_key_create.c +++ b/src/thread/pthread_key_create.c @@ -3,54 +3,123 @@ volatile size_t __pthread_tsd_size = sizeof(void *) * PTHREAD_KEYS_MAX; void *__pthread_tsd_main[PTHREAD_KEYS_MAX] = { 0 }; -static void (*volatile keys[PTHREAD_KEYS_MAX])(void *); +static void (*keys[PTHREAD_KEYS_MAX])(void *); + +static pthread_rwlock_t key_lock = PTHREAD_RWLOCK_INITIALIZER; + +static pthread_key_t next_key; static void nodtor(void *dummy) { } +static void dirty(void *dummy) +{ +} + +struct cleanup_args { + pthread_t caller; + int ret; +}; + +static void clean_dirty_tsd_callback(void *p) +{ + struct cleanup_args *args = p; + pthread_t self = __pthread_self(); + pthread_key_t i; + for (i=0; itsd[i]) + self->tsd[i] = 0; + } + /* Arbitrary choice to avoid data race. */ + if (args->caller == self) args->ret = 0; +} + +static int clean_dirty_tsd(void) +{ + struct cleanup_args args = { + .caller = __pthread_self(), + .ret = EAGAIN + }; + __pthread_key_delete_synccall(clean_dirty_tsd_callback, &args); + return args.ret; +} + int __pthread_key_create(pthread_key_t *k, void (*dtor)(void *)) { - unsigned i = (uintptr_t)&k / 16 % PTHREAD_KEYS_MAX; - unsigned j = i; + pthread_key_t j = next_key; pthread_t self = __pthread_self(); + int found_dirty = 0; /* This can only happen in the main thread before * pthread_create has been called. */ if (!self->tsd) self->tsd = __pthread_tsd_main; + /* Purely a sentinel value since null means slot is free. */ if (!dtor) dtor = nodtor; + + pthread_rwlock_wrlock(&key_lock); do { - if (!a_cas_p(keys+j, 0, (void *)dtor)) { - *k = j; + if (!keys[j]) { + keys[next_key = *k = j] = dtor; + pthread_rwlock_unlock(&key_lock); return 0; + } else if (keys[j] == dirty) { + found_dirty = 1; } - } while ((j=(j+1)%PTHREAD_KEYS_MAX) != i); - return EAGAIN; + } while ((j=(j+1)%PTHREAD_KEYS_MAX) != next_key); + + /* It's possible that all slots are in use or __synccall fails. */ + if (!found_dirty || clean_dirty_tsd()) { + pthread_rwlock_unlock(&key_lock); + return EAGAIN; + } + + /* If this point is reached there is necessarily a newly-cleaned + * slot to allocate to satisfy the caller's request. Find it and + * mark any additional previously-dirty slots clean. */ + for (j=0; jtsd_used; - for (j=0; not_finished && jtsd_used && jtsd_used = 0; for (i=0; itsd[i] && keys[i]) { - void *tmp = self->tsd[i]; - self->tsd[i] = 0; - keys[i](tmp); - not_finished = 1; + void *val = self->tsd[i]; + void (*dtor)(void *) = keys[i]; + self->tsd[i] = 0; + if (val && dtor && dtor != nodtor && dtor != dirty) { + pthread_rwlock_unlock(&key_lock); + dtor(val); + pthread_rwlock_rdlock(&key_lock); } } + pthread_rwlock_unlock(&key_lock); } } -weak_alias(__pthread_key_delete, pthread_key_delete); weak_alias(__pthread_key_create, pthread_key_create);