diff options
author | Brian Behlendorf <[email protected]> | 2009-07-14 14:16:26 -0700 |
---|---|---|
committer | Brian Behlendorf <[email protected]> | 2009-07-14 14:16:26 -0700 |
commit | 864e44e7e03e7ade1810617f55d5f2fb14fc1e94 (patch) | |
tree | aa15daa312bafa656563358336669fde198f4adb | |
parent | 57ae3b59e9c32f8dcfe1b9a790b8d06c099208c9 (diff) |
Provide correct kthread API semantics using pthreads.
The intent here is to fully remove the previous Solaris thread
implementation so we don't need to simulate both Solaris kernel
and user space thread APIs. The few user space consumers of the
thread API have been updated to use the kthread API. In order
to support this we needed to more fully support the kthread API
and that means not doing crazy things like casting a thread id
to a pointer and using that as was done before. This first
implementation is not effecient but it does provide all the
corrent semantics. If/when performance becomes and issue we
can and should just natively adopt pthreads which is portable.
Let me finish by saying I'm not proud of any of this and I would
love to see it improved. However, this slow implementation does
at least provide all the correct kthread API semantics whereas
the previous method of casting the thread ID to a pointer was
dodgy at best.
-rw-r--r-- | cmd/ztest/ztest.c | 19 | ||||
-rw-r--r-- | lib/libzpool/include/sys/zfs_context.h | 37 | ||||
-rw-r--r-- | lib/libzpool/kernel.c | 247 | ||||
-rw-r--r-- | lib/libzpool/taskq.c | 20 | ||||
-rw-r--r-- | module/zfs/txg.c | 4 |
5 files changed, 228 insertions, 99 deletions
diff --git a/cmd/ztest/ztest.c b/cmd/ztest/ztest.c index facb2354a..e171e91de 100644 --- a/cmd/ztest/ztest.c +++ b/cmd/ztest/ztest.c @@ -141,6 +141,7 @@ typedef struct ztest_args { objset_t *za_os; zilog_t *za_zilog; kthread_t *za_thread; + kt_did_t za_threadid; uint64_t za_instance; uint64_t za_random; uint64_t za_diroff; @@ -3806,6 +3807,8 @@ ztest_resume_thread(void *arg) (void) poll(NULL, 0, 1000); ztest_resume(spa); } + + thread_exit(); return (NULL); } @@ -3870,6 +3873,7 @@ ztest_thread(void *arg) break; } + thread_exit(); return (NULL); } @@ -3885,6 +3889,7 @@ ztest_run(char *pool) spa_t *spa; char name[100]; kthread_t *resume_thread; + kt_did_t resume_id; ztest_exiting = B_FALSE; @@ -3963,8 +3968,9 @@ ztest_run(char *pool) /* * Create a thread to periodically resume suspended I/O. */ - resume_thread = thread_create(NULL, 0, ztest_resume_thread, spa, - THR_BOUND, NULL, 0, 0); + VERIFY3P((resume_thread = thread_create(NULL, 0, ztest_resume_thread, + spa, THR_BOUND, NULL, 0, 0)), !=, NULL); + resume_id = resume_thread->t_tid; /* * Verify that we can safely inquire about about any object, @@ -4040,12 +4046,13 @@ ztest_run(char *pool) za[d].za_zilog = zil_open(za[d].za_os, NULL); } - za[t].za_thread = thread_create(NULL, 0, ztest_thread, &za[t], - THR_BOUND, NULL, 0, 0); + VERIFY3P((za[t].za_thread = thread_create(NULL, 0, ztest_thread, + &za[t], THR_BOUND, NULL, 0, 0)), !=, NULL); + za[t].za_threadid = za[t].za_thread->t_tid; } while (--t >= 0) { - VERIFY(thr_join(za[t].za_thread, NULL, NULL) == 0); + VERIFY(thread_join(za[t].za_threadid, NULL, NULL) == 0); if (t < zopt_datasets) { zil_close(za[t].za_zilog); dmu_objset_close(za[t].za_os); @@ -4084,7 +4091,7 @@ ztest_run(char *pool) /* Kill the resume thread */ ztest_exiting = B_TRUE; - VERIFY(thr_join(resume_thread, NULL, NULL) == 0); + VERIFY(thread_join(resume_id, NULL, NULL) == 0); ztest_resume(spa); /* diff --git a/lib/libzpool/include/sys/zfs_context.h b/lib/libzpool/include/sys/zfs_context.h index 15f7665cd..cad7553cc 100644 --- a/lib/libzpool/include/sys/zfs_context.h +++ b/lib/libzpool/include/sys/zfs_context.h @@ -192,24 +192,34 @@ _NOTE(CONSTCOND) } while (0) /* * Threads */ -#define tsd_get(key) pthread_getspecific(key) -#define tsd_set(key, val) pthread_setspecific(key, val) +#define THR_BOUND 0x00000001 +#define TS_RUN 0x00000002 typedef void (*thread_func_t)(void *); +typedef pthread_t kt_did_t; + typedef struct kthread { list_node_t t_node; - pthread_t t_id; + kt_did_t t_tid; pthread_attr_t t_attr; } kthread_t; -#define curthread zk_curthread() -#define thread_create(stk, stksize, func, arg, len, pp, state, pri) \ - zk_thread_create((thread_func_t)func, arg) -#define thr_join(kt, v1, v2) pthread_join(kt->t_id, v2) - -extern kthread_t *zk_curthread(void); -extern kthread_t *zk_thread_create(thread_func_t func, void *arg); -extern void thread_exit(void); +#define tsd_get(key) pthread_getspecific(key) +#define tsd_set(key, val) pthread_setspecific(key, val) +#define curthread zk_thread_current() +#define thread_exit zk_thread_exit +#define thread_create(stk, stksize, func, arg, len, pp, state, pri) \ + zk_thread_create(stk, stksize, (thread_func_t)func, arg, \ + len, NULL, state, pri) +#define thread_join(tid, dtid, status) \ + zk_thread_join(tid, dtid, status) + +extern kthread_t *zk_thread_current(void); +extern void zk_thread_exit(void); +extern kthread_t *zk_thread_create(caddr_t stk, size_t stksize, + thread_func_t func, void *arg, size_t len, + void *pp, int state, pri_t pri); +extern int zk_thread_join(kt_did_t tid, kthread_t *dtid, void **status); #define issig(why) (FALSE) #define ISSIG(thr, why) (FALSE) @@ -218,6 +228,8 @@ extern void thread_exit(void); * Mutexes */ #define MTX_MAGIC 0x9522f51362a6e326ull +#define MTX_INIT (void *)NULL +#define MTX_DEST (void *)-1UL typedef struct kmutex { void *m_owner; uint64_t m_magic; @@ -238,6 +250,8 @@ extern void *mutex_owner(kmutex_t *mp); * RW locks */ #define RW_MAGIC 0x4d31fb123648e78aull +#define RW_INIT (void *)NULL +#define RW_DEST (void *)-1UL typedef struct krwlock { void *rw_owner; void *rw_wr_owner; @@ -339,6 +353,7 @@ extern void taskq_destroy(taskq_t *); extern void taskq_wait(taskq_t *); extern int taskq_member(taskq_t *, void *); extern void system_taskq_init(void); +extern void system_taskq_fini(void); #define XVA_MAPSIZE 3 #define XVA_MAGIC 0x78766174 diff --git a/lib/libzpool/kernel.c b/lib/libzpool/kernel.c index 161f3d5ba..19776bb8a 100644 --- a/lib/libzpool/kernel.c +++ b/lib/libzpool/kernel.c @@ -30,6 +30,7 @@ #include <stdlib.h> #include <string.h> #include <zlib.h> +#include <sys/signal.h> #include <sys/spa.h> #include <sys/stat.h> #include <sys/processor.h> @@ -57,64 +58,157 @@ struct utsname utsname = { * ========================================================================= */ -kmutex_t kthread_lock; +/* NOTE: Tracking each tid on a list and using it for curthread lookups + * is slow at best but it provides an easy way to provide a kthread + * style API on top of pthreads. For now we just want ztest to work + * to validate correctness. Performance is not much of an issue + * since that is what the in-kernel version is for. That said + * reworking this to track the kthread_t structure as thread + * specific data would be probably the best way to speed this up. + */ + +pthread_cond_t kthread_cond = PTHREAD_COND_INITIALIZER; +pthread_mutex_t kthread_lock = PTHREAD_MUTEX_INITIALIZER; list_t kthread_list; -kthread_t * -zk_curthread(void) +static int +thread_count(void) { kthread_t *kt; - pthread_t tid; + int count = 0; + + for (kt = list_head(&kthread_list); kt != NULL; + kt = list_next(&kthread_list, kt)) + count++; + + return count; +} + +static void +thread_init(void) +{ + kthread_t *kt; + + /* Initialize list for tracking kthreads */ + list_create(&kthread_list, sizeof (kthread_t), + offsetof(kthread_t, t_node)); - tid = pthread_self(); - mutex_enter(&kthread_lock); - for (kt = list_head(&kthread_list); kt != NULL; - kt = list_next(&kthread_list, kt)) { + /* Create entry for primary kthread */ + kt = umem_zalloc(sizeof(kthread_t), UMEM_NOFAIL); + list_link_init(&kt->t_node); + VERIFY3U(kt->t_tid = pthread_self(), !=, 0); + VERIFY3S(pthread_attr_init(&kt->t_attr), ==, 0); + VERIFY3S(pthread_mutex_lock(&kthread_lock), ==, 0); + list_insert_head(&kthread_list, kt); + VERIFY3S(pthread_mutex_unlock(&kthread_lock), ==, 0); +} + +static void +thread_fini(void) +{ + kthread_t *kt; + struct timespec ts = { 0 }; + int count; + + /* Wait for all threads to exit via thread_exit() */ + VERIFY3S(pthread_mutex_lock(&kthread_lock), ==, 0); + while ((count = thread_count()) > 1) { + printf("Waiting for %d\n", count); + clock_gettime(CLOCK_REALTIME, &ts); + ts.tv_sec += 1; + pthread_cond_timedwait(&kthread_cond, &kthread_lock, &ts); + } + + ASSERT3S(thread_count(), ==, 1); + kt = list_head(&kthread_list); + list_remove(&kthread_list, kt); + VERIFY3S(pthread_mutex_unlock(&kthread_lock), ==, 0); + + VERIFY(pthread_attr_destroy(&kt->t_attr) == 0); + umem_free(kt, sizeof(kthread_t)); - if (kt->t_id == tid) { - mutex_exit(&kthread_lock); - return kt; + /* Cleanup list for tracking kthreads */ + list_destroy(&kthread_list); +} + +kthread_t * +zk_thread_current(void) +{ + kt_did_t tid = pthread_self(); + kthread_t *kt; + int count = 1; + + /* + * Because a newly created thread may call zk_thread_current() + * before the thread parent has had time to add the thread's tid + * to our lookup list. We will loop as long as there are tid + * which have not yet been set which must be one of ours. + * Yes it's a hack, at some point we can just use native pthreads. + */ + while (count > 0) { + count = 0; + VERIFY3S(pthread_mutex_lock(&kthread_lock), ==, 0); + for (kt = list_head(&kthread_list); kt != NULL; + kt = list_next(&kthread_list, kt)) { + + if (kt->t_tid == tid) { + VERIFY3S(pthread_mutex_unlock( + &kthread_lock), ==, 0); + return kt; + } + + if (kt->t_tid == (kt_did_t)-1) + count++; } + VERIFY3S(pthread_mutex_unlock(&kthread_lock), ==, 0); } - mutex_exit(&kthread_lock); + /* Unreachable */ + ASSERT(0); return NULL; } kthread_t * -zk_thread_create(thread_func_t func, void *arg) +zk_thread_create(caddr_t stk, size_t stksize, thread_func_t func, void *arg, + size_t len, void *pp, int state, pri_t pri) { kthread_t *kt; kt = umem_zalloc(sizeof(kthread_t), UMEM_NOFAIL); - + kt->t_tid = (kt_did_t)-1; + list_link_init(&kt->t_node); VERIFY(pthread_attr_init(&kt->t_attr) == 0); - VERIFY(pthread_attr_setdetachstate(&kt->t_attr, - PTHREAD_CREATE_DETACHED) == 0); - VERIFY(pthread_create(&kt->t_id, &kt->t_attr, - (void *(*)(void *))func, arg) == 0); - mutex_enter(&kthread_lock); + VERIFY3S(pthread_mutex_lock(&kthread_lock), ==, 0); list_insert_head(&kthread_list, kt); - mutex_exit(&kthread_lock); + VERIFY3S(pthread_mutex_unlock(&kthread_lock), ==, 0); + + VERIFY(pthread_create(&kt->t_tid, &kt->t_attr, + (void *(*)(void *))func, arg) == 0); return kt; } +int +zk_thread_join(kt_did_t tid, kthread_t *dtid, void **status) +{ + return pthread_join(tid, status); +} + void -thread_exit(void) +zk_thread_exit(void) { kthread_t *kt; - VERIFY((kt = curthread) != NULL); - - mutex_enter(&kthread_lock); + VERIFY3P(kt = curthread, !=, NULL); + VERIFY3S(pthread_mutex_lock(&kthread_lock), ==, 0); list_remove(&kthread_list, kt); - mutex_exit(&kthread_lock); + VERIFY3S(pthread_mutex_unlock(&kthread_lock), ==, 0); VERIFY(pthread_attr_destroy(&kt->t_attr) == 0); umem_free(kt, sizeof(kthread_t)); + pthread_cond_broadcast(&kthread_cond); pthread_exit(NULL); } @@ -149,14 +243,14 @@ kstat_delete(kstat_t *ksp) void mutex_init(kmutex_t *mp, char *name, int type, void *cookie) { - ASSERT(type == MUTEX_DEFAULT); - ASSERT(cookie == NULL); + ASSERT3S(type, ==, MUTEX_DEFAULT); + ASSERT3P(cookie, ==, NULL); #ifdef IM_FEELING_LUCKY - ASSERT(mp->m_magic != MTX_MAGIC); + ASSERT3U(mp->m_magic, !=, MTX_MAGIC); #endif - mp->m_owner = NULL; + mp->m_owner = MTX_INIT; mp->m_magic = MTX_MAGIC; VERIFY3S(pthread_mutex_init(&mp->m_lock, NULL), ==, 0); } @@ -164,31 +258,31 @@ mutex_init(kmutex_t *mp, char *name, int type, void *cookie) void mutex_destroy(kmutex_t *mp) { - ASSERT(mp->m_magic == MTX_MAGIC); - ASSERT(mp->m_owner == NULL); + ASSERT3U(mp->m_magic, ==, MTX_MAGIC); + ASSERT3P(mp->m_owner, ==, MTX_INIT); VERIFY3S(pthread_mutex_destroy(&(mp)->m_lock), ==, 0); - mp->m_owner = (void *)-1UL; + mp->m_owner = MTX_DEST; mp->m_magic = 0; } void mutex_enter(kmutex_t *mp) { - ASSERT(mp->m_magic == MTX_MAGIC); - ASSERT(mp->m_owner != (void *)-1UL); - ASSERT(mp->m_owner != curthread); + ASSERT3U(mp->m_magic, ==, MTX_MAGIC); + ASSERT3P(mp->m_owner, !=, MTX_DEST); + ASSERT3P(mp->m_owner, !=, curthread); VERIFY3S(pthread_mutex_lock(&mp->m_lock), ==, 0); - ASSERT(mp->m_owner == NULL); + ASSERT3P(mp->m_owner, ==, MTX_INIT); mp->m_owner = curthread; } int mutex_tryenter(kmutex_t *mp) { - ASSERT(mp->m_magic == MTX_MAGIC); - ASSERT(mp->m_owner != (void *)-1UL); + ASSERT3U(mp->m_magic, ==, MTX_MAGIC); + ASSERT3P(mp->m_owner, !=, MTX_DEST); if (0 == pthread_mutex_trylock(&mp->m_lock)) { - ASSERT(mp->m_owner == NULL); + ASSERT3P(mp->m_owner, ==, MTX_INIT); mp->m_owner = curthread; return (1); } else { @@ -199,16 +293,16 @@ mutex_tryenter(kmutex_t *mp) void mutex_exit(kmutex_t *mp) { - ASSERT(mp->m_magic == MTX_MAGIC); - ASSERT(mutex_owner(mp) == curthread); - mp->m_owner = NULL; + ASSERT3U(mp->m_magic, ==, MTX_MAGIC); + ASSERT3P(mutex_owner(mp), ==, curthread); + mp->m_owner = MTX_INIT; VERIFY3S(pthread_mutex_unlock(&mp->m_lock), ==, 0); } void * mutex_owner(kmutex_t *mp) { - ASSERT(mp->m_magic == MTX_MAGIC); + ASSERT3U(mp->m_magic, ==, MTX_MAGIC); return (mp->m_owner); } @@ -221,16 +315,16 @@ mutex_owner(kmutex_t *mp) void rw_init(krwlock_t *rwlp, char *name, int type, void *arg) { - ASSERT(type == RW_DEFAULT); - ASSERT(arg == NULL); + ASSERT3S(type, ==, RW_DEFAULT); + ASSERT3P(arg, ==, NULL); #ifdef IM_FEELING_LUCKY - ASSERT(rwlp->rw_magic != RW_MAGIC); + ASSERT3U(rwlp->rw_magic, !=, RW_MAGIC); #endif VERIFY3S(pthread_rwlock_init(&rwlp->rw_lock, NULL), ==, 0); - rwlp->rw_owner = NULL; - rwlp->rw_wr_owner = NULL; + rwlp->rw_owner = RW_INIT; + rwlp->rw_wr_owner = RW_INIT; rwlp->rw_readers = 0; rwlp->rw_magic = RW_MAGIC; } @@ -238,7 +332,7 @@ rw_init(krwlock_t *rwlp, char *name, int type, void *arg) void rw_destroy(krwlock_t *rwlp) { - ASSERT(rwlp->rw_magic == RW_MAGIC); + ASSERT3U(rwlp->rw_magic, ==, RW_MAGIC); VERIFY3S(pthread_rwlock_destroy(&rwlp->rw_lock), ==, 0); rwlp->rw_magic = 0; @@ -247,18 +341,18 @@ rw_destroy(krwlock_t *rwlp) void rw_enter(krwlock_t *rwlp, krw_t rw) { - ASSERT(rwlp->rw_magic == RW_MAGIC); - ASSERT(rwlp->rw_owner != curthread); - ASSERT(rwlp->rw_wr_owner != curthread); + ASSERT3U(rwlp->rw_magic, ==, RW_MAGIC); + ASSERT3P(rwlp->rw_owner, !=, curthread); + ASSERT3P(rwlp->rw_wr_owner, !=, curthread); if (rw == RW_READER) { VERIFY3S(pthread_rwlock_rdlock(&rwlp->rw_lock), ==, 0); - ASSERT(rwlp->rw_wr_owner == NULL); + ASSERT3P(rwlp->rw_wr_owner, ==, RW_INIT); atomic_inc_uint(&rwlp->rw_readers); } else { VERIFY3S(pthread_rwlock_wrlock(&rwlp->rw_lock), ==, 0); - ASSERT(rwlp->rw_wr_owner == NULL); + ASSERT3P(rwlp->rw_wr_owner, ==, RW_INIT); ASSERT3U(rwlp->rw_readers, ==, 0); rwlp->rw_wr_owner = curthread; @@ -270,15 +364,15 @@ rw_enter(krwlock_t *rwlp, krw_t rw) void rw_exit(krwlock_t *rwlp) { - ASSERT(rwlp->rw_magic == RW_MAGIC); + ASSERT3U(rwlp->rw_magic, ==, RW_MAGIC); ASSERT(RW_LOCK_HELD(rwlp)); if (RW_READ_HELD(rwlp)) atomic_dec_uint(&rwlp->rw_readers); else - rwlp->rw_wr_owner = NULL; + rwlp->rw_wr_owner = RW_INIT; - rwlp->rw_owner = NULL; + rwlp->rw_owner = RW_INIT; VERIFY3S(pthread_rwlock_unlock(&rwlp->rw_lock), ==, 0); } @@ -287,7 +381,7 @@ rw_tryenter(krwlock_t *rwlp, krw_t rw) { int rv; - ASSERT(rwlp->rw_magic == RW_MAGIC); + ASSERT3U(rwlp->rw_magic, ==, RW_MAGIC); if (rw == RW_READER) rv = pthread_rwlock_tryrdlock(&rwlp->rw_lock); @@ -295,7 +389,7 @@ rw_tryenter(krwlock_t *rwlp, krw_t rw) rv = pthread_rwlock_trywrlock(&rwlp->rw_lock); if (rv == 0) { - ASSERT(rwlp->rw_wr_owner == NULL); + ASSERT3P(rwlp->rw_wr_owner, ==, RW_INIT); if (rw == RW_READER) atomic_inc_uint(&rwlp->rw_readers); @@ -317,7 +411,7 @@ rw_tryenter(krwlock_t *rwlp, krw_t rw) int rw_tryupgrade(krwlock_t *rwlp) { - ASSERT(rwlp->rw_magic == RW_MAGIC); + ASSERT3U(rwlp->rw_magic, ==, RW_MAGIC); return (0); } @@ -331,10 +425,10 @@ rw_tryupgrade(krwlock_t *rwlp) void cv_init(kcondvar_t *cv, char *name, int type, void *arg) { - ASSERT(type == CV_DEFAULT); + ASSERT3S(type, ==, CV_DEFAULT); #ifdef IM_FEELING_LUCKY - ASSERT(cv->cv_magic != CV_MAGIC); + ASSERT3U(cv->cv_magic, !=, CV_MAGIC); #endif cv->cv_magic = CV_MAGIC; @@ -345,7 +439,7 @@ cv_init(kcondvar_t *cv, char *name, int type, void *arg) void cv_destroy(kcondvar_t *cv) { - ASSERT(cv->cv_magic == CV_MAGIC); + ASSERT3U(cv->cv_magic, ==, CV_MAGIC); VERIFY3S(pthread_cond_destroy(&cv->cv), ==, 0); cv->cv_magic = 0; } @@ -353,9 +447,9 @@ cv_destroy(kcondvar_t *cv) void cv_wait(kcondvar_t *cv, kmutex_t *mp) { - ASSERT(cv->cv_magic == CV_MAGIC); - ASSERT(mutex_owner(mp) == curthread); - mp->m_owner = NULL; + ASSERT3U(cv->cv_magic, ==, CV_MAGIC); + ASSERT3P(mutex_owner(mp), ==, curthread); + mp->m_owner = MTX_INIT; int ret = pthread_cond_wait(&cv->cv, &mp->m_lock); if (ret != 0) VERIFY3S(ret, ==, EINTR); @@ -370,7 +464,7 @@ cv_timedwait(kcondvar_t *cv, kmutex_t *mp, clock_t abstime) timestruc_t ts; clock_t delta; - ASSERT(cv->cv_magic == CV_MAGIC); + ASSERT3U(cv->cv_magic, ==, CV_MAGIC); top: delta = abstime - lbolt; @@ -386,8 +480,8 @@ top: ts.tv_nsec -= NANOSEC; } - ASSERT(mutex_owner(mp) == curthread); - mp->m_owner = NULL; + ASSERT3P(mutex_owner(mp), ==, curthread); + mp->m_owner = MTX_INIT; error = pthread_cond_timedwait(&cv->cv, &mp->m_lock, &ts); mp->m_owner = curthread; @@ -405,14 +499,14 @@ top: void cv_signal(kcondvar_t *cv) { - ASSERT(cv->cv_magic == CV_MAGIC); + ASSERT3U(cv->cv_magic, ==, CV_MAGIC); VERIFY3S(pthread_cond_signal(&cv->cv), ==, 0); } void cv_broadcast(kcondvar_t *cv) { - ASSERT(cv->cv_magic == CV_MAGIC); + ASSERT3U(cv->cv_magic, ==, CV_MAGIC); VERIFY3S(pthread_cond_broadcast(&cv->cv), ==, 0); } @@ -896,12 +990,8 @@ kernel_init(int mode) VERIFY((random_fd = open("/dev/random", O_RDONLY)) != -1); VERIFY((urandom_fd = open("/dev/urandom", O_RDONLY)) != -1); - mutex_init(&kthread_lock, NULL, MUTEX_DEFAULT, NULL); - list_create(&kthread_list, sizeof (kthread_t), - offsetof(kthread_t, t_node)); - + thread_init(); system_taskq_init(); - spa_init(mode); } @@ -909,9 +999,8 @@ void kernel_fini(void) { spa_fini(); - - list_destroy(&kthread_list); - mutex_destroy(&kthread_lock); + system_taskq_fini(); + thread_fini(); close(random_fd); close(urandom_fd); diff --git a/lib/libzpool/taskq.c b/lib/libzpool/taskq.c index 40dfb6789..42e2dd3f4 100644 --- a/lib/libzpool/taskq.c +++ b/lib/libzpool/taskq.c @@ -43,6 +43,7 @@ struct taskq { kcondvar_t tq_dispatch_cv; kcondvar_t tq_wait_cv; kthread_t **tq_threadlist; + kt_did_t *tq_idlist; int tq_flags; int tq_active; int tq_nthreads; @@ -163,6 +164,7 @@ taskq_thread(void *arg) tq->tq_nthreads--; cv_broadcast(&tq->tq_wait_cv); mutex_exit(&tq->tq_lock); + thread_exit(); return (NULL); } @@ -198,7 +200,10 @@ taskq_create(const char *name, int nthreads, pri_t pri, tq->tq_maxalloc = maxalloc; tq->tq_task.task_next = &tq->tq_task; tq->tq_task.task_prev = &tq->tq_task; - tq->tq_threadlist = kmem_alloc(nthreads*sizeof(kthread_t *), KM_SLEEP); + VERIFY3P((tq->tq_threadlist = kmem_alloc(tq->tq_nthreads * + sizeof(kthread_t *), KM_SLEEP)), !=, NULL); + VERIFY3P((tq->tq_idlist = kmem_alloc(tq->tq_nthreads * + sizeof(kt_did_t), KM_SLEEP)), !=, NULL); if (flags & TASKQ_PREPOPULATE) { mutex_enter(&tq->tq_lock); @@ -207,9 +212,11 @@ taskq_create(const char *name, int nthreads, pri_t pri, mutex_exit(&tq->tq_lock); } - for (t = 0; t < nthreads; t++) + for (t = 0; t < tq->tq_nthreads; t++) { VERIFY((tq->tq_threadlist[t] = thread_create(NULL, 0, taskq_thread, tq, THR_BOUND, NULL, 0, 0)) != NULL); + tq->tq_idlist[t] = tq->tq_threadlist[t]->t_tid; + } return (tq); } @@ -239,9 +246,10 @@ taskq_destroy(taskq_t *tq) mutex_exit(&tq->tq_lock); for (t = 0; t < nthreads; t++) - (void) thr_join(tq->tq_threadlist[t], NULL, NULL); + VERIFY3S(thread_join(tq->tq_idlist[t], NULL, NULL), ==, 0); kmem_free(tq->tq_threadlist, nthreads * sizeof(kthread_t *)); + kmem_free(tq->tq_idlist, nthreads * sizeof(kt_did_t)); rw_destroy(&tq->tq_threadlock); mutex_destroy(&tq->tq_lock); @@ -272,3 +280,9 @@ system_taskq_init(void) system_taskq = taskq_create("system_taskq", 64, minclsyspri, 4, 512, TASKQ_DYNAMIC | TASKQ_PREPOPULATE); } + +void +system_taskq_fini(void) +{ + taskq_destroy(system_taskq); +} diff --git a/module/zfs/txg.c b/module/zfs/txg.c index b5fcc8c4a..15745b3e9 100644 --- a/module/zfs/txg.c +++ b/module/zfs/txg.c @@ -446,6 +446,8 @@ txg_sync_thread(dsl_pool_t *dp) rw_exit(&tx->tx_suspend); cv_broadcast(&tx->tx_sync_done_cv); } + + thread_exit(); } static void @@ -490,6 +492,8 @@ txg_quiesce_thread(dsl_pool_t *dp) cv_broadcast(&tx->tx_sync_more_cv); cv_broadcast(&tx->tx_quiesce_done_cv); } + + thread_exit(); } /* |