summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMatthew Ahrens <[email protected]>2020-07-22 09:51:47 -0700
committerGitHub <[email protected]>2020-07-22 09:51:47 -0700
commit5dd92909c6d1ed606aac7c6ffe3df68513bb5ed1 (patch)
tree6e54b699d4a1044f53124de3ba816103ef7111e0
parent317dbea1735f3cd3eaec70bcb31b3a7ba0721923 (diff)
Adjust ARC terminology
The process of evicting data from the ARC is referred to as `arc_adjust`. This commit changes the term to `arc_evict`, which is more specific. Reviewed-by: George Wilson <[email protected]> Reviewed-by: Brian Behlendorf <[email protected]> Reviewed-by: George Melikov <[email protected]> Signed-off-by: Matthew Ahrens <[email protected]> Closes #10592
-rw-r--r--include/sys/arc_impl.h8
-rw-r--r--module/os/freebsd/zfs/arc_os.c10
-rw-r--r--module/os/linux/zfs/arc_os.c20
-rw-r--r--module/zfs/arc.c156
4 files changed, 97 insertions, 97 deletions
diff --git a/include/sys/arc_impl.h b/include/sys/arc_impl.h
index e43cf57d8..1ec4a67d3 100644
--- a/include/sys/arc_impl.h
+++ b/include/sys/arc_impl.h
@@ -878,10 +878,10 @@ extern hrtime_t arc_growtime;
extern boolean_t arc_warm;
extern int arc_grow_retry;
extern int arc_shrink_shift;
-extern zthr_t *arc_adjust_zthr;
-extern kmutex_t arc_adjust_lock;
-extern kcondvar_t arc_adjust_waiters_cv;
-extern boolean_t arc_adjust_needed;
+extern zthr_t *arc_evict_zthr;
+extern kmutex_t arc_evict_lock;
+extern kcondvar_t arc_evict_waiters_cv;
+extern boolean_t arc_evict_needed;
extern kmutex_t arc_prune_mtx;
extern list_t arc_prune_list;
extern aggsum_t arc_size;
diff --git a/module/os/freebsd/zfs/arc_os.c b/module/os/freebsd/zfs/arc_os.c
index 87e517b58..fc88e6e59 100644
--- a/module/os/freebsd/zfs/arc_os.c
+++ b/module/os/freebsd/zfs/arc_os.c
@@ -223,9 +223,9 @@ arc_lowmem(void *arg __unused, int howto __unused)
DTRACE_PROBE2(arc__needfree, int64_t, free_memory, int64_t, to_free);
arc_reduce_target_size(to_free);
- mutex_enter(&arc_adjust_lock);
- arc_adjust_needed = B_TRUE;
- zthr_wakeup(arc_adjust_zthr);
+ mutex_enter(&arc_evict_lock);
+ arc_evict_needed = B_TRUE;
+ zthr_wakeup(arc_evict_zthr);
/*
* It is unsafe to block here in arbitrary threads, because we can come
@@ -233,8 +233,8 @@ arc_lowmem(void *arg __unused, int howto __unused)
* with ARC reclaim thread.
*/
if (curproc == pageproc)
- (void) cv_wait(&arc_adjust_waiters_cv, &arc_adjust_lock);
- mutex_exit(&arc_adjust_lock);
+ (void) cv_wait(&arc_evict_waiters_cv, &arc_evict_lock);
+ mutex_exit(&arc_evict_lock);
}
void
diff --git a/module/os/linux/zfs/arc_os.c b/module/os/linux/zfs/arc_os.c
index 2a119d4e8..e9c0f349d 100644
--- a/module/os/linux/zfs/arc_os.c
+++ b/module/os/linux/zfs/arc_os.c
@@ -194,12 +194,12 @@ arc_shrinker_scan(struct shrinker *shrink, struct shrink_control *sc)
return (SHRINK_STOP);
/* Reclaim in progress */
- if (mutex_tryenter(&arc_adjust_lock) == 0) {
+ if (mutex_tryenter(&arc_evict_lock) == 0) {
ARCSTAT_INCR(arcstat_need_free, ptob(sc->nr_to_scan));
return (0);
}
- mutex_exit(&arc_adjust_lock);
+ mutex_exit(&arc_evict_lock);
/*
* Evict the requested number of pages by shrinking arc_c the
@@ -213,17 +213,17 @@ arc_shrinker_scan(struct shrinker *shrink, struct shrink_control *sc)
* drastically, potentially all the way to arc_c_min. While
* arc_c is below arc_size, ZFS can't process read/write
* requests, because arc_get_data_impl() will block. To
- * ensure that arc_c doesn't shrink faster than the adjust
+ * ensure that arc_c doesn't shrink faster than the evict
* thread can keep up, we wait for eviction here.
*/
- mutex_enter(&arc_adjust_lock);
+ mutex_enter(&arc_evict_lock);
if (arc_is_overflowing()) {
- arc_adjust_needed = B_TRUE;
- zthr_wakeup(arc_adjust_zthr);
- (void) cv_wait(&arc_adjust_waiters_cv,
- &arc_adjust_lock);
+ arc_evict_needed = B_TRUE;
+ zthr_wakeup(arc_evict_zthr);
+ (void) cv_wait(&arc_evict_waiters_cv,
+ &arc_evict_lock);
}
- mutex_exit(&arc_adjust_lock);
+ mutex_exit(&arc_evict_lock);
if (current_is_kswapd())
arc_kmem_reap_soon();
@@ -232,7 +232,7 @@ arc_shrinker_scan(struct shrinker *shrink, struct shrink_control *sc)
/*
* We've shrunk what we can, wake up threads.
*/
- cv_broadcast(&arc_adjust_waiters_cv);
+ cv_broadcast(&arc_evict_waiters_cv);
} else
pages = SHRINK_STOP;
diff --git a/module/zfs/arc.c b/module/zfs/arc.c
index 5b9df43d2..785ad9775 100644
--- a/module/zfs/arc.c
+++ b/module/zfs/arc.c
@@ -317,13 +317,13 @@ static zthr_t *arc_reap_zthr;
/*
* This thread's job is to keep arc_size under arc_c, by calling
- * arc_adjust(), which improves arc_is_overflowing().
+ * arc_evict(), which improves arc_is_overflowing().
*/
-zthr_t *arc_adjust_zthr;
+zthr_t *arc_evict_zthr;
-kmutex_t arc_adjust_lock;
-kcondvar_t arc_adjust_waiters_cv;
-boolean_t arc_adjust_needed = B_FALSE;
+kmutex_t arc_evict_lock;
+kcondvar_t arc_evict_waiters_cv;
+boolean_t arc_evict_needed = B_FALSE;
/*
* The number of headers to evict in arc_evict_state_impl() before
@@ -3954,13 +3954,13 @@ arc_evict_state_impl(multilist_t *ml, int idx, arc_buf_hdr_t *marker,
*
* If threads are left sleeping, due to not
* using cv_broadcast here, they will be woken
- * up via cv_broadcast in arc_adjust_cb() just
- * before arc_adjust_zthr sleeps.
+ * up via cv_broadcast in arc_evict_cb() just
+ * before arc_evict_zthr sleeps.
*/
- mutex_enter(&arc_adjust_lock);
+ mutex_enter(&arc_evict_lock);
if (!arc_is_overflowing())
- cv_signal(&arc_adjust_waiters_cv);
- mutex_exit(&arc_adjust_lock);
+ cv_signal(&arc_evict_waiters_cv);
+ mutex_exit(&arc_evict_lock);
} else {
ARCSTAT_BUMP(arcstat_mutex_miss);
}
@@ -4021,7 +4021,7 @@ arc_evict_state(arc_state_t *state, uint64_t spa, int64_t bytes,
/*
* A b_spa of 0 is used to indicate that this header is
- * a marker. This fact is used in arc_adjust_type() and
+ * a marker. This fact is used in arc_evict_type() and
* arc_evict_state_impl().
*/
markers[i]->b_spa = 0;
@@ -4156,7 +4156,7 @@ arc_flush_state(arc_state_t *state, uint64_t spa, arc_buf_contents_t type,
* evict everything it can, when passed a negative value for "bytes".
*/
static uint64_t
-arc_adjust_impl(arc_state_t *state, uint64_t spa, int64_t bytes,
+arc_evict_impl(arc_state_t *state, uint64_t spa, int64_t bytes,
arc_buf_contents_t type)
{
int64_t delta;
@@ -4188,7 +4188,7 @@ arc_adjust_impl(arc_state_t *state, uint64_t spa, int64_t bytes,
* available for reclaim.
*/
static uint64_t
-arc_adjust_meta_balanced(uint64_t meta_used)
+arc_evict_meta_balanced(uint64_t meta_used)
{
int64_t delta, prune = 0, adjustmnt;
uint64_t total_evicted = 0;
@@ -4198,7 +4198,7 @@ arc_adjust_meta_balanced(uint64_t meta_used)
restart:
/*
* This slightly differs than the way we evict from the mru in
- * arc_adjust because we don't have a "target" value (i.e. no
+ * arc_evict because we don't have a "target" value (i.e. no
* "meta" arc_p). As a result, I think we can completely
* cannibalize the metadata in the MRU before we evict the
* metadata from the MFU. I think we probably need to implement a
@@ -4210,7 +4210,7 @@ restart:
zfs_refcount_count(&arc_mru->arcs_esize[type]) > 0) {
delta = MIN(zfs_refcount_count(&arc_mru->arcs_esize[type]),
adjustmnt);
- total_evicted += arc_adjust_impl(arc_mru, 0, delta, type);
+ total_evicted += arc_evict_impl(arc_mru, 0, delta, type);
adjustmnt -= delta;
}
@@ -4228,7 +4228,7 @@ restart:
zfs_refcount_count(&arc_mfu->arcs_esize[type]) > 0) {
delta = MIN(zfs_refcount_count(&arc_mfu->arcs_esize[type]),
adjustmnt);
- total_evicted += arc_adjust_impl(arc_mfu, 0, delta, type);
+ total_evicted += arc_evict_impl(arc_mfu, 0, delta, type);
}
adjustmnt = meta_used - arc_meta_limit;
@@ -4237,7 +4237,7 @@ restart:
zfs_refcount_count(&arc_mru_ghost->arcs_esize[type]) > 0) {
delta = MIN(adjustmnt,
zfs_refcount_count(&arc_mru_ghost->arcs_esize[type]));
- total_evicted += arc_adjust_impl(arc_mru_ghost, 0, delta, type);
+ total_evicted += arc_evict_impl(arc_mru_ghost, 0, delta, type);
adjustmnt -= delta;
}
@@ -4245,7 +4245,7 @@ restart:
zfs_refcount_count(&arc_mfu_ghost->arcs_esize[type]) > 0) {
delta = MIN(adjustmnt,
zfs_refcount_count(&arc_mfu_ghost->arcs_esize[type]));
- total_evicted += arc_adjust_impl(arc_mfu_ghost, 0, delta, type);
+ total_evicted += arc_evict_impl(arc_mfu_ghost, 0, delta, type);
}
/*
@@ -4280,7 +4280,7 @@ restart:
* capped by the arc_meta_limit tunable.
*/
static uint64_t
-arc_adjust_meta_only(uint64_t meta_used)
+arc_evict_meta_only(uint64_t meta_used)
{
uint64_t total_evicted = 0;
int64_t target;
@@ -4296,7 +4296,7 @@ arc_adjust_meta_only(uint64_t meta_used)
(int64_t)(zfs_refcount_count(&arc_anon->arcs_size) +
zfs_refcount_count(&arc_mru->arcs_size) - arc_p));
- total_evicted += arc_adjust_impl(arc_mru, 0, target, ARC_BUFC_METADATA);
+ total_evicted += arc_evict_impl(arc_mru, 0, target, ARC_BUFC_METADATA);
/*
* Similar to the above, we want to evict enough bytes to get us
@@ -4307,18 +4307,18 @@ arc_adjust_meta_only(uint64_t meta_used)
(int64_t)(zfs_refcount_count(&arc_mfu->arcs_size) -
(arc_c - arc_p)));
- total_evicted += arc_adjust_impl(arc_mfu, 0, target, ARC_BUFC_METADATA);
+ total_evicted += arc_evict_impl(arc_mfu, 0, target, ARC_BUFC_METADATA);
return (total_evicted);
}
static uint64_t
-arc_adjust_meta(uint64_t meta_used)
+arc_evict_meta(uint64_t meta_used)
{
if (zfs_arc_meta_strategy == ARC_STRATEGY_META_ONLY)
- return (arc_adjust_meta_only(meta_used));
+ return (arc_evict_meta_only(meta_used));
else
- return (arc_adjust_meta_balanced(meta_used));
+ return (arc_evict_meta_balanced(meta_used));
}
/*
@@ -4330,7 +4330,7 @@ arc_adjust_meta(uint64_t meta_used)
* returned.
*/
static arc_buf_contents_t
-arc_adjust_type(arc_state_t *state)
+arc_evict_type(arc_state_t *state)
{
multilist_t *data_ml = state->arcs_list[ARC_BUFC_DATA];
multilist_t *meta_ml = state->arcs_list[ARC_BUFC_METADATA];
@@ -4400,7 +4400,7 @@ arc_adjust_type(arc_state_t *state)
* Evict buffers from the cache, such that arc_size is capped by arc_c.
*/
static uint64_t
-arc_adjust(void)
+arc_evict(void)
{
uint64_t total_evicted = 0;
uint64_t bytes;
@@ -4412,7 +4412,7 @@ arc_adjust(void)
* If we're over arc_meta_limit, we want to correct that before
* potentially evicting data buffers below.
*/
- total_evicted += arc_adjust_meta(ameta);
+ total_evicted += arc_evict_meta(ameta);
/*
* Adjust MRU size
@@ -4436,9 +4436,9 @@ arc_adjust(void)
* type. If we cannot satisfy the number of bytes from this
* type, spill over into the next type.
*/
- if (arc_adjust_type(arc_mru) == ARC_BUFC_METADATA &&
+ if (arc_evict_type(arc_mru) == ARC_BUFC_METADATA &&
ameta > arc_meta_min) {
- bytes = arc_adjust_impl(arc_mru, 0, target, ARC_BUFC_METADATA);
+ bytes = arc_evict_impl(arc_mru, 0, target, ARC_BUFC_METADATA);
total_evicted += bytes;
/*
@@ -4448,9 +4448,9 @@ arc_adjust(void)
target -= bytes;
total_evicted +=
- arc_adjust_impl(arc_mru, 0, target, ARC_BUFC_DATA);
+ arc_evict_impl(arc_mru, 0, target, ARC_BUFC_DATA);
} else {
- bytes = arc_adjust_impl(arc_mru, 0, target, ARC_BUFC_DATA);
+ bytes = arc_evict_impl(arc_mru, 0, target, ARC_BUFC_DATA);
total_evicted += bytes;
/*
@@ -4460,7 +4460,7 @@ arc_adjust(void)
target -= bytes;
total_evicted +=
- arc_adjust_impl(arc_mru, 0, target, ARC_BUFC_METADATA);
+ arc_evict_impl(arc_mru, 0, target, ARC_BUFC_METADATA);
}
/*
@@ -4479,9 +4479,9 @@ arc_adjust(void)
*/
target = asize - arc_c;
- if (arc_adjust_type(arc_mfu) == ARC_BUFC_METADATA &&
+ if (arc_evict_type(arc_mfu) == ARC_BUFC_METADATA &&
ameta > arc_meta_min) {
- bytes = arc_adjust_impl(arc_mfu, 0, target, ARC_BUFC_METADATA);
+ bytes = arc_evict_impl(arc_mfu, 0, target, ARC_BUFC_METADATA);
total_evicted += bytes;
/*
@@ -4491,9 +4491,9 @@ arc_adjust(void)
target -= bytes;
total_evicted +=
- arc_adjust_impl(arc_mfu, 0, target, ARC_BUFC_DATA);
+ arc_evict_impl(arc_mfu, 0, target, ARC_BUFC_DATA);
} else {
- bytes = arc_adjust_impl(arc_mfu, 0, target, ARC_BUFC_DATA);
+ bytes = arc_evict_impl(arc_mfu, 0, target, ARC_BUFC_DATA);
total_evicted += bytes;
/*
@@ -4503,7 +4503,7 @@ arc_adjust(void)
target -= bytes;
total_evicted +=
- arc_adjust_impl(arc_mfu, 0, target, ARC_BUFC_METADATA);
+ arc_evict_impl(arc_mfu, 0, target, ARC_BUFC_METADATA);
}
/*
@@ -4520,13 +4520,13 @@ arc_adjust(void)
target = zfs_refcount_count(&arc_mru->arcs_size) +
zfs_refcount_count(&arc_mru_ghost->arcs_size) - arc_c;
- bytes = arc_adjust_impl(arc_mru_ghost, 0, target, ARC_BUFC_DATA);
+ bytes = arc_evict_impl(arc_mru_ghost, 0, target, ARC_BUFC_DATA);
total_evicted += bytes;
target -= bytes;
total_evicted +=
- arc_adjust_impl(arc_mru_ghost, 0, target, ARC_BUFC_METADATA);
+ arc_evict_impl(arc_mru_ghost, 0, target, ARC_BUFC_METADATA);
/*
* We assume the sum of the mru list and mfu list is less than
@@ -4539,13 +4539,13 @@ arc_adjust(void)
target = zfs_refcount_count(&arc_mru_ghost->arcs_size) +
zfs_refcount_count(&arc_mfu_ghost->arcs_size) - arc_c;
- bytes = arc_adjust_impl(arc_mfu_ghost, 0, target, ARC_BUFC_DATA);
+ bytes = arc_evict_impl(arc_mfu_ghost, 0, target, ARC_BUFC_DATA);
total_evicted += bytes;
target -= bytes;
total_evicted +=
- arc_adjust_impl(arc_mfu_ghost, 0, target, ARC_BUFC_METADATA);
+ arc_evict_impl(arc_mfu_ghost, 0, target, ARC_BUFC_METADATA);
return (total_evicted);
}
@@ -4596,11 +4596,11 @@ arc_reduce_target_size(int64_t to_free)
}
if (asize > arc_c) {
- /* See comment in arc_adjust_cb_check() on why lock+flag */
- mutex_enter(&arc_adjust_lock);
- arc_adjust_needed = B_TRUE;
- mutex_exit(&arc_adjust_lock);
- zthr_wakeup(arc_adjust_zthr);
+ /* See comment in arc_evict_cb_check() on why lock+flag */
+ mutex_enter(&arc_evict_lock);
+ arc_evict_needed = B_TRUE;
+ mutex_exit(&arc_evict_lock);
+ zthr_wakeup(arc_evict_zthr);
}
}
@@ -4665,7 +4665,7 @@ arc_kmem_reap_soon(void)
/* ARGSUSED */
static boolean_t
-arc_adjust_cb_check(void *arg, zthr_t *zthr)
+arc_evict_cb_check(void *arg, zthr_t *zthr)
{
/*
* This is necessary so that any changes which may have been made to
@@ -4684,7 +4684,7 @@ arc_adjust_cb_check(void *arg, zthr_t *zthr)
* this call, these commands may show stale stats for the
* anon, mru, mru_ghost, mfu, and mfu_ghost lists. Even
* with this change, the data might be up to 1 second
- * out of date(the arc_adjust_zthr has a maximum sleep
+ * out of date(the arc_evict_zthr has a maximum sleep
* time of 1 second); but that should suffice. The
* arc_state_t structures can be queried directly if more
* accurate information is needed.
@@ -4693,60 +4693,60 @@ arc_adjust_cb_check(void *arg, zthr_t *zthr)
arc_ksp->ks_update(arc_ksp, KSTAT_READ);
/*
- * We have to rely on arc_get_data_impl() to tell us when to adjust,
+ * We have to rely on arc_get_data_impl() to tell us when to evict,
* rather than checking if we are overflowing here, so that we are
* sure to not leave arc_get_data_impl() waiting on
- * arc_adjust_waiters_cv. If we have become "not overflowing" since
+ * arc_evict_waiters_cv. If we have become "not overflowing" since
* arc_get_data_impl() checked, we need to wake it up. We could
* broadcast the CV here, but arc_get_data_impl() may have not yet
* gone to sleep. We would need to use a mutex to ensure that this
* function doesn't broadcast until arc_get_data_impl() has gone to
- * sleep (e.g. the arc_adjust_lock). However, the lock ordering of
+ * sleep (e.g. the arc_evict_lock). However, the lock ordering of
* such a lock would necessarily be incorrect with respect to the
* zthr_lock, which is held before this function is called, and is
* held by arc_get_data_impl() when it calls zthr_wakeup().
*/
- return (arc_adjust_needed);
+ return (arc_evict_needed);
}
/*
- * Keep arc_size under arc_c by running arc_adjust which evicts data
+ * Keep arc_size under arc_c by running arc_evict which evicts data
* from the ARC.
*/
/* ARGSUSED */
static void
-arc_adjust_cb(void *arg, zthr_t *zthr)
+arc_evict_cb(void *arg, zthr_t *zthr)
{
uint64_t evicted = 0;
fstrans_cookie_t cookie = spl_fstrans_mark();
/* Evict from cache */
- evicted = arc_adjust();
+ evicted = arc_evict();
/*
* If evicted is zero, we couldn't evict anything
- * via arc_adjust(). This could be due to hash lock
+ * via arc_evict(). This could be due to hash lock
* collisions, but more likely due to the majority of
* arc buffers being unevictable. Therefore, even if
* arc_size is above arc_c, another pass is unlikely to
* be helpful and could potentially cause us to enter an
* infinite loop. Additionally, zthr_iscancelled() is
* checked here so that if the arc is shutting down, the
- * broadcast will wake any remaining arc adjust waiters.
+ * broadcast will wake any remaining arc evict waiters.
*/
- mutex_enter(&arc_adjust_lock);
- arc_adjust_needed = !zthr_iscancelled(arc_adjust_zthr) &&
+ mutex_enter(&arc_evict_lock);
+ arc_evict_needed = !zthr_iscancelled(arc_evict_zthr) &&
evicted > 0 && aggsum_compare(&arc_size, arc_c) > 0;
- if (!arc_adjust_needed) {
+ if (!arc_evict_needed) {
/*
* We're either no longer overflowing, or we
* can't evict anything more, so we should wake
* arc_get_data_impl() sooner.
*/
- cv_broadcast(&arc_adjust_waiters_cv);
+ cv_broadcast(&arc_evict_waiters_cv);
arc_need_free = 0;
}
- mutex_exit(&arc_adjust_lock);
+ mutex_exit(&arc_evict_lock);
spl_fstrans_unmark(cookie);
}
@@ -4785,7 +4785,7 @@ arc_reap_cb_check(void *arg, zthr_t *zthr)
/*
* Keep enough free memory in the system by reaping the ARC's kmem
* caches. To cause more slabs to be reapable, we may reduce the
- * target size of the cache (arc_c), causing the arc_adjust_cb()
+ * target size of the cache (arc_c), causing the arc_evict_cb()
* to free more buffers.
*/
/* ARGSUSED */
@@ -5036,7 +5036,7 @@ arc_get_data_impl(arc_buf_hdr_t *hdr, uint64_t size, void *tag)
* overflowing; thus we don't use a while loop here.
*/
if (arc_is_overflowing()) {
- mutex_enter(&arc_adjust_lock);
+ mutex_enter(&arc_evict_lock);
/*
* Now that we've acquired the lock, we may no longer be
@@ -5050,12 +5050,12 @@ arc_get_data_impl(arc_buf_hdr_t *hdr, uint64_t size, void *tag)
* shouldn't cause any harm.
*/
if (arc_is_overflowing()) {
- arc_adjust_needed = B_TRUE;
- zthr_wakeup(arc_adjust_zthr);
- (void) cv_wait(&arc_adjust_waiters_cv,
- &arc_adjust_lock);
+ arc_evict_needed = B_TRUE;
+ zthr_wakeup(arc_evict_zthr);
+ (void) cv_wait(&arc_evict_waiters_cv,
+ &arc_evict_lock);
}
- mutex_exit(&arc_adjust_lock);
+ mutex_exit(&arc_evict_lock);
}
VERIFY3U(hdr->b_type, ==, type);
@@ -7268,8 +7268,8 @@ void
arc_init(void)
{
uint64_t percent, allmem = arc_all_memory();
- mutex_init(&arc_adjust_lock, NULL, MUTEX_DEFAULT, NULL);
- cv_init(&arc_adjust_waiters_cv, NULL, CV_DEFAULT, NULL);
+ mutex_init(&arc_evict_lock, NULL, MUTEX_DEFAULT, NULL);
+ cv_init(&arc_evict_waiters_cv, NULL, CV_DEFAULT, NULL);
arc_min_prefetch_ms = 1000;
arc_min_prescient_prefetch_ms = 6000;
@@ -7339,8 +7339,8 @@ arc_init(void)
kstat_install(arc_ksp);
}
- arc_adjust_zthr = zthr_create_timer(arc_adjust_cb_check,
- arc_adjust_cb, NULL, SEC2NSEC(1));
+ arc_evict_zthr = zthr_create_timer(arc_evict_cb_check,
+ arc_evict_cb, NULL, SEC2NSEC(1));
arc_reap_zthr = zthr_create_timer(arc_reap_cb_check,
arc_reap_cb, NULL, SEC2NSEC(1));
@@ -7398,11 +7398,11 @@ arc_fini(void)
list_destroy(&arc_prune_list);
mutex_destroy(&arc_prune_mtx);
- (void) zthr_cancel(arc_adjust_zthr);
+ (void) zthr_cancel(arc_evict_zthr);
(void) zthr_cancel(arc_reap_zthr);
- mutex_destroy(&arc_adjust_lock);
- cv_destroy(&arc_adjust_waiters_cv);
+ mutex_destroy(&arc_evict_lock);
+ cv_destroy(&arc_evict_waiters_cv);
/*
* buf_fini() must proceed arc_state_fini() because buf_fin() may
@@ -7417,7 +7417,7 @@ arc_fini(void)
* torn down to avoid the case of them receiving any
* wakeup() signals after they are destroyed.
*/
- zthr_destroy(arc_adjust_zthr);
+ zthr_destroy(arc_evict_zthr);
zthr_destroy(arc_reap_zthr);
ASSERT0(arc_loaned_bytes);
@@ -10273,7 +10273,7 @@ ZFS_MODULE_PARAM(zfs_arc, zfs_arc_, meta_prune, INT, ZMOD_RW,
"Meta objects to scan for prune");
ZFS_MODULE_PARAM(zfs_arc, zfs_arc_, meta_adjust_restarts, INT, ZMOD_RW,
- "Limit number of restarts in arc_adjust_meta");
+ "Limit number of restarts in arc_evict_meta");
ZFS_MODULE_PARAM(zfs_arc, zfs_arc_, meta_strategy, INT, ZMOD_RW,
"Meta reclaim strategy");