diff options
author | Debabrata Banerjee <[email protected]> | 2017-03-15 19:31:56 -0400 |
---|---|---|
committer | Brian Behlendorf <[email protected]> | 2017-05-02 15:50:13 -0400 |
commit | 44813aefad9a685be92214a331f413806974a419 (patch) | |
tree | 7866833f84224050b39f2c32573138d08b572df0 /module | |
parent | 1a31dcf53cd71c3c71db343fc4c9ab1dfa326d3f (diff) |
Don't run the reaper if we didn't shrink the cache
Calling it when nothing is evictable will cause extra kswapd cpu. Also
if we didn't shrink it's unlikely to have memory to reap because we
likely just called it microseconds ago. The exception is if we are in
direct reclaim.
You can see how hard this is being hit in kswapd with a light test
workload:
34.95% [zfs] [k] arc_kmem_reap_now
5.40% [spl] [k] spl_kmem_cache_reap_now
3.79% [kernel] [k] _raw_spin_lock
2.86% [spl] [k] __spl_kmem_cache_generic_shrinker.isra.7
2.70% [kernel] [k] shrink_slab.part.37
1.93% [kernel] [k] isolate_lru_pages.isra.43
1.55% [kernel] [k] __wake_up_bit
1.20% [kernel] [k] super_cache_count
1.20% [kernel] [k] __radix_tree_lookup
With ZFS just mounted but only ext4/pagecache memory pressure
arc_kmem_reap_now still consumes excessive CPU:
12.69% [kernel] [k] isolate_lru_pages.isra.43
10.76% [kernel] [k] free_pcppages_bulk
7.98% [kernel] [k] drop_buffers
7.31% [kernel] [k] shrink_page_list
6.44% [zfs] [k] arc_kmem_reap_now
4.19% [kernel] [k] free_hot_cold_page
4.00% [kernel] [k] __slab_free
3.95% [kernel] [k] __isolate_lru_page
3.09% [kernel] [k] __radix_tree_lookup
Same pagecache only workload as above with this patch series:
11.58% [kernel] [k] isolate_lru_pages.isra.43
11.20% [kernel] [k] drop_buffers
9.67% [kernel] [k] free_pcppages_bulk
8.44% [kernel] [k] shrink_page_list
4.86% [kernel] [k] __isolate_lru_page
4.43% [kernel] [k] free_hot_cold_page
4.00% [kernel] [k] __slab_free
3.44% [kernel] [k] __radix_tree_lookup
(arc_kmem_reap_now has 0 samples in perf)
AKAMAI: zfs: CR 3695042
Reviewed-by: Tim Chase <[email protected]>
Reviewed-by: Richard Yao <[email protected]>
Reviewed-by: Brian Behlendorf <[email protected]>
Signed-off-by: Debabrata Banerjee <[email protected]>
Issue #6035
Diffstat (limited to 'module')
-rw-r--r-- | module/zfs/arc.c | 11 |
1 files changed, 5 insertions, 6 deletions
diff --git a/module/zfs/arc.c b/module/zfs/arc.c index 79b04f455..efac973ba 100644 --- a/module/zfs/arc.c +++ b/module/zfs/arc.c @@ -4423,12 +4423,12 @@ __arc_shrinker_func(struct shrinker *shrink, struct shrink_control *sc) /* * Evict the requested number of pages by shrinking arc_c the - * requested amount. If there is nothing left to evict just - * reap whatever we can from the various arc slabs. + * requested amount. */ if (pages > 0) { arc_shrink(ptob(sc->nr_to_scan)); - arc_kmem_reap_now(); + if (current_is_kswapd()) + arc_kmem_reap_now(); #ifdef HAVE_SPLIT_SHRINKER_CALLBACK pages = MAX(pages - btop(arc_evictable_memory()), 0); #else @@ -4439,10 +4439,8 @@ __arc_shrinker_func(struct shrinker *shrink, struct shrink_control *sc) */ cv_broadcast(&arc_reclaim_waiters_cv); - } else { - arc_kmem_reap_now(); + } else pages = SHRINK_STOP; - } /* * When direct reclaim is observed it usually indicates a rapid @@ -4455,6 +4453,7 @@ __arc_shrinker_func(struct shrinker *shrink, struct shrink_control *sc) ARCSTAT_BUMP(arcstat_memory_indirect_count); } else { arc_no_grow = B_TRUE; + arc_kmem_reap_now(); ARCSTAT_BUMP(arcstat_memory_direct_count); } |