[Top] [All Lists]

[PATCH] libxfs: increase hash chain depth when we run out of slots

To: xfs-oss <xfs@xxxxxxxxxxx>
Subject: [PATCH] libxfs: increase hash chain depth when we run out of slots
From: Eric Sandeen <sandeen@xxxxxxxxxx>
Date: Thu, 17 Sep 2009 11:06:16 -0500
Cc: Tomek Kruszona <bloodyscarion@xxxxxxxxx>, Riku Paananen <riku.paananen@xxxxxxxxxxx>
User-agent: Thunderbird (X11/20090320)
A couple people reported xfs_repair hangs after
"Traversing filesystem ..." in xfs_repair.  This happens
when all slots in the cache are full and referenced, and the
loop in cache_node_get() which tries to shake unused entries
fails to find any - it just keeps upping the priority and goes

This can be worked around by restarting xfs_repair with
-P and/or "-o bhash=<largersize>" for older xfs_repair.

I started down the path of increasing the number of hash buckets
on the fly, but Barry suggested simply increasing the max allowed
depth which is much simpler (thanks!)

Resizing the hash lengths does mean that cache_report ends up with
most things in the "greater-than" category:

Hash buckets with  23 entries      3 (  3%)
Hash buckets with  24 entries      3 (  3%)
Hash buckets with >24 entries     50 ( 85%)

but I think I'll save that fix for another patch unless there's
real concern right now.

I tested this on the metadump image provided by Tomek.

Signed-off-by: Eric Sandeen <sandeen@xxxxxxxxxxx>
Reported-by: Tomek Kruszona <bloodyscarion@xxxxxxxxx>
Reported-by: Riku Paananen <riku.paananen@xxxxxxxxxxx>

diff --git a/libxfs/cache.c b/libxfs/cache.c
index 48f91d7..56b24e7 100644
--- a/libxfs/cache.c
+++ b/libxfs/cache.c
@@ -83,6 +83,18 @@ cache_init(
+       struct cache *          cache)
+       pthread_mutex_lock(&cache->c_mutex);
+       fprintf(stderr, "doubling cache size to %d\n", 2 * cache->c_maxcount);
+       cache->c_maxcount *= 2;
+       pthread_mutex_unlock(&cache->c_mutex);
        struct cache *          cache,
        cache_walk_t            visit)
@@ -344,6 +356,15 @@ cache_node_get(
                if (node)
                priority = cache_shake(cache, priority, 0);
+               /*
+                * We start at 0; if we free CACHE_SHAKE_COUNT we get
+                * back the same priority, if not we get back priority+1.
+                * If we exceed CACHE_MAX_PRIORITY all slots are full; grow it.
+                */
+               if (priority > CACHE_MAX_PRIORITY) {
+                       priority = 0;
+                       cache_expand(cache);
+               }
        node->cn_hashidx = hashidx;

<Prev in Thread] Current Thread [Next in Thread>