xfs
[Top] [All Lists]

xfs_repair trouble

To: xfs@xxxxxxxxxxx
Subject: xfs_repair trouble
From: "Simon J" <simonjj@xxxxxxxxx>
Date: Sat, 14 Jun 2008 12:56:34 -0400
Dkim-signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:message-id:date:from:to :subject:mime-version:content-type; bh=VNDcFKTOewML7SXRRQsnt4pfQhcNNoJsFX+d89obg08=; b=Zlat2JgBhyQWgPk3mXk26La13GMU6ZL9zNd/DzqoQ1QLkJRcLnBNC1RSWH2KjyijAg SYqvndxd2QOqVdZhGDDBekJKKZ/iPo2PEADBgTNig68W9bWGbvpcx1bYmln5x+DZtnWR rVXhs6t81m379k5MY6IErfdNw914+jNYKizpQ=
Domainkey-signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=message-id:date:from:to:subject:mime-version:content-type; b=qNSm04+Y+hcS2XfM8DR7BzkAEtKMgiHFMZlMNgQnpGMzC9UHyv0WKSbkEWjzqP4YCb cyawWenjft5Sl30KahBfWkC0fmCnFVsKbwv4O2fQXzo1O/UnF1HIQaU2YsLX6byF9uYx SghCZLFNZrK6gSh7/L0nNVkp2Mx0B0xoAfyqU=
Sender: xfs-bounce@xxxxxxxxxxx
Hello all,


I have a corrupted xfs filesystem on top of a RAID 5 (1TB in size).
The RAID is still fully intact but the filesystem was damaged. When
trying to repair the filesystem xfs_repair fails. I have tried various
versions of xfs_repair, the latest stable (2.9.8) and the latest trunk
(from CVS). I'd love to investigate and/or fix the issue further but I
am a bit confused about some of my xfs_repair runs (both done with
trunk).
Could someone shed some light on where the problem could be, I'd be
happy to continue digging if I would only know where roughly

Run 1 ============================================
./xfs_repair -P -m 170 /dev/evms/monster_evms
... more output ...
bad hash table for directory inode 4842 (no data entry): rebuilding
rebuilding directory inode 4842
entry ".." in directory inode 27930072 points to free inode 2013274702
bad hash table for directory inode 27930072 (no data entry): rebuilding
rebuilding directory inode 27930072
bad hash table for directory inode 28776251 (no data entry): rebuilding
rebuilding directory inode 28776251
fixing i8count in inode 29111010
xfs_repair: phase6.c:3411: shortform_dir2_entry_check: Assertion
`bytes_deleted > 0' failed.
Aborted
=================================================
Exits with an Assert, it would be interesting to know why this Assert
is there and what it means for bytes_deleted to be 0.


Run 2 ============================================
./xfs_repair -P -m 166 /dev/evms/monster_evms
... more output ...
bad hash table for directory inode 2013274673 (no data entry): rebuilding
rebuilding directory inode 2013274673
bad hash table for directory inode 2029477400 (no data entry): rebuilding
rebuilding directory inode 2029477400
bad hash table for directory inode 2037825112 (no data entry): rebuilding
rebuilding directory inode 2037825112

fatal error -- malloc failed in longform_dir2_entry_check (2585598792 bytes)
=================================================
Exists with an malloc failed, even though I am using the -m option
which I think was created to avoid exactly this scenario. Run 3


Run 3 ============================================
./xfs_repair -P -m 200 /dev/evms/monster_evms
... more output ...
      - agno = 38
      - agno = 39
Phase 5 - rebuild AG headers and trees...
      - reset superblock...
Phase 6 - check inode connectivity...
      - resetting contents of realtime bitmap and summary inodes
      - traversing filesystem ...
bad hash table for directory inode 4842 (no data entry): rebuilding
rebuilding directory inode 4842
bad hash table for directory inode 28776251 (no data entry): rebuilding
rebuilding directory inode 28776251
fixing i8count in inode 29111010
corrupt block 0 in directory inode 321685982: junking block
Segmentation fault
=================================================
Exists with a Seg Fault. I found this one interesting especially since
it happens at the exact same inode as Run 1


All comments appreciated.
Simon


[[HTML alternate version deleted]]


<Prev in Thread] Current Thread [Next in Thread>