XFS File system in trouble

Leslie Rhorer lrhorer at mygrande.net
Wed Jul 22 20:45:07 CDT 2015


On 7/20/2015 6:17 AM, Brian Foster wrote:
> On Sat, Jul 18, 2015 at 08:02:50PM -0500, Leslie Rhorer wrote:
>>
>> 	I found the problem with md5sum (and probably nfs, as well).  One of the
>> memory modules in the server was bad.  The problem with XFS persists.  Every
>> time tar tried to create the directory:
>>
>> /RAID/Server-Main/Equipment/Drive Controllers/HighPoint Adapters/Rocket 2722/Driver/RR276x/Driver/Linux/openSUSE/rr276x-suse-11.2-i386/linux/suse/i386-11.1
>>
>> 	It would begin spitting out errors, starting with "Cannot mkdir: Structure
>> needs cleaning".  At that point, XFS had shut down.  I went into
>> /RAID/Server-Main/Equipment/Drive Controllers/HighPoint Adapters/Rocket
>> 2722/Driver/RR276x/Driver/Linux/openSUSE/rr276x-suse-11.2-i386/linux/suse/
>> and created the i386-11.1 directory by hand, and tar no longer starts
>> spitting out errors at that point, but it does start up again at
>> RR2782/Windows/Vista-Win2008-Win7-legacy_single/x64.
>>
>
> So is this untar problem a reliable reproducer? If so, here's what I

	Absolutely reliable producer.  The only change is if I create the 
offending directory by hand (after recovering the filesystem, of course) 
and then start the tar again.  Then it copies all the files into the 
previously offending directory, failing the next time it tries to create 
a directory.

> would try to hopefully isolate a filesystem problem from something
> underneath:

	OK.  Frankly, I fail to find it at all likely to be anything above.  I 
can read and write 100s of megabytes of data without an error.  The only 
thing that I can find failing is creating directories, and that is only 
when tar attempts it.  The directory structure is going to be written to 
different inodes as time goes by, so a failure of mdadm or some 
structure above it should cause other widesperead issues.  I need to try 
some other tarballs when I get the chance, and also try dumping that tar 
on a different directory.


> xfs_metadump -go /dev/md0 /somewhere/on/rootfs/md0.metadump
> xfs_mdrestore -g /somewhere/on/rootfs/md0.metadump /.../fileonrootfs.img

	How big are those files going to be, do you think?  The root partition 
is not all that huge.  There is only a little over 80G free.

> mount /.../fileonrootfs.img /mnt/
>
> ... and repeat the test on that mount using the original tarball (if
> it's on the associated fs, the version from the dump will have no data).

	It is.  I've tried copying it to another fs, and it works fine, there.

> This will create a metadata only dump of the original fs onto another
> storage device (e.g., whatever holds the root fs), restore the metadump
> to a file and mount it loopback. The resulting fs will not contain any
> file data, but will contain all of the metadata such as directory
> structure, etc. and is otherwise mountable and usable for experimental
> purposes.
>
> If the problem is in the filesystem or "above" (as in kernel, memory
> issue, etc.), the test should fail on this mount. If the problem is
> beneath the fs such as somewhere in the storage stack (assuming the
> rootfs storage stack is reliable), it probably shouldn't fail.

	I'll look into this when I can.  Right now I have some critical 
operations going on both servers (primary and backup), and I can't take 
down a file system or even risk doing so.  Hopefully I will get around 
to it this weekend.



More information about the xfs mailing list