XFS File system in trouble
Leslie Rhorer
lrhorer at mygrande.net
Wed Jul 22 20:45:07 CDT 2015
On 7/20/2015 6:17 AM, Brian Foster wrote:
> On Sat, Jul 18, 2015 at 08:02:50PM -0500, Leslie Rhorer wrote:
>>
>> I found the problem with md5sum (and probably nfs, as well). One of the
>> memory modules in the server was bad. The problem with XFS persists. Every
>> time tar tried to create the directory:
>>
>> /RAID/Server-Main/Equipment/Drive Controllers/HighPoint Adapters/Rocket 2722/Driver/RR276x/Driver/Linux/openSUSE/rr276x-suse-11.2-i386/linux/suse/i386-11.1
>>
>> It would begin spitting out errors, starting with "Cannot mkdir: Structure
>> needs cleaning". At that point, XFS had shut down. I went into
>> /RAID/Server-Main/Equipment/Drive Controllers/HighPoint Adapters/Rocket
>> 2722/Driver/RR276x/Driver/Linux/openSUSE/rr276x-suse-11.2-i386/linux/suse/
>> and created the i386-11.1 directory by hand, and tar no longer starts
>> spitting out errors at that point, but it does start up again at
>> RR2782/Windows/Vista-Win2008-Win7-legacy_single/x64.
>>
>
> So is this untar problem a reliable reproducer? If so, here's what I
Absolutely reliable producer. The only change is if I create the
offending directory by hand (after recovering the filesystem, of course)
and then start the tar again. Then it copies all the files into the
previously offending directory, failing the next time it tries to create
a directory.
> would try to hopefully isolate a filesystem problem from something
> underneath:
OK. Frankly, I fail to find it at all likely to be anything above. I
can read and write 100s of megabytes of data without an error. The only
thing that I can find failing is creating directories, and that is only
when tar attempts it. The directory structure is going to be written to
different inodes as time goes by, so a failure of mdadm or some
structure above it should cause other widesperead issues. I need to try
some other tarballs when I get the chance, and also try dumping that tar
on a different directory.
> xfs_metadump -go /dev/md0 /somewhere/on/rootfs/md0.metadump
> xfs_mdrestore -g /somewhere/on/rootfs/md0.metadump /.../fileonrootfs.img
How big are those files going to be, do you think? The root partition
is not all that huge. There is only a little over 80G free.
> mount /.../fileonrootfs.img /mnt/
>
> ... and repeat the test on that mount using the original tarball (if
> it's on the associated fs, the version from the dump will have no data).
It is. I've tried copying it to another fs, and it works fine, there.
> This will create a metadata only dump of the original fs onto another
> storage device (e.g., whatever holds the root fs), restore the metadump
> to a file and mount it loopback. The resulting fs will not contain any
> file data, but will contain all of the metadata such as directory
> structure, etc. and is otherwise mountable and usable for experimental
> purposes.
>
> If the problem is in the filesystem or "above" (as in kernel, memory
> issue, etc.), the test should fail on this mount. If the problem is
> beneath the fs such as somewhere in the storage stack (assuming the
> rootfs storage stack is reliable), it probably shouldn't fail.
I'll look into this when I can. Right now I have some critical
operations going on both servers (primary and backup), and I can't take
down a file system or even risk doing so. Hopefully I will get around
to it this weekend.
More information about the xfs
mailing list