Xfs_repair segfaults.

Filippo Stenico filippo.stenico at gmail.com
Tue May 7 08:36:42 CDT 2013


On Tue, May 7, 2013 at 3:20 PM, Eric Sandeen <sandeen at sandeen.net> wrote:

> On 5/7/13 4:27 AM, Filippo Stenico wrote:
> > Hello,
> > this is a start-over to try hard to recover some more data out of my
> raid5 - lvm - xfs toasted volume.
> > My goal is either to try the best to get some more data out of the
> volume, and see if I can reproduce the segfault.
> > I compiled xfsprogs 3.1.9 from deb-source. I ran a xfs_metarestore to
> put original metadata on the cloned raid volume i had zeroed the log on
> before via xfs_repair -L (i figured none of the actual data was modified
> before as I am just working on metadata.. right?).
> > Then I ran a mount, checked a dir that I knew it was corrupted, unmount
> and try an xfs_repair (commands.txt attached for details)
> > I went home to sleep, but at morning I found out that kernel paniced due
> "out of memory and no killable process".
> > I ran repair without -P... Should I try now disabling inode prefetch?
> > Attached are also output of "free" and "top" at time of panic, as well
> as the output of xfs_repair and strace attached to it. Dont think gdb
> symbols would help here....
> >
>
> >
>
> Ho hum, well, no segfault this time, just an out of memory error?
>
That's right....

> No real way to know where it went from the available data I think.
>
> A few things:
>
> > root at ws1000:~# mount /dev/mapper/vg0-lv0 /raid0/data/
> > mount: Structure needs cleaning
>
> mount failed?  Now's the time to look at dmesg to see why.
>

>From attached logs it seems to be:
>
> > XFS internal error xlog_valid_rec_header(1) at line 3466 of file
> [...2.6.32...]/fs/xfs/xfs_log_recover.c
> > XFS: log mount/recovery failed: error 117
>
> > root at ws1000:~# mount
>
> <no raid0 mounted>
>
> > root at ws1000:~# mount /dev/mapper/vg0-lv0 /raid0/data/
> > root at ws1000:~# mount | grep raid0
> > /dev/mapper/vg0-lv0 on /raid0/data type xfs (rw,relatime,attr2,noquota)
>
> Uh, now it worked, with no other steps in between?  That's a little odd.
>
Looks odd to me too. But i just copied the commands issued as they where on
my console... so yes, nothing in between.

> It found a clean log this time:
>
> > XFS mounting filesystem dm-1
> > Ending clean XFS mount for filesystem: dm-1
>
> which is unexpected.
>
> So the memory consumption might be a bug but there's not enough info to go
> on here.
>
> > PS. Let me know if you wish reports like this one on list.
>
> worth reporting, but I'm not sure what we can do with it.
> Your storage is in pretty bad shape, and xfs_repair can't make something
> out
> of nothing.
>
> -Eric
>

I still got back around 6TB out of 7.2 TB of total data stored, so this
tells xfs is reliable even when major faults occur...

Thanks anyways, I am trying with a "-L" repair, at this step I expect
another fail (due out of memory or something, as it happened last time)
then I will try with "xfs_repair -L -vv -P" and I expect to see that
segfault again.

Will report next steps, maybe something interesting for you will pop up...
for me is not a waste of time, since this last try is worth being made.

-- 
F
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://oss.sgi.com/pipermail/xfs/attachments/20130507/8c02aeb4/attachment.html>


More information about the xfs mailing list