xfs
[Top] [All Lists]

Re: XFS crashed twice, once in 2.6.16.20, next in 2.6.17, reproducable

To: Adam Sj?gren <asjo@xxxxxxxxxxxx>
Subject: Re: XFS crashed twice, once in 2.6.16.20, next in 2.6.17, reproducable
From: Nathan Scott <nathans@xxxxxxx>
Date: Tue, 4 Jul 2006 11:47:21 +1000
Cc: linux-xfs@xxxxxxxxxxx
In-reply-to: <87veqhfmo4.fsf@topper.koldfront.dk>; from asjo@koldfront.dk on Sat, Jul 01, 2006 at 04:02:35PM +0200
References: <3aa654a40606190044q43dca571qdc06ee13d82d979@mail.gmail.com> <20060620161006.C1079661@wobbly.melbourne.sgi.com> <3aa654a40606192338v751150fp5645d1d2943316ea@mail.gmail.com> <20060620164338.A1080488@wobbly.melbourne.sgi.com> <3aa654a40606192350w5c469670t466dfc1344e23a4@mail.gmail.com> <20060620165209.C1080488@wobbly.melbourne.sgi.com> <3aa654a40606200120v5baf0304ka205f1ad8f136ad9@mail.gmail.com> <20060622125640.C1135236__28424.0987770774$1150945513$gmane$org@wobbly.melbourne.sgi.com> <87veqhfmo4.fsf@topper.koldfront.dk>
Sender: xfs-bounce@xxxxxxxxxxx
User-agent: Mutt/1.2.5i
On Sat, Jul 01, 2006 at 04:02:35PM +0200, Adam Sj?gren wrote:
> On Thu, 22 Jun 2006 12:56:40 +1000, Nathan wrote:
> The subsequent xfs_repair run gave different, expected I guess, output
> (it noticed the modified directory inode and stuff), but at the end of
> Phase 7, it said:
> 
>  cache_purge: share on cache 0x100930b0 left 1 nodes!?
>  cache_purge: share on cache 0x100930b0 left 1 nodes!?
>  done

Thats some diagnostic stuff for us developers.  Don't worry about
it, its just telling us we have inode refcounting issues which we
will have to resolve for a parallel xfs_repair.

> ...
> I proceeded to run xfs_repair. In Phase 3 under '- agno 0' it said:
> 
>  bad dir magic number 0x30 in inode 180 bno = 8388608
> 
> It continued and in Phase 6 after saying '- traversing filesystem
> starting at / ...' it said:
> 
>  rebuilding directory inode 128

Thats your root directory.  Its being rebuilt due to a lost+found
directory being earlier unlinked, most likely.

> And a flurry of "disconnected inode [number], moving to lost+found"
> followed.

Yeah, anything that was in lost+found after the first run, gets
put back there each time.

> Any hints on how to fully repair the filesystem, in place?

It sounds like your filesystem is repaired at this point - you
need to fix up (remove) lost+found, and you should get no issues
reported thereafter.

cheers.

-- 
Nathan


<Prev in Thread] Current Thread [Next in Thread>