On 2/26/2012 1:22 AM, MikeJeezy wrote:
> On 02/25/2012 10:35pm, Stan Hoeppner wrote:
>> Can you run xfs_check on the filesystem to determine if a freespace
>> tree is corrupted (post the output if it is), then run xfs_repair
>> to rebuild them?"
> Thank you for responding. This is a 24/7 production server and I did not
> anticipate getting a response this late on a Saturday, so I panicked quite
> frankly, and went ahead and ran "xfs_repair -L" on both volumes. I can now
I wasn't sure how big a pickle you were in so I jumped in and tried to
help best I could.
> mount the volumes and everything looks okay as far as I can tell. There
> were only 2 files in the "lost+found" directory after the repair. Does that
> mean only two files were lost? Is there any way to tell how many files were
I'm not sure. If this is free space btree corruption then you shouldn't
have lost any user files. Others might answer this better than me.
>> This corruption could have happened a long time ago in the past, and
>> it may simply be coincidental that you've tripped over this at
>> roughly the same time you upgraded the kernel.
Note the text above is something I quoted from Dave's 2008 response to
another user with the same problem. In that case he had just upgraded
his kernel and suspected that as the cause. It was not.
> It would be nice to find out why this happened. I suspect it is as you
> suggested, previous corruption and not a hardware issue, because I have
> other volumes mounted to other VM's that are attached to the same SAN
> controller / RAID6 Array... and they did not have any issues - only this one
Are those other VMs using XFS filesystems?
Found this in the list archive:
On 9/19/2011 9:27 AM, Christoph Hellwig wrote:
> Given that before ~2.6.35 LVM/device mapper was not able to pass through
> cache flush requests that is your most likely culprit. A repair will
> rebuild the freespace btrees, and make sure to keep the write caches
> down the whole stack disabled.
What kernel version are you running? Are you using LVM under XFS? What
fstab mount options? Does your SAN array have battery backed write
cache? Are the individual drive caches in the underlying array disabled?
>> So, run "xfs_check /dev/sde1" and post the output here. Then await
>> further instructions.
> Can I still do this (or anything) to help uncover any causes or is it too
If you already ran a repair and it fixed the damage then the check won't
> late? I have also run yum update on the server because it was out of date.
Answering the questions above may lead us to a possible/plausible cause.