I'm having a severe problem with a configuration, but I'm not sure where
this problem lies. Thus, I would like to request from the XFS team, any
known issues regarding kernel 2.4.19, xfs, and 2.4.19-rc5-aa1 patch on
quad P4 Xeon boxen.
I don't know if the issue is XFS, but I don't know it's *not* either.
The server stats right now, are that I have two Dell 6650's with 8GB
ram, RH 7.2 from the xfs 1.1 installer base, with most recent redhat
updates, QLogic 2300Fs atached to ~1TB - 3TB of FC2 storage using LVM
for volume management.
The symptoms are that after about 6 days of uptime with loadavg ~10-12%
most of the time, running oracle8i and some other apps we have, the box
locks.
Usually bad enough that none of the volumes ever get's shut down in any
way, and usually corrupting the xfs log header, so that I have to run
"xfs_repair -L /dev/vg/somevol" to let the volume mount. I keep hitting
the pagebuf.c mount bug which was fixed some time ago, where mount on an
unrepaired file system causes a kernel BUG() at page_buf.c:32(or
something simliar, I can send a ksymoops output if anyone can help).
I'm only curious if there is something in XFS or in the 2.4.19-rc5-aa1
kernel patch that could cause this behaviour, and of course what that
might be.
This is very urgent, so if anyone can mail me directly to help resolve
this, I'm more than happy to take the time to talk and try suggestions.
TIA.
p.s. We've performed all the usual diagnostics, troubleshooting, etc
with all vendors in this situation, but still cannot find a resolution
because when the box crashes, there are no bad log entries, or
noticeable bad messages anywhere.
--
Austin Gonyou <austin@xxxxxxxxxxxxxxx>
Coremetrics, Inc.
signature.asc
Description: This is a digitally signed message part
|