I'm using 2.4.19-rc5-aa1 patch on a standard 2.4.19 kernel with ~1TB of
disks FC2 disks. (14*69.87GB)
I'm using LVM 1.0.3 and associated tools, as well as XFS as my FS of
choice. I'm not sure what's happened recently, but after upgrading my
disk subsystem to full FC2 (we were just FC2 to the switches before),
when mounting any of the volumes on the StorageTek D178 subsystem, I get
the following bug message.
I was curious if I was hitting an XFS bug, or if it's because of the
many LIP resets I get from the Qlogic Driver and the kernel got
overflowed or XFS's mount got overflowed or something. Not sure, but it
doesn't sound good. TIA
scsi1: Topology - (FL_Port), Host Loop address 0x0
Starting XFS recovery on filesystem: lvm(58,0) (dev: 58/0)
kernel BUG at page_buf.c:602!
invalid operand: 0000 2.4.19 #2 SMP Sun Aug 25 23:33:55 CDT 2002
CPU: 4
EIP: 0010:[<c01ff6e4>] Not tainted
EFLAGS: 00010246
eax: 00000000 ebx: 00000002 ecx: 00080221 edx: 00000000
esi: c92d5180 edi: 00000001 ebp: 00000000 esp: c92e5720
ds: 0018 es: 0018 ss: 0018
Process mount (pid: 1381, stackpage=c92e5000)
Stack: c92d51e0 00000002 000001f0 c78eb28c c92e4000 00000000 00000000
000001f0
00000000 00000001 0010000c 00000002 00000000 00000002 00800050
00000000
00002205 00000000 c01ffc76 c92d5180 c92e9934 00002205 00002000
00002205
Call Trace: [<c01ffc76>] [<c01e68c7>] [<c01e7854>] [<c01e5d97>]
[<c01e792f>] [<c01e7a3e>] [<c01e8544>] [<c01e89d3>] [<c01e8a3b>]
[<c01e8c53>] [<c01e13b4>] [<c01ea0bc>] [<c0201a3c>] [<c01e95b4>]
[<c01e980b>] [<c01f36be>] [<c01f392f>] [<c01f3977>] [<c020c753>]
[<c013ed95>] [<c01352b5>] [<c013d02f>] [<c012f57d>] [<c013d02f>]
[<c014dd2a>] [<c014e7ac>] [<c014dc7b>] [<c014eb8c>] [<c01652e8>]
[<c0165780>] [<c01655c9>] [<c0165d3f>] [<c01095bb>]
Code: 0f 0b 5a 02 a0 4e 2f c0 8b 4c 24 4c 8b 51 58 81 49 08 04 00
--
Austin Gonyou <austin@xxxxxxxxxxxxxxx>
Coremetrics, Inc.
|