<div dir="ltr">A meager non-expert user question with full ignorance of glusterfs: Why are you having I/O errors once every two weeks?<br><div><div><div class="gmail_extra"><br></div><div class="gmail_extra">This looks like XFS behavior I've seen under 2 conditions: 1) when I test XFS on the device-mapper flakey object, using XFS without an external journal, and 2) when I try to press my hard-drive connectors against the motherboard while the PC is still running. Your error message looks more like the result of (2) than of (1).<br>
<br></div><div class="gmail_extra">XFS behavior on flakey is not the best, and I wish it would recover in such situations. In Case (2), I'm fairly sure that the PC is confused on a hardware level because the drive light does not go out. Then again, seeing the behavior of other file systems that fight through the errors, maybe it's for the best. If you're fighting I/O errors, there is no winner, and it's best to get rid of the I/O error.<br>
<br></div><div class="gmail_extra">OK, I'm off the soapbox and will quietly wait for a RAID expert like Dave or Stan to jump in and make me feel like a complete amateur...<br><br>MIchael<br></div><div class="gmail_extra">
<br><div class="gmail_quote">On Tue, Apr 9, 2013 at 9:03 AM, 符永涛 <span dir="ltr"><<a href="mailto:yongtaofu@gmail.com" target="_blank">yongtaofu@gmail.com</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
BTW<br>
xfs_info /dev/sdb<br>
meta-data=/dev/sdb isize=256 agcount=28, agsize=268435440 blks<br>
= sectsz=512 attr=2<br>
data = bsize=4096 blocks=<a href="tel:7324303360" value="+17324303360">7324303360</a>, imaxpct=5<br>
= sunit=16 swidth=160 blks<br>
naming =version 2 bsize=4096 ascii-ci=0<br>
log =internal bsize=4096 blocks=521728, version=2<br>
= sectsz=512 sunit=16 blks, lazy-count=1<br>
realtime =none extsz=4096 blocks=0, rtextents=0<br>
<br>
2013/4/9, 符永涛 <<a href="mailto:yongtaofu@gmail.com">yongtaofu@gmail.com</a>>:<br>
> Dear xfs experts,<br>
> I really need your help sincerely!!! In our production enviroment we<br>
> run glusterfs over top of xfs on Dell x720D(Raid 6). And the xfs file<br>
> system crash on some of the server frequently about every two weeks.<br>
> Can you help to give me a direction about how to debug this issue and<br>
> how to avoid it? Thank you very very much!<br>
><br>
> uname -a<br>
> Linux cqdx.miaoyan.cluster1.node11.qiyi.domain 2.6.32-279.el6.x86_64<br>
> #1 SMP Wed Jun 13 18:24:36 EDT 2012 x86_64 x86_64 x86_64 GNU/Linux<br>
><br>
> Every time the crash log is same, as following<br>
><br>
> 038 Apr 9 09:41:36 cqdx kernel: XFS (sdb): xfs_iunlink_remove:<br>
> xfs_inotobp() returned error 22.<br>
> 1039 Apr 9 09:41:36 cqdx kernel: XFS (sdb): xfs_inactive: xfs_ifree<br>
> returned error 22<br>
> 1040 Apr 9 09:41:36 cqdx kernel: XFS (sdb):<br>
> xfs_do_force_shutdown(0x1) called from line 1184 of file<br>
> fs/xfs/xfs_vnodeops.c. Return address = 0xffffffffa02ee20a<br>
> 1041 Apr 9 09:41:36 cqdx kernel: XFS (sdb): I/O Error Detected.<br>
> Shutting down filesystem<br>
> 1042 Apr 9 09:41:36 cqdx kernel: XFS (sdb): Please umount the<br>
> filesystem and rectify the problem(s)<br>
> 1043 Apr 9 09:41:53 cqdx kernel: XFS (sdb): xfs_log_force: error 5<br>
> returned.<br>
> 1044 Apr 9 09:42:23 cqdx kernel: XFS (sdb): xfs_log_force: error 5<br>
> returned.<br>
> 1045 Apr 9 09:42:53 cqdx kernel: XFS (sdb): xfs_log_force: error 5<br>
> returned.<br>
> 1046 Apr 9 09:43:23 cqdx kernel: XFS (sdb): xfs_log_force: error 5<br>
> returned.<br>
<span class="HOEnZb"><font color="#888888">><br>
> --<br>
> 符永涛<br>
><br>
<br>
<br>
--<br>
符永涛<br>
<br>
_______________________________________________<br>
xfs mailing list<br>
<a href="mailto:xfs@oss.sgi.com">xfs@oss.sgi.com</a><br>
<a href="http://oss.sgi.com/mailman/listinfo/xfs" target="_blank">http://oss.sgi.com/mailman/listinfo/xfs</a><br>
</font></span></blockquote></div><br></div></div></div></div>