On 8/7/2012 12:54 AM, Guk-Bong, Kwon wrote:
> HI all
> I tested xfs over nfs using bonnie++
> xfs and nfs hang when xfs filesystem filled
> What's the problem?
The problem is likely the OP has been given enough rope to hang himself. ;)
> b. lvcreate -L 90G -n test ld1
> data = bsize=4096 blocks=23592960, imaxpct=25
4096*23592960/1048576/1000= ~92GB filesystem
> bonnie++ -s 0 -n 200:1024000:1024000 -r 32G -d /test/ -u 0 -g 0 -q &
((((200*1024=204800 files)*(1024000 bytes))/1048576)/1000)= ~200GB
If my understanding of the bonnie++ options, and my math, are correct,
you are attempting to write 200GB of 1MB files, in parallel, over NFS,
to a 90GB filesystem. Adding insult to injury, you're mounting with
inode32, causing allocation to serialize on AG0, which will cause head
thrashing when alternating between writing directory information and
So, first and foremost, you're attempting to write twice as many bytes
as the filesystem can hold. You're then hamstringing the filesystem's
ability to allocate in parallel. Inode64 would be a better choice here.
You didn't describe the underlying storage hardware, which will likely
have a role to play in the 120 second blocking and the unresponsiveness,
or "hang" as you describe it.
In summary, you're intentionally writing twice the bytes of the FS
capacity. Processes block due to latency, and the FS hangs.
What result were you expecting to see as a result of trying to
intentionally break things?