| To: | Joshua Baker-LePain <jlb17@xxxxxxxx>, Linux xfs mailing list <linux-xfs@xxxxxxxxxxx> |
|---|---|
| Subject: | Re: Performance question |
| From: | Seth Mos <knuffie@xxxxxxxxx> |
| Date: | Wed, 18 Feb 2004 19:41:45 +0100 |
| In-reply-to: | <Pine.LNX.4.58.0402181131210.25541@chaos.egr.duke.edu> |
| Sender: | linux-xfs-bounce@xxxxxxxxxxx |
At 12:57 18-2-2004 -0500, Joshua Baker-LePain wrote:
I've pretty much ruled out hardware. I've swapped the 3ware and rebuilt the array, and the disks all show good SMART data. Have you considered raid 10 configuration? It will about halve the storage space but increase IO by a multiple. You could flip the cover and see if the disks are pushed hard by the controller (leds on the 3ware controller). Although from what you described above it's probably a directory growing a bit on the large side. In narrowing down the problem, it seems that one particular (large) directory is the main culprit. This dir is 471,401,788 KB big and has 3,377,520 files (~140KB/file average). Is the large number of files the entire culprit? If so, is there anything I can do to alleviate the problem? I already 'mount -o logbufs=8'. Here's xfs_info on that partition: Perhaps creating the filessystem with a larger inode size like 512. You could also use version logs instead of version which mostly helped software raid 5 although this might help hardware as well. Cheers -- Seth I don't make sense, I don't pretend to either. Questions? |
| <Prev in Thread] | Current Thread | [Next in Thread> |
|---|---|---|
| ||
| Previous by Date: | [Bug 310] Zeroed files on / partition after clean unmount, bugzilla-daemon |
|---|---|
| Next by Date: | Re: Performance question, Christian Guggenberger |
| Previous by Thread: | Performance question, Joshua Baker-LePain |
| Next by Thread: | Re: Performance question, Christian Guggenberger |
| Indexes: | [Date] [Thread] [Top] [All Lists] |