| To: | xfs@xxxxxxxxxxx |
|---|---|
| Subject: | High Fragmentation with XFS and NFS Sync |
| From: | Nick Fisk <friskyfisk10@xxxxxxxxxxxxxx> |
| Date: | Sat, 2 Jul 2016 09:52:40 +0100 |
| Delivered-to: | xfs@xxxxxxxxxxx |
| Dkim-signature: | v=1; a=rsa-sha256; c=relaxed/relaxed; d=googlemail.com; s=20120113; h=mime-version:from:date:message-id:subject:to; bh=GsZdNL9XWkRl7jD1SOUX9qNTF85vpJx0avs5rBtHTR4=; b=0rz/rPUYCShYiR3Es8VTIm9GB8ovlb8+itDmFqDJJxE28BRxYQKoa3QUfCR8CpFSWN LId7+cdmXjCO9Ce04KIzmQDoY7L6OI+BnjcSdQ+7tnrpnkb/ktkIF7M5B4giOr+gqvDf ZPteM0c4GAuPeVAp5uLlENs9QdgGRC883Oa0LwSkySdcY++PgwHRkxwb5LIqRJ72BT38 qa6VUHToiH1heKaqWfRERBVerhR6ahlbKlVqZF0eO4v3DLuQHOmbJldtHdlumDmgNPAb fQULVxo+FmINVD0sfQCq+ohyBJIyZVNQ3z5tL7/yiOwKmT5JdTJRRVyZ05A9EoZz6w9t LAtA== |
|
Hi, hope someone can help me here.
I'm exporting some XFS fs's to ESX via NFS with the sync option enabled. I'm seeing really heavy fragmentation when multiple VM's are copied onto the share at the same time. I'm also seeing kmem_alloc failures, which is probably the biggest problem as this effectively takes everything down. Underlying storage is a Ceph RBD, the server the FS is running on, is running kernel 4.5.7. Mount options are currently default. I'm seeing Millions of extents, where the ideal is listed as a couple of thousand when running xfs_db, there is only a couple of 100 files on the FS. It looks like roughly the extent sizes roughly match the IO size that the VM's were written to XFS with. So it looks like each parallel IO thread is being allocated next to each other rather than at spaced out regions of the disk. From what I understand, this is because each NFS write opens and closes the file which throws off any chance that XFS will be able to use its allocation features to stop parallel write streams from interleaving with each other. Is there anything I can tune to try and give each write to each file a little bit of space, so that it at least gives readahead a chance when reading, that it might hit at least a few MB of sequential data? I have read that inode32 allocates more randomly compared to inode64, so I'm not sure if it's worth trying this as there will likely be less than a 1000 files per FS. Or am I best just to run fsr after everything has been copied on? Thanks for any advice Nick |
| <Prev in Thread] | Current Thread | [Next in Thread> |
|---|---|---|
| ||
| Previous by Date: | Re: [PATCH] xfs: add readahead bufs to lru early to prevent post-unmount panic, Brian Foster |
|---|---|
| Next by Date: | Re: High Fragmentation with XFS and NFS Sync, Richard Scobie |
| Previous by Thread: | [PATCH] xfs_io: implement 'copy_range' command, Anna Schumaker |
| Next by Thread: | Re: High Fragmentation with XFS and NFS Sync, Darrick J. Wong |
| Indexes: | [Date] [Thread] [Top] [All Lists] |