xfs
[Top] [All Lists]

Need advice on building a new XFS setup for large files

To: xfs@xxxxxxxxxxx
Subject: Need advice on building a new XFS setup for large files
From: Alvin Ong <alvin.ong@xxxxxxxxxxxxxxxxx>
Date: Tue, 22 Jan 2013 12:22:42 +0800
Delivered-to: xfs@xxxxxxxxxxx
Hi,

We are building a solution with a web front end for our users to store large files.
Large files starting from the size of 500GB and above and can grow up to 1-2TB's per file.
This is the reason we are trying out XFS to see if we can get a test system running.

We plan to use a 6+2 RAID6 to start off with. Then when it gets filled up to maybe 60-70% we will
expand by adding another 6+2 RAID6 to the array.
The max we can grow this configuration is up to 252TB usable which should be enough for a year.
Our requirements might grow up to 2PB in 2 years time if all goes well.

So I have been testing all of this out on a VM running 3 vmdk's and using LVM to create a single logical volume of the 3 disks.
I noticed that out of sdb, sdc and sdd, files keep getting written to sdc.
This is probably due to our web app creating a single folder and all files are written under that folder.
This is the nature of the Allocation Group of XFS? Is there a way to avoid this? As we will have files keep writing to the same disk thus creating a hot spot.
Although it might not hurt us that much if we fill up a single RAID6 to 60-70% then adding another RAID6 to the mix. We could go up to a total of 14 RAID6 sets.

Is LVM a good choice of doing this configuration? Or do you have a better recommendation?
The reason we thought LVM would be good was so that we could easily grow XFS.

Here is some info:

# xfs_info /xfs
meta-data="" isize=256    agcount=4, agsize=32768000 blks
         =                       sectsz=512   attr=2
data     =                       bsize=4096   blocks=131072000, imaxpct=25
         =                       sunit=0      swidth=0 blks
naming   =version 2              bsize=4096   ascii-ci=0
log      =external               bsize=4096   blocks=65536, version=2
         =                       sectsz=512   sunit=0 blks, lazy-count=1
realtime =none                   extsz=4096   blocks=0, rtextents=0


Mount options:
/dev/mapper/vg_xfs-lv_xfs on /xfs type xfs (rw,noatime,nodiratime,logdev=/dev/vg_xfs/lv_log_xfs,nobarrier,inode64,logbsize=262144,allocsize=512m)

Is I was to use the 8-disk RAID6 array with a 256kB stripe size will have a sunit of 512 and a swidth of (8-2)*512=3072. 
# mkfs.xfs -d sunit=512,swidth=3072 /dev/mapper/vg_xfs-lv_xfs
# mount -o remount,sunit=512,swidth=3072
Correct?

What about the logdev option? What is the optimal size to create for it?

Hope to get some guidance from you guru's please.

Thank you.
Alvin



<Prev in Thread] Current Thread [Next in Thread>