[Top] [All Lists]

Re: stable xfs

To: Linux XFS <linux-xfs@xxxxxxxxxxx>
Subject: Re: stable xfs
From: pg_xfs@xxxxxxxxxxxxxxxxxx (Peter Grandi)
Date: Wed, 19 Jul 2006 00:14:21 +0100
In-reply-to: <1153262166.2669.267.camel@xxxxxxxxxxxxxxxxxxxxx>
References: <1153150223.4532.24.camel@xxxxxxxxxxxxxxxxxxxxx> <17595.47312.720883.451573@xxxxxxxxxxxxxxxxxx> <1153262166.2669.267.camel@xxxxxxxxxxxxxxxxxxxxx>
Sender: xfs-bounce@xxxxxxxxxxx
>>> On Tue, 18 Jul 2006 18:36:06 -0400, Ming Zhang
>>> <mingz@xxxxxxxxxxx> said:

mingz> [ .. ] example on what is an improper use?

Well, this mailing list is full of them :-). However it is
easier to say what is an optimal use:

  * A 64 bit system.
  * With a large, parallel storage system.
  * The block IO system handles all storage errors.
  * With backups of the contents of the storage system.

In other words, an Altix in an enterprise computing room... :-)

Something like 64 bit systems running a UNIX-like OS, one system
production and one for backup, each with some TiB of RAID10
storage, both with UPSes giving a significant amount of uptime,
and extensive hot swapping abilities. If you got that, XFS can
give really good performance quite safely.

My impression is that the design of XFS was based on a focus on
performance, at the file system level, via on-disk layout,
massive ''transactions'', and parallel IO requests, assuming
that the block IO subsystem handles every storage error issue
both transparently and gracefully.

It is _possible_, and may even be appropriate after carefully
thinking it through, to use XFS in a 32 bit system without UPS,
and with no storage system redundancy, and with device errors
not handled by the block IO system, and with little parallelism
in the storage subsystem; e.g. a SOHO desktop or server.

But then I have seen people building RAIDs stuffing in a couple
dozen drives from the same shipping box, so improper use of XFS
is definitely a second order issue at that kind of level :-).

<Prev in Thread] Current Thread [Next in Thread>