xfs
[Top] [All Lists]

Re: [PATCH v2 2/3] xfs: run an eofblocks scan on ENOSPC/EDQUOT

To: Brian Foster <bfoster@xxxxxxxxxx>
Subject: Re: [PATCH v2 2/3] xfs: run an eofblocks scan on ENOSPC/EDQUOT
From: Dave Chinner <david@xxxxxxxxxxxxx>
Date: Tue, 27 May 2014 08:57:55 +1000
Cc: xfs@xxxxxxxxxxx
Delivered-to: xfs@xxxxxxxxxxx
In-reply-to: <1400845950-41435-3-git-send-email-bfoster@xxxxxxxxxx>
References: <1400845950-41435-1-git-send-email-bfoster@xxxxxxxxxx> <1400845950-41435-3-git-send-email-bfoster@xxxxxxxxxx>
User-agent: Mutt/1.5.21 (2010-09-15)
On Fri, May 23, 2014 at 07:52:29AM -0400, Brian Foster wrote:
> Speculative preallocation and and the associated throttling metrics
> assume we're working with large files on large filesystems. Users have
> reported inefficiencies in these mechanisms when we happen to be dealing
> with large files on smaller filesystems. This can occur because while
> prealloc throttling is aggressive under low free space conditions, it is
> not active until we reach 5% free space or less.
> 
> For example, a 40GB filesystem has enough space for several files large
> enough to have multi-GB preallocations at any given time. If those files
> are slow growing, they might reserve preallocation for long periods of
> time as well as avoid the background scanner due to frequent
> modification. If a new file is written under these conditions, said file
> has no access to this already reserved space and premature ENOSPC is
> imminent.
> 
> To handle this scenario, modify the buffered write ENOSPC handling and
> retry sequence to invoke an eofblocks scan. In the smaller filesystem
> scenario, the eofblocks scan resets the usage of preallocation such that
> when the 5% free space threshold is met, throttling effectively takes
> over to provide fair and efficient preallocation until legitimate
> ENOSPC.
> 
> The eofblocks scan is selective based on the nature of the failure. For
> example, an EDQUOT failure in a particular quota will use a filtered
> scan for that quota. Because we don't know which quota might have caused
> an allocation failure at any given time, we run a scan against each
> applicable quota determined to be under low free space conditions.
> 
> Signed-off-by: Brian Foster <bfoster@xxxxxxxxxx>
> ---
>  fs/xfs/xfs_dquot.h  | 15 +++++++++++++++
>  fs/xfs/xfs_file.c   | 23 +++++++++++++++++++----
>  fs/xfs/xfs_icache.c | 47 +++++++++++++++++++++++++++++++++++++++++++++++
>  fs/xfs/xfs_icache.h |  1 +
>  4 files changed, 82 insertions(+), 4 deletions(-)
> 
> diff --git a/fs/xfs/xfs_dquot.h b/fs/xfs/xfs_dquot.h
> index 68a68f7..c24c67e 100644
> --- a/fs/xfs/xfs_dquot.h
> +++ b/fs/xfs/xfs_dquot.h
> @@ -139,6 +139,21 @@ static inline xfs_dquot_t *xfs_inode_dquot(struct 
> xfs_inode *ip, int type)
>       }
>  }
>  
> +/*
> + * Check whether a dquot is under low free space conditions. We assume the 
> quota
> + * is enabled and enforced.
> + */
> +static inline bool xfs_dquot_lowsp(struct xfs_dquot *dqp)
> +{
> +     int64_t freesp;
> +
> +     freesp = be64_to_cpu(dqp->q_core.d_blk_hardlimit) - dqp->q_res_bcount;
> +     if (freesp < dqp->q_low_space[XFS_QLOWSP_1_PCNT])
> +             return true;
> +
> +     return false;
> +}
> +
>  #define XFS_DQ_IS_LOCKED(dqp)        (mutex_is_locked(&((dqp)->q_qlock)))
>  #define XFS_DQ_IS_DIRTY(dqp) ((dqp)->dq_flags & XFS_DQ_DIRTY)
>  #define XFS_QM_ISUDQ(dqp)    ((dqp)->dq_flags & XFS_DQ_USER)
> diff --git a/fs/xfs/xfs_file.c b/fs/xfs/xfs_file.c
> index 1b8160d..2e0e73b 100644
> --- a/fs/xfs/xfs_file.c
> +++ b/fs/xfs/xfs_file.c
> @@ -38,6 +38,7 @@
>  #include "xfs_trace.h"
>  #include "xfs_log.h"
>  #include "xfs_dinode.h"
> +#include "xfs_icache.h"
>  
>  #include <linux/aio.h>
>  #include <linux/dcache.h>
> @@ -741,14 +742,28 @@ write_retry:
>       ret = generic_perform_write(file, &from, pos);
>       if (likely(ret >= 0))
>               iocb->ki_pos = pos + ret;
> +
>       /*
> -      * If we just got an ENOSPC, try to write back all dirty inodes to
> -      * convert delalloc space to free up some of the excess reserved
> -      * metadata space.
> +      * If we hit a space limit, try to free up some lingering preallocated
> +      * space before returning an error. In the case of ENOSPC, first try to
> +      * write back all dirty inodes to free up some of the excess reserved
> +      * metadata space. This reduces the chances that the eofblocks scan
> +      * waits on dirty mappings. Since xfs_flush_inodes() is serialized, this
> +      * also behaves as a filter to prevent too many eofblocks scans from
> +      * running at the same time.
>        */
> -     if (ret == -ENOSPC && !enospc) {
> +     if (ret == -EDQUOT && !enospc) {
> +             enospc = xfs_inode_free_quota_eofblocks(ip);
> +             if (enospc)
> +                     goto write_retry;
> +     } else if (ret == -ENOSPC && !enospc) {
> +             struct xfs_eofblocks eofb = {0};
> +
>               enospc = 1;
>               xfs_flush_inodes(ip->i_mount);
> +             eofb.eof_scan_owner = ip->i_ino; /* for locking */
> +             eofb.eof_flags = XFS_EOF_FLAGS_SYNC;
> +             xfs_icache_free_eofblocks(ip->i_mount, &eofb);
>               goto write_retry;
>       }
>  
> diff --git a/fs/xfs/xfs_icache.c b/fs/xfs/xfs_icache.c
> index f4191f6..3cceb1b 100644
> --- a/fs/xfs/xfs_icache.c
> +++ b/fs/xfs/xfs_icache.c
> @@ -33,6 +33,9 @@
>  #include "xfs_trace.h"
>  #include "xfs_icache.h"
>  #include "xfs_bmap_util.h"
> +#include "xfs_quota.h"
> +#include "xfs_dquot_item.h"
> +#include "xfs_dquot.h"
>  
>  #include <linux/kthread.h>
>  #include <linux/freezer.h>
> @@ -1270,6 +1273,50 @@ xfs_icache_free_eofblocks(
>                                        eofb, XFS_ICI_EOFBLOCKS_TAG);
>  }
>  
> +/*
> + * Run eofblocks scans on the quotas applicable to the inode. For inodes with
> + * multiple quotas, we don't know exactly which quota caused an allocation
> + * failure. We make a best effort by running scans for each quota considered
> + * to be under low free space conditions (less than 1% available free space).
> + */
> +int
> +xfs_inode_free_quota_eofblocks(
> +     struct xfs_inode *ip)
> +{
> +     int scanned = 0;
> +     struct xfs_eofblocks eofb = {0,};
> +     struct xfs_dquot *dq;
> +
> +     ASSERT(xfs_isilocked(ip, XFS_IOLOCK_EXCL));
> +
> +     /* set the scan owner to avoid potential livelock */
> +     eofb.eof_scan_owner = ip->i_ino;
> +
> +     if (XFS_IS_UQUOTA_ENFORCED(ip->i_mount)) {
> +             dq = xfs_inode_dquot(ip, XFS_DQ_USER);
> +             if (dq && xfs_dquot_lowsp(dq)) {
> +                     eofb.eof_uid = VFS_I(ip)->i_uid;
> +                     eofb.eof_flags = XFS_EOF_FLAGS_SYNC|
> +                                      XFS_EOF_FLAGS_UID;
> +                     xfs_icache_free_eofblocks(ip->i_mount, &eofb);
> +                     scanned = 1;
> +             }
> +     }
> +
> +     if (XFS_IS_GQUOTA_ENFORCED(ip->i_mount)) {
> +             dq = xfs_inode_dquot(ip, XFS_DQ_GROUP);
> +             if (dq && xfs_dquot_lowsp(dq)) {
> +                     eofb.eof_gid = VFS_I(ip)->i_gid;
> +                     eofb.eof_flags = XFS_EOF_FLAGS_SYNC|
> +                                      XFS_EOF_FLAGS_GID;
> +                     xfs_icache_free_eofblocks(ip->i_mount, &eofb);
> +                     scanned = 1;
> +             }
> +     }

Rather that doing two scans here, wouldn't it be more efficient
to do:

        eofb.eof_flags = XFS_EOF_FLAGS_SYNC;
        scan = false;
        if (uquota is low) {
                eofb.eof_uid = VFS_I(ip)->i_uid;
                eofb.eof_flags |= XFS_EOF_FLAGS_UID;
                scan = true;
        }
        if (gquota is low) {
                eofb.eof_gid = VFS_I(ip)->i_gid;
                eofb.eof_flags |= XFS_EOF_FLAGS_GID;
                scan = true;
        }
        if (scan)
                xfs_icache_free_eofblocks(ip->i_mount, &eofb);

and change xfs_inode_match_id() to be able to check against multiple
flags on a single inode? That way we only scan the inode cache
once, regardless of the number of quota types that are enabled and
are tracking low space thresholds.

Cheers,

Dave.
-- 
Dave Chinner
david@xxxxxxxxxxxxx

<Prev in Thread] Current Thread [Next in Thread>