xfs
[Top] [All Lists]

Re: [PATCH 1/3] xfs: build bios directly in xfs_add_to_ioend

To: Christoph Hellwig <hch@xxxxxx>
Subject: Re: [PATCH 1/3] xfs: build bios directly in xfs_add_to_ioend
From: Brian Foster <bfoster@xxxxxxxxxx>
Date: Thu, 3 Mar 2016 10:17:11 -0500
Cc: xfs@xxxxxxxxxxx, Dave Chinner <dchinner@xxxxxxxxxx>
Delivered-to: xfs@xxxxxxxxxxx
In-reply-to: <1456302011-18915-2-git-send-email-hch@xxxxxx>
References: <1456302011-18915-1-git-send-email-hch@xxxxxx> <1456302011-18915-2-git-send-email-hch@xxxxxx>
User-agent: Mutt/1.5.24 (2015-08-30)
On Wed, Feb 24, 2016 at 09:20:09AM +0100, Christoph Hellwig wrote:
> From: Dave Chinner <dchinner@xxxxxxxxxx>
> 
> Currently adding a buffer to the ioend and then building a bio from
> the buffer list are two separate operations. We don't build the bios
> and submit them until the ioend is submitted, and this places a
> fixed dependency on bufferhead chaining in the ioend.
> 
> The first step to removing the bufferhead chaining in the ioend is
> on the IO submission side. We can build the bio directly as we add
> the buffers to the ioend chain, thereby removing the need for a
> latter "buffer-to-bio" submission loop. This allows us to submit
> bios on large ioends as soon as we cannot add more data to the bio.
> 
> These bios then get captured by the active plug, and hence will be
> dispatched as soon as either the plug overflows or we schedule away
> from the writeback context. This will reduce submission latency for
> large IOs, but will also allow more timely request queue based
> writeback blocking when the device becomes congested.
> 
> Signed-off-by: Dave Chinner <dchinner@xxxxxxxxxx>
> ---

Reviewed-by: Brian Foster <bfoster@xxxxxxxxxx>

>  fs/xfs/xfs_aops.c | 117 
> ++++++++++++++++++++++++++----------------------------
>  fs/xfs/xfs_aops.h |   1 +
>  2 files changed, 57 insertions(+), 61 deletions(-)
> 
> diff --git a/fs/xfs/xfs_aops.c b/fs/xfs/xfs_aops.c
> index 7a467b3..90e6e3a 100644
> --- a/fs/xfs/xfs_aops.c
> +++ b/fs/xfs/xfs_aops.c
> @@ -274,6 +274,7 @@ xfs_alloc_ioend(
>       xfs_ioend_t             *ioend;
>  
>       ioend = mempool_alloc(xfs_ioend_pool, GFP_NOFS);
> +     memset(ioend, 0, sizeof(*ioend));
>  
>       /*
>        * Set the count to 1 initially, which will prevent an I/O
> @@ -281,16 +282,9 @@ xfs_alloc_ioend(
>        * all the I/O from calling the completion routine too early.
>        */
>       atomic_set(&ioend->io_remaining, 1);
> -     ioend->io_error = 0;
>       INIT_LIST_HEAD(&ioend->io_list);
>       ioend->io_type = type;
>       ioend->io_inode = inode;
> -     ioend->io_buffer_head = NULL;
> -     ioend->io_buffer_tail = NULL;
> -     ioend->io_offset = 0;
> -     ioend->io_size = 0;
> -     ioend->io_append_trans = NULL;
> -
>       INIT_WORK(&ioend->io_work, xfs_end_io);
>       return ioend;
>  }
> @@ -452,13 +446,18 @@ static inline int xfs_bio_add_buffer(struct bio *bio, 
> struct buffer_head *bh)
>  }
>  
>  /*
> - * Submit all of the bios for an ioend. We are only passed a single ioend at 
> a
> - * time; the caller is responsible for chaining prior to submission.
> + * Submit the bio for an ioend. We are passed an ioend with a bio attached to
> + * it, and we submit that bio. The ioend may be used for multiple bio
> + * submissions, so we only want to allocate an append transaction for the 
> ioend
> + * once. In the case of multiple bio submission, each bio will take an IO
> + * reference to the ioend to ensure that the ioend completion is only done 
> once
> + * all bios have been submitted and the ioend is really done.
>   *
>   * If @fail is non-zero, it means that we have a situation where some part of
>   * the submission process has failed after we have marked paged for writeback
> - * and unlocked them. In this situation, we need to fail the ioend chain 
> rather
> - * than submit it to IO. This typically only happens on a filesystem 
> shutdown.
> + * and unlocked them. In this situation, we need to fail the bio and ioend
> + * rather than submit it to IO. This typically only happens on a filesystem
> + * shutdown.
>   */
>  STATIC int
>  xfs_submit_ioend(
> @@ -466,48 +465,34 @@ xfs_submit_ioend(
>       xfs_ioend_t             *ioend,
>       int                     status)
>  {
> -     struct buffer_head      *bh;
> -     struct bio              *bio;
> -     sector_t                lastblock = 0;
> +     if (!ioend->io_bio || status)
> +             goto error_finish;
>  
>       /* Reserve log space if we might write beyond the on-disk inode size. */
> -     if (!status &&
> -          ioend->io_type != XFS_IO_UNWRITTEN && xfs_ioend_is_append(ioend))
> +     if (ioend->io_type != XFS_IO_UNWRITTEN && xfs_ioend_is_append(ioend) &&
> +         !ioend->io_append_trans) {
>               status = xfs_setfilesize_trans_alloc(ioend);
> -     /*
> -      * If we are failing the IO now, just mark the ioend with an
> -      * error and finish it. This will run IO completion immediately
> -      * as there is only one reference to the ioend at this point in
> -      * time.
> -      */
> -     if (status) {
> -             ioend->io_error = status;
> -             xfs_finish_ioend(ioend);
> -             return status;
> +             if (status)
> +                     goto error_finish;
>       }
>  
> -     bio = NULL;
> -     for (bh = ioend->io_buffer_head; bh; bh = bh->b_private) {
> -
> -             if (!bio) {
> -retry:
> -                     bio = xfs_alloc_ioend_bio(bh);
> -             } else if (bh->b_blocknr != lastblock + 1) {
> -                     xfs_submit_ioend_bio(wbc, ioend, bio);
> -                     goto retry;
> -             }
> -
> -             if (xfs_bio_add_buffer(bio, bh) != bh->b_size) {
> -                     xfs_submit_ioend_bio(wbc, ioend, bio);
> -                     goto retry;
> -             }
> -
> -             lastblock = bh->b_blocknr;
> -     }
> -     if (bio)
> -             xfs_submit_ioend_bio(wbc, ioend, bio);
> +     xfs_submit_ioend_bio(wbc, ioend, ioend->io_bio);
> +     ioend->io_bio = NULL;
>       xfs_finish_ioend(ioend);
>       return 0;
> +
> +     /*
> +      * If we are failing the IO now, just mark the ioend with an error and
> +      * finish it, releasing the active bio if there is one.  This will run
> +      * IO completion immediately as there is only one reference to the ioend
> +      * at this point in time.
> +      */
> +error_finish:
> +     if (ioend->io_bio)
> +             bio_put(ioend->io_bio);
> +     ioend->io_error = status;
> +     xfs_finish_ioend(ioend);
> +     return status;
>  }
>  
>  /*
> @@ -523,27 +508,37 @@ xfs_add_to_ioend(
>       struct buffer_head      *bh,
>       xfs_off_t               offset,
>       struct xfs_writepage_ctx *wpc,
> +     struct writeback_control *wbc,
>       struct list_head        *iolist)
>  {
> -     if (!wpc->ioend || wpc->io_type != wpc->ioend->io_type ||
> -         bh->b_blocknr != wpc->last_block + 1) {
> -             struct xfs_ioend        *new;
> +     struct xfs_ioend        *ioend = wpc->ioend;
>  
> -             if (wpc->ioend)
> -                     list_add(&wpc->ioend->io_list, iolist);
> +     if (!ioend || wpc->io_type != ioend->io_type ||
> +         bh->b_blocknr != wpc->last_block + 1) {
> +             if (ioend)
> +                     list_add(&ioend->io_list, iolist);
>  
> -             new = xfs_alloc_ioend(inode, wpc->io_type);
> -             new->io_offset = offset;
> -             new->io_buffer_head = bh;
> -             new->io_buffer_tail = bh;
> -             wpc->ioend = new;
> +             ioend = wpc->ioend = xfs_alloc_ioend(inode, wpc->io_type);
> +             ioend->io_offset = offset;
> +             ioend->io_buffer_head = bh;
> +             ioend->io_buffer_tail = bh;
>       } else {
> -             wpc->ioend->io_buffer_tail->b_private = bh;
> -             wpc->ioend->io_buffer_tail = bh;
> +             ioend->io_buffer_tail->b_private = bh;
> +             ioend->io_buffer_tail = bh;
>       }
> -
>       bh->b_private = NULL;
> -     wpc->ioend->io_size += bh->b_size;
> +
> +retry:
> +     if (!ioend->io_bio)
> +             ioend->io_bio = xfs_alloc_ioend_bio(bh);
> +
> +     if (xfs_bio_add_buffer(ioend->io_bio, bh) != bh->b_size) {
> +             xfs_submit_ioend_bio(wbc, ioend, ioend->io_bio);
> +             ioend->io_bio = NULL;
> +             goto retry;
> +     }
> +
> +     ioend->io_size += bh->b_size;
>       wpc->last_block = bh->b_blocknr;
>       xfs_start_buffer_writeback(bh);
>  }
> @@ -802,7 +797,7 @@ xfs_writepage_map(
>                       lock_buffer(bh);
>                       if (wpc->io_type != XFS_IO_OVERWRITE)
>                               xfs_map_at_offset(inode, bh, &wpc->imap, 
> offset);
> -                     xfs_add_to_ioend(inode, bh, offset, wpc, &submit_list);
> +                     xfs_add_to_ioend(inode, bh, offset, wpc, wbc, 
> &submit_list);
>                       count++;
>               }
>  
> diff --git a/fs/xfs/xfs_aops.h b/fs/xfs/xfs_aops.h
> index 4e01bd5..c89c3bd 100644
> --- a/fs/xfs/xfs_aops.h
> +++ b/fs/xfs/xfs_aops.h
> @@ -52,6 +52,7 @@ typedef struct xfs_ioend {
>       xfs_off_t               io_offset;      /* offset in the file */
>       struct work_struct      io_work;        /* xfsdatad work queue */
>       struct xfs_trans        *io_append_trans;/* xact. for size update */
> +     struct bio              *io_bio;        /* bio being built */
>  } xfs_ioend_t;
>  
>  extern const struct address_space_operations xfs_address_space_operations;
> -- 
> 2.1.4
> 
> _______________________________________________
> xfs mailing list
> xfs@xxxxxxxxxxx
> http://oss.sgi.com/mailman/listinfo/xfs

<Prev in Thread] Current Thread [Next in Thread>
  • Re: [PATCH 1/3] xfs: build bios directly in xfs_add_to_ioend, Brian Foster <=