On Wed, Feb 24, 2016 at 09:20:09AM +0100, Christoph Hellwig wrote:
> From: Dave Chinner <dchinner@xxxxxxxxxx>
>
> Currently adding a buffer to the ioend and then building a bio from
> the buffer list are two separate operations. We don't build the bios
> and submit them until the ioend is submitted, and this places a
> fixed dependency on bufferhead chaining in the ioend.
>
> The first step to removing the bufferhead chaining in the ioend is
> on the IO submission side. We can build the bio directly as we add
> the buffers to the ioend chain, thereby removing the need for a
> latter "buffer-to-bio" submission loop. This allows us to submit
> bios on large ioends as soon as we cannot add more data to the bio.
>
> These bios then get captured by the active plug, and hence will be
> dispatched as soon as either the plug overflows or we schedule away
> from the writeback context. This will reduce submission latency for
> large IOs, but will also allow more timely request queue based
> writeback blocking when the device becomes congested.
>
> Signed-off-by: Dave Chinner <dchinner@xxxxxxxxxx>
> ---
Reviewed-by: Brian Foster <bfoster@xxxxxxxxxx>
> fs/xfs/xfs_aops.c | 117
> ++++++++++++++++++++++++++----------------------------
> fs/xfs/xfs_aops.h | 1 +
> 2 files changed, 57 insertions(+), 61 deletions(-)
>
> diff --git a/fs/xfs/xfs_aops.c b/fs/xfs/xfs_aops.c
> index 7a467b3..90e6e3a 100644
> --- a/fs/xfs/xfs_aops.c
> +++ b/fs/xfs/xfs_aops.c
> @@ -274,6 +274,7 @@ xfs_alloc_ioend(
> xfs_ioend_t *ioend;
>
> ioend = mempool_alloc(xfs_ioend_pool, GFP_NOFS);
> + memset(ioend, 0, sizeof(*ioend));
>
> /*
> * Set the count to 1 initially, which will prevent an I/O
> @@ -281,16 +282,9 @@ xfs_alloc_ioend(
> * all the I/O from calling the completion routine too early.
> */
> atomic_set(&ioend->io_remaining, 1);
> - ioend->io_error = 0;
> INIT_LIST_HEAD(&ioend->io_list);
> ioend->io_type = type;
> ioend->io_inode = inode;
> - ioend->io_buffer_head = NULL;
> - ioend->io_buffer_tail = NULL;
> - ioend->io_offset = 0;
> - ioend->io_size = 0;
> - ioend->io_append_trans = NULL;
> -
> INIT_WORK(&ioend->io_work, xfs_end_io);
> return ioend;
> }
> @@ -452,13 +446,18 @@ static inline int xfs_bio_add_buffer(struct bio *bio,
> struct buffer_head *bh)
> }
>
> /*
> - * Submit all of the bios for an ioend. We are only passed a single ioend at
> a
> - * time; the caller is responsible for chaining prior to submission.
> + * Submit the bio for an ioend. We are passed an ioend with a bio attached to
> + * it, and we submit that bio. The ioend may be used for multiple bio
> + * submissions, so we only want to allocate an append transaction for the
> ioend
> + * once. In the case of multiple bio submission, each bio will take an IO
> + * reference to the ioend to ensure that the ioend completion is only done
> once
> + * all bios have been submitted and the ioend is really done.
> *
> * If @fail is non-zero, it means that we have a situation where some part of
> * the submission process has failed after we have marked paged for writeback
> - * and unlocked them. In this situation, we need to fail the ioend chain
> rather
> - * than submit it to IO. This typically only happens on a filesystem
> shutdown.
> + * and unlocked them. In this situation, we need to fail the bio and ioend
> + * rather than submit it to IO. This typically only happens on a filesystem
> + * shutdown.
> */
> STATIC int
> xfs_submit_ioend(
> @@ -466,48 +465,34 @@ xfs_submit_ioend(
> xfs_ioend_t *ioend,
> int status)
> {
> - struct buffer_head *bh;
> - struct bio *bio;
> - sector_t lastblock = 0;
> + if (!ioend->io_bio || status)
> + goto error_finish;
>
> /* Reserve log space if we might write beyond the on-disk inode size. */
> - if (!status &&
> - ioend->io_type != XFS_IO_UNWRITTEN && xfs_ioend_is_append(ioend))
> + if (ioend->io_type != XFS_IO_UNWRITTEN && xfs_ioend_is_append(ioend) &&
> + !ioend->io_append_trans) {
> status = xfs_setfilesize_trans_alloc(ioend);
> - /*
> - * If we are failing the IO now, just mark the ioend with an
> - * error and finish it. This will run IO completion immediately
> - * as there is only one reference to the ioend at this point in
> - * time.
> - */
> - if (status) {
> - ioend->io_error = status;
> - xfs_finish_ioend(ioend);
> - return status;
> + if (status)
> + goto error_finish;
> }
>
> - bio = NULL;
> - for (bh = ioend->io_buffer_head; bh; bh = bh->b_private) {
> -
> - if (!bio) {
> -retry:
> - bio = xfs_alloc_ioend_bio(bh);
> - } else if (bh->b_blocknr != lastblock + 1) {
> - xfs_submit_ioend_bio(wbc, ioend, bio);
> - goto retry;
> - }
> -
> - if (xfs_bio_add_buffer(bio, bh) != bh->b_size) {
> - xfs_submit_ioend_bio(wbc, ioend, bio);
> - goto retry;
> - }
> -
> - lastblock = bh->b_blocknr;
> - }
> - if (bio)
> - xfs_submit_ioend_bio(wbc, ioend, bio);
> + xfs_submit_ioend_bio(wbc, ioend, ioend->io_bio);
> + ioend->io_bio = NULL;
> xfs_finish_ioend(ioend);
> return 0;
> +
> + /*
> + * If we are failing the IO now, just mark the ioend with an error and
> + * finish it, releasing the active bio if there is one. This will run
> + * IO completion immediately as there is only one reference to the ioend
> + * at this point in time.
> + */
> +error_finish:
> + if (ioend->io_bio)
> + bio_put(ioend->io_bio);
> + ioend->io_error = status;
> + xfs_finish_ioend(ioend);
> + return status;
> }
>
> /*
> @@ -523,27 +508,37 @@ xfs_add_to_ioend(
> struct buffer_head *bh,
> xfs_off_t offset,
> struct xfs_writepage_ctx *wpc,
> + struct writeback_control *wbc,
> struct list_head *iolist)
> {
> - if (!wpc->ioend || wpc->io_type != wpc->ioend->io_type ||
> - bh->b_blocknr != wpc->last_block + 1) {
> - struct xfs_ioend *new;
> + struct xfs_ioend *ioend = wpc->ioend;
>
> - if (wpc->ioend)
> - list_add(&wpc->ioend->io_list, iolist);
> + if (!ioend || wpc->io_type != ioend->io_type ||
> + bh->b_blocknr != wpc->last_block + 1) {
> + if (ioend)
> + list_add(&ioend->io_list, iolist);
>
> - new = xfs_alloc_ioend(inode, wpc->io_type);
> - new->io_offset = offset;
> - new->io_buffer_head = bh;
> - new->io_buffer_tail = bh;
> - wpc->ioend = new;
> + ioend = wpc->ioend = xfs_alloc_ioend(inode, wpc->io_type);
> + ioend->io_offset = offset;
> + ioend->io_buffer_head = bh;
> + ioend->io_buffer_tail = bh;
> } else {
> - wpc->ioend->io_buffer_tail->b_private = bh;
> - wpc->ioend->io_buffer_tail = bh;
> + ioend->io_buffer_tail->b_private = bh;
> + ioend->io_buffer_tail = bh;
> }
> -
> bh->b_private = NULL;
> - wpc->ioend->io_size += bh->b_size;
> +
> +retry:
> + if (!ioend->io_bio)
> + ioend->io_bio = xfs_alloc_ioend_bio(bh);
> +
> + if (xfs_bio_add_buffer(ioend->io_bio, bh) != bh->b_size) {
> + xfs_submit_ioend_bio(wbc, ioend, ioend->io_bio);
> + ioend->io_bio = NULL;
> + goto retry;
> + }
> +
> + ioend->io_size += bh->b_size;
> wpc->last_block = bh->b_blocknr;
> xfs_start_buffer_writeback(bh);
> }
> @@ -802,7 +797,7 @@ xfs_writepage_map(
> lock_buffer(bh);
> if (wpc->io_type != XFS_IO_OVERWRITE)
> xfs_map_at_offset(inode, bh, &wpc->imap,
> offset);
> - xfs_add_to_ioend(inode, bh, offset, wpc, &submit_list);
> + xfs_add_to_ioend(inode, bh, offset, wpc, wbc,
> &submit_list);
> count++;
> }
>
> diff --git a/fs/xfs/xfs_aops.h b/fs/xfs/xfs_aops.h
> index 4e01bd5..c89c3bd 100644
> --- a/fs/xfs/xfs_aops.h
> +++ b/fs/xfs/xfs_aops.h
> @@ -52,6 +52,7 @@ typedef struct xfs_ioend {
> xfs_off_t io_offset; /* offset in the file */
> struct work_struct io_work; /* xfsdatad work queue */
> struct xfs_trans *io_append_trans;/* xact. for size update */
> + struct bio *io_bio; /* bio being built */
> } xfs_ioend_t;
>
> extern const struct address_space_operations xfs_address_space_operations;
> --
> 2.1.4
>
> _______________________________________________
> xfs mailing list
> xfs@xxxxxxxxxxx
> http://oss.sgi.com/mailman/listinfo/xfs
|