xfs
[Top] [All Lists]

Re: ENOSPC but df and df -i show free space

To: Dave Chinner <david@xxxxxxxxxxxxx>
Subject: Re: ENOSPC but df and df -i show free space
From: Andy Isaacson <adi@xxxxxxxxxxxxx>
Date: Sun, 19 Jun 2011 16:55:18 -0700
Cc: xfs@xxxxxxxxxxx
In-reply-to: <20110619232339.GK561@dastard>
References: <20110619215039.GA17820@xxxxxxxxxxxxx> <20110619221852.GH561@dastard> <20110619225834.GY21778@xxxxxxxxxxxxx> <20110619232339.GK561@dastard>
User-agent: Mutt/1.5.18 (2008-05-17)
On Mon, Jun 20, 2011 at 09:23:39AM +1000, Dave Chinner wrote:
> No allocation algorithm is perfect in all circumstances. The
> alogrithms in XFS tend to degrade when large contiguous freespace
> regions are not available, resulting in more fragmentation of data
> extents and subsequent freespace fragmentation when those files are
> removed or defragmented. The algorithms will recover if you free up
> enough space that large contiguous freespace extents re-form, but
> that can require removing a large amount of data....

Thanks for the background explanation!

> > > > % df -i /d1
> > > > Filesystem            Inodes   IUsed   IFree IUse% Mounted on
> > > > /dev/mapper/vg0-d1   167509008 11806336 155702672    8% /d1
> > > > % sudo xfs_growfs -n /d1
> > > > meta-data=/dev/mapper/vg0-d1     isize=256    agcount=18, 
> > > > agsize=13107200 blks
> > > >          =                       sectsz=512   attr=2
> > > > data     =                       bsize=4096   blocks=235929600, 
> > > > imaxpct=25
> > > >          =                       sunit=0      swidth=0 blks
> > > > naming   =version 2              bsize=4096   ascii-ci=0
> > > > log      =internal               bsize=4096   blocks=25600, version=2
> > > >          =                       sectsz=512   sunit=0 blks, lazy-count=1
> > > > realtime =none                   extsz=4096   blocks=0, rtextents=0
> > > > % grep d1 /proc/mounts
> > > > /dev/mapper/vg0-d1 /d1 xfs rw,relatime,attr2,noquota 0 0
> > > > 
> > > > Obviously I'm missing something, but what?
> > > 
> > > Most likely is that you have no contiguous free space large enough
> > > to create a new inode chunk.  using xfs_db to dump the freespace
> > > size histogram will tell you if this is the case or not.
> > 
> > % sudo xfs_db -c freesp /dev/vg0/d1
> >    from      to extents  blocks    pct
> >       1       1  168504  168504   1.71
> >       2       3     446    1135   0.01
> >       4       7    5550   37145   0.38
> >       8      15   49159  524342   5.33
> >      16      31    1383   29223   0.30
> > 2097152 4194303       1 2931455  29.78
> > 4194304 8388607       1 6150953  62.49
> > 
> > I don't really grok that output.
> 
> It's the historgram of free space extent sizes. You have 168504
> single free block regions (4k in size) in the filesystem, 446
> between 8k and 12k (2-3 blocks), etc.

Ah, OK!  Now it makes sense.

> Inode allocation requires aligned 16k allocations (64x256 byte
> inodes), so you need free extents in the 4-7 block range or larger,
> which you appear to have so it should not be failing. Did you dump
> this histogram while touch was giving ENOSPC errors?

Yes, that was before I grew the filesystem again to get back to
a working state.  I killed all the processes using the filesystem,
unmounted it, and ran xfs_db.

> Also, it might be worthwhile dumping the per-ag histograms (use a
> for loop and the "freesp -a <x>" command) - it may be that certain
> AGs are out of contiguous freespace and that is causing the issue...

I've now grown the filesystem again to get it back into a working state;
it's obviously not "production" per se given my janky configuration, but
it is more convenient if I can create files. :)

It's a shame that we lost the chance to do more debugging though.

> FWIW, you shoul drun "echo 1 > /proc/sys/vm/drop_caches" before
> running the xfsdb comand so that it is not reading stale metadata
> from cache...

I unmounted the filesystem before running xfs_db, so it should be fine.
Didn't even occur to me that it might work to run xfs_db on a block
device that's mounted and active...

I did notice that the unmount command took a minute or two to complete.

-andy

<Prev in Thread] Current Thread [Next in Thread>