[Top] [All Lists]

xfs trouble

To: linux-xfs@xxxxxxxxxxx
Subject: xfs trouble
From: Thomas Graichen <news-innominate.list.sgi.xfs@xxxxxxxxxxxxx>
Date: 18 Aug 2000 19:33:57 GMT
Distribution: local
Organization: innominate AG, Berlin, Germany
Reply-to: Thomas Graichen <graichen@xxxxxxxxxxxxx>
Reply-to: thomas.graichen@xxxxxxxxxxxxx
Sender: owner-linux-xfs@xxxxxxxxxxx
User-agent: tin/1.4.2-20000205 ("Possession") (UNIX) (Linux/2.2.16-local (i586))
just built xfs fromm an rsync of the current cvs tree and wanted to
move a system to xfs (for some more realy world testing) - but after
mkfs.xfs'ing the device (ide, no kiobufs), mounting it, umounting it
and again mounting it i got

[root@ozon /root]# mkfs -t xfs -f /dev/hda3
meta-data=/dev/hda3              isize=256    agcount=8, agsize=65766 blks
data     =                       bsize=4096   blocks=526128, imaxpct=25
         =                       sunit=0      swidth=0 blks, unwritten=0
naming   =version 2              bsize=4096  
log      =internal log           bsize=4096   blocks=1200
realtime =none                   extsz=65536  blocks=0, rtextents=0
[root@ozon /root]# mount -t xfs /dev/hda3 /mnt
Start mounting filesystem: ide0(3,3)
Ending clean XFS mount for filesystem: ide0(3,3)
[root@ozon /root]# umount /mnt
[root@ozon /root]# mount -t xfs /dev/hda3 /mnt
Start mounting filesystem: ide0(3,3)
XFS: xlog_find_verify_log_record: need to backup
XFS: failed to find log head

XFS: log mount/recovery failed

XFS: log mount failed

mount: wrong fs type, bad option, bad superblock on /dev/hda3,
       or too many mounted file systems
[root@ozon /root]# 

any idea ? is this reproducable for others too ?

(it is on redhat 6.2 - thus egcs, 233mmx pentium, also the cmd's
fresh rebuild - from real current sources)

a lot of thanks in advance


technical director                                       innominate AG
clustering & security                                networking people
tel: +49.30.308806-13  fax: -77                   http://innominate.de

<Prev in Thread] Current Thread [Next in Thread>