[Ocfs2-users] more ocfs2_delete_inode dmesg questions

Brian Kroth bpkroth at gmail.com
Tue Oct 6 15:44:05 PDT 2009


So, we've found that this has actually been causing some dropped mail
and backlogs.

Here's the situation:

MX servers filter into the main mail server, all running sendmail.  The
main mail server has an OCFS2 spool volume which will periodically throw
those error messages in dmesg that I listed earlier.

Sendmail returns one of these two sequences:

Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   0: fl=0x0, mode=20666: CHR: dev=0/13, ino=679, nlink=1, u/gid=0/0, size=0
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   1: fl=0x802, mode=140777: SOCK smtp/25->mx2/54625
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   2: fl=0x1, mode=20666: CHR: dev=0/13, ino=679, nlink=1, u/gid=0/0, size=0
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   3: fl=0x2, mode=140777: SOCK localhost->[[UNIX: /dev/log]]
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   4: fl=0x802, mode=140777: SOCK smtp/25->mx2/54625
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   5: fl=0x0, mode=100640: dev=8/33, ino=580655, nlink=1, u/gid=0/25, size=164461
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   6: fl=0x8000, mode=100644: dev=8/1, ino=175982, nlink=1, u/gid=107/25, size=12288
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   7: fl=0x8000, mode=100644: dev=8/1, ino=175982, nlink=1, u/gid=107/25, size=12288
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   8: fl=0x8000, mode=100644: dev=8/1, ino=175663, nlink=1, u/gid=107/25, size=49152
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:   9: fl=0x802, mode=140777: SOCK smtp/25->mx2/54625
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:  10: fl=0x8000, mode=100644: dev=8/1, ino=175663, nlink=1, u/gid=107/25, size=49152
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:  11: fl=0x8000, mode=100644: dev=8/1, ino=175662, nlink=1, u/gid=107/25, size=2621440
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:  12: fl=0x8000, mode=100644: dev=8/1, ino=175662, nlink=1, u/gid=107/25, size=2621440
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:  13: fl=0x8000, mode=100644: dev=8/1, ino=175622, nlink=1, u/gid=107/25, size=2543616
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407:  14: fl=0x8000, mode=100644: dev=8/1, ino=175622, nlink=1, u/gid=107/25, size=2543616
Oct  6 16:46:02 iris sm-mta[14407]: n96Lk23h014407: SYSERR(root): queueup: cannot create queue file ./qfn96Lk23h014407, euid=0, fd=-1, fp=0x0: File exists

This results in dropped mail.

Oct  6 16:26:09 iris sm-mta[29393]: n96LQ4uK029393: SYSERR(root): collect: bfcommit(./dfn96LQ4uK029393): already on disk, size=0: File exists

This results in a failure being propagated to the mx, which then
backlogs mail for a while before it retries.

The debufs.ocfs2 command allows me to identify the file, which turns out
to be a 0 byte file on the spool volume.  If it's a data file I can
usually examine the cooresponding control file, or vise-versa, but after
not too long, both are removed, or rather replaced by new files.  I'm
not sure if that's just sendmail shoving lots of files around and
therefore needing to reuse inodes, or sendmail cleaning up after itself,
or ocfs2 cleaning up after itself.

For now I've moved the spool volume back to a local disk, but may try to
test this out some more on our test setup.

Any thoughts?

Thanks,
Brian

Brian Kroth <bpkroth at gmail.com> 2009-08-25 08:52:
> Sunil Mushran <sunil.mushran at oracle.com> 2009-08-24 18:12:
> > So a delete was called for some inodes that had not been orphaned.
> > The pre-checks detected the same and correctly aborted the deletes.
> > No harm done.
> 
> Very good to hear.
> 
> > No, the messages do not pinpoint the device. It's something we discussed
> > adding, but have not done it as yet.
> > 
> > Next time this happens and you can identify the volume, do:
> > # debugfs.ocfs2 -R "findpath <613069>" /dev/sdX
> > 
> > This will tell you the pathname for the inode#. Then see if you can remember
> > performing any op on that file. Anything. It may help us narrow down the
> > issue.
> > 
> > Sunil
> 
> Will do.
> 
> Thanks again,
> Brian
> 
> > Brian Kroth wrote:
> > > I recently brought up a mail server with two ocfs2 volumes on it, one
> > > large one for the user maildirs, and one small one for queue/spool
> > > directories.  More information on the specifics below.  When flushing
> > > the queues from the MXs I saw the messages listed below fly by, but
> > > since then nothing.
> > >
> > > A couple of questions:
> > > - Should I be worried about these?  They seemed similar yet different to
> > >   a number of other "out of space" and "failure to delete" reports of
> > >   late.
> > > - How can I tell which volume has the problem inodes?
> > > - Is there anything to be done about them?
> > >
> > > Here's the snip from the tail of dmesg:
> > >
> > > [   34.578787] netconsole: network logging started
> > > [   36.695679] ocfs2: Registered cluster interface o2cb
> > > [   43.354897] OCFS2 1.5.0
> > > [   43.373100] ocfs2_dlm: Nodes in domain ("94468EF57C9F4CA18C8D218C63E99A9C"): 1 
> > > [   43.386623] kjournald2 starting: pid 2328, dev sdb1:36, commit interval 5 seconds
> > > [   43.395413] ocfs2: Mounting device (8,17) on (node 1, slot 0) with ordered data mode.
> > > [   44.984201] eth1: no IPv6 routers present
> > > [   54.362580] warning: `ntpd' uses 32-bit capabilities (legacy support in use)
> > > [ 1601.560932] ocfs2_dlm: Nodes in domain ("10BBA4EB7687450496F7FCF0475F9372"): 1 
> > > [ 1601.581106] kjournald2 starting: pid 7803, dev sdc1:36, commit interval 5 seconds
> > > [ 1601.593065] ocfs2: Mounting device (8,33) on (node 1, slot 0) with ordered data mode.
> > > [ 3858.778792] (26441,0):ocfs2_query_inode_wipe:882 ERROR: Inode 613069 (on-disk 613069) not orphaned! Disk flags  0x1, inode flags 0x80
> > > [ 3858.779005] (26441,0):ocfs2_delete_inode:1010 ERROR: status = -17
> > > [ 4451.007580] (5053,0):ocfs2_query_inode_wipe:882 ERROR: Inode 613118 (on-disk 613118) not orphaned! Disk flags  0x1, inode flags 0x80
> > > [ 4451.007711] (5053,0):ocfs2_delete_inode:1010 ERROR: status = -17
> > > [ 4807.908463] (11859,0):ocfs2_query_inode_wipe:882 ERROR: Inode 612899 (on-disk 612899) not orphaned! Disk flags  0x1, inode flags 0x80
> > > [ 4807.908611] (11859,0):ocfs2_delete_inode:1010 ERROR: status = -17
> > > [ 5854.377155] (31074,1):ocfs2_query_inode_wipe:882 ERROR: Inode 612867 (on-disk 612867) not orphaned! Disk flags  0x1, inode flags 0x80
> > > [ 5854.377302] (31074,1):ocfs2_delete_inode:1010 ERROR: status = -17
> > > [ 6136.297464] (3463,0):ocfs2_query_inode_wipe:882 ERROR: Inode 612959 (on-disk 612959) not orphaned! Disk flags  0x1, inode flags 0x80
> > > [ 6136.297555] (3463,0):ocfs2_delete_inode:1010 ERROR: status = -17
> > > [19179.000100] NOHZ: local_softirq_pending 80
> > >
> > >
> > > There's actually three nodes, all VMs, that are setup for the ocfs2
> > > cluster volumes, but only one has it mounted.  The others are available
> > > as cold standbys that may eventually be managed by heartbeat, so there
> > > shouldn't be any locking contention going on.
> > >
> > > All nodes are running 2.6.30 with ocfs2-tools 1.4.2.
> > >
> > > Here's the commands used to make the volumes:
> > > mkfs.ocfs2 -v -L ocfs2mailcluster2 -N 8 -T mail /dev/sdb1
> > > mkfs.ocfs2 -v -L ocfs2mailcluster2spool -N 8 -T mail /dev/sdc1
> > >
> > > The features the were setup with:
> > > tunefs.ocfs2 -Q "Label: %V\nFeatures: %H %O\n" /dev/sdb1
> > > Label: ocfs2mailcluster2
> > > Features: sparse inline-data unwritten
> > >
> > > tunefs.ocfs2 -Q "Label: %V\nFeatures: %H %O\n" /dev/sdc1
> > > Label: ocfs2mailcluster2spool
> > > Features: sparse inline-data unwritten
> > >
> > > And their mount options:
> > > mount | grep cluster
> > > /dev/sdb1 on /cluster type ocfs2 (rw,noexec,nodev,_netdev,relatime,localflocks,heartbeat=local)
> > > /dev/sdc1 on /cluster-spool type ocfs2 (rw,noexec,nodev,_netdev,relatime,localflocks,heartbeat=local)
> > >
> > > localflocks because I ran into a problem with them previously, and since
> > > it's a single active node model currently there's no reason for them
> > > anyways.
> > >
> > > Let me know if you need any other information.
> > >
> > > Thanks,
> > > Brian
> > >
> > > _______________________________________________
> > > Ocfs2-users mailing list
> > > Ocfs2-users at oss.oracle.com
> > > http://oss.oracle.com/mailman/listinfo/ocfs2-users
> > >   
> > 
> > 
> > _______________________________________________
> > Ocfs2-users mailing list
> > Ocfs2-users at oss.oracle.com
> > http://oss.oracle.com/mailman/listinfo/ocfs2-users



More information about the Ocfs2-users mailing list