[Ocfs2-users] more ocfs2_delete_inode dmesg questions

Brian Kroth bpkroth at gmail.com
Mon Aug 24 06:26:12 PDT 2009


I recently brought up a mail server with two ocfs2 volumes on it, one
large one for the user maildirs, and one small one for queue/spool
directories.  More information on the specifics below.  When flushing
the queues from the MXs I saw the messages listed below fly by, but
since then nothing.

A couple of questions:
- Should I be worried about these?  They seemed similar yet different to
  a number of other "out of space" and "failure to delete" reports of
  late.
- How can I tell which volume has the problem inodes?
- Is there anything to be done about them?

Here's the snip from the tail of dmesg:

[   34.578787] netconsole: network logging started
[   36.695679] ocfs2: Registered cluster interface o2cb
[   43.354897] OCFS2 1.5.0
[   43.373100] ocfs2_dlm: Nodes in domain ("94468EF57C9F4CA18C8D218C63E99A9C"): 1 
[   43.386623] kjournald2 starting: pid 2328, dev sdb1:36, commit interval 5 seconds
[   43.395413] ocfs2: Mounting device (8,17) on (node 1, slot 0) with ordered data mode.
[   44.984201] eth1: no IPv6 routers present
[   54.362580] warning: `ntpd' uses 32-bit capabilities (legacy support in use)
[ 1601.560932] ocfs2_dlm: Nodes in domain ("10BBA4EB7687450496F7FCF0475F9372"): 1 
[ 1601.581106] kjournald2 starting: pid 7803, dev sdc1:36, commit interval 5 seconds
[ 1601.593065] ocfs2: Mounting device (8,33) on (node 1, slot 0) with ordered data mode.
[ 3858.778792] (26441,0):ocfs2_query_inode_wipe:882 ERROR: Inode 613069 (on-disk 613069) not orphaned! Disk flags  0x1, inode flags 0x80
[ 3858.779005] (26441,0):ocfs2_delete_inode:1010 ERROR: status = -17
[ 4451.007580] (5053,0):ocfs2_query_inode_wipe:882 ERROR: Inode 613118 (on-disk 613118) not orphaned! Disk flags  0x1, inode flags 0x80
[ 4451.007711] (5053,0):ocfs2_delete_inode:1010 ERROR: status = -17
[ 4807.908463] (11859,0):ocfs2_query_inode_wipe:882 ERROR: Inode 612899 (on-disk 612899) not orphaned! Disk flags  0x1, inode flags 0x80
[ 4807.908611] (11859,0):ocfs2_delete_inode:1010 ERROR: status = -17
[ 5854.377155] (31074,1):ocfs2_query_inode_wipe:882 ERROR: Inode 612867 (on-disk 612867) not orphaned! Disk flags  0x1, inode flags 0x80
[ 5854.377302] (31074,1):ocfs2_delete_inode:1010 ERROR: status = -17
[ 6136.297464] (3463,0):ocfs2_query_inode_wipe:882 ERROR: Inode 612959 (on-disk 612959) not orphaned! Disk flags  0x1, inode flags 0x80
[ 6136.297555] (3463,0):ocfs2_delete_inode:1010 ERROR: status = -17
[19179.000100] NOHZ: local_softirq_pending 80


There's actually three nodes, all VMs, that are setup for the ocfs2
cluster volumes, but only one has it mounted.  The others are available
as cold standbys that may eventually be managed by heartbeat, so there
shouldn't be any locking contention going on.

All nodes are running 2.6.30 with ocfs2-tools 1.4.2.

Here's the commands used to make the volumes:
mkfs.ocfs2 -v -L ocfs2mailcluster2 -N 8 -T mail /dev/sdb1
mkfs.ocfs2 -v -L ocfs2mailcluster2spool -N 8 -T mail /dev/sdc1

The features the were setup with:
tunefs.ocfs2 -Q "Label: %V\nFeatures: %H %O\n" /dev/sdb1
Label: ocfs2mailcluster2
Features: sparse inline-data unwritten

tunefs.ocfs2 -Q "Label: %V\nFeatures: %H %O\n" /dev/sdc1
Label: ocfs2mailcluster2spool
Features: sparse inline-data unwritten

And their mount options:
mount | grep cluster
/dev/sdb1 on /cluster type ocfs2 (rw,noexec,nodev,_netdev,relatime,localflocks,heartbeat=local)
/dev/sdc1 on /cluster-spool type ocfs2 (rw,noexec,nodev,_netdev,relatime,localflocks,heartbeat=local)

localflocks because I ran into a problem with them previously, and since
it's a single active node model currently there's no reason for them
anyways.

Let me know if you need any other information.

Thanks,
Brian



More information about the Ocfs2-users mailing list