[Ocfs2-devel] ocfs2: Possible deadlock in dlm?

Tao Ma tao.ma at oracle.com
Tue Feb 9 00:43:23 PST 2010


Hi Sunil/Joel,
	I just got a lockdep warning today when I enable it to
reflink check.

So the question is that dlm_domain_lock and dlm->spinlock are
spin_locked in different order.

In dlm_run_purge_list, we lock dlm->spinlock first and then in
dlm_lockres_release->dlm_put we lock dlm_domain_lock.
While in dlm_mark_domain_leaving we use the reverse order.

So is this a problem or these 2 scenarios can never happen together?
I have attached the lockdep print below.

Regards,
Tao

=======================================================
[ INFO: possible circular locking dependency detected ]
2.6.33-rc6 #2
-------------------------------------------------------
umount/3880 is trying to acquire lock:
 (&(&dlm->spinlock)->rlock){+.+...}, at: [<ffffffffa045c6e0>] dlm_unregister_domain+0x465/0x7ce [ocfs2_dlm]

but task is already holding lock:
 (dlm_domain_lock){+.+...}, at: [<ffffffffa045c6d7>] dlm_unregister_domain+0x45c/0x7ce [ocfs2_dlm]

which lock already depends on the new lock.


the existing dependency chain (in reverse order) is:

-> #1 (dlm_domain_lock){+.+...}:
       [<ffffffff82065a01>] validate_chain+0xa40/0xd38
       [<ffffffff820664a6>] __lock_acquire+0x7ad/0x813
       [<ffffffff820665d3>] lock_acquire+0xc7/0xe4
       [<ffffffff8233d3da>] _raw_spin_lock+0x31/0x66
       [<ffffffffa045bf74>] dlm_put+0x1f/0x3e [ocfs2_dlm]
       [<ffffffffa046b6f6>] dlm_lockres_release+0x132/0x30e [ocfs2_dlm]
       [<ffffffff8218bbea>] kref_put+0x43/0x4f
       [<ffffffffa046999d>] dlm_lockres_put+0x14/0x16 [ocfs2_dlm]
       [<ffffffffa0460277>] dlm_run_purge_list+0x494/0x4df [ocfs2_dlm]
       [<ffffffffa04605c7>] dlm_thread+0x9d/0xe32 [ocfs2_dlm]
       [<ffffffff82054d4d>] kthread+0x7d/0x85
       [<ffffffff82003794>] kernel_thread_helper+0x4/0x10

-> #0 (&(&dlm->spinlock)->rlock){+.+...}:
       [<ffffffff820656ed>] validate_chain+0x72c/0xd38
       [<ffffffff820664a6>] __lock_acquire+0x7ad/0x813
       [<ffffffff820665d3>] lock_acquire+0xc7/0xe4
       [<ffffffff8233d3da>] _raw_spin_lock+0x31/0x66
       [<ffffffffa045c6e0>] dlm_unregister_domain+0x465/0x7ce [ocfs2_dlm]
       [<ffffffffa04941d5>] o2cb_cluster_disconnect+0x38/0x4a [ocfs2_stack_o2cb]
       [<ffffffffa04012ba>] ocfs2_cluster_disconnect+0x2a/0x4e [ocfs2_stackglue]
       [<ffffffffa04e95f4>] ocfs2_dlm_shutdown+0xf9/0x167 [ocfs2]
       [<ffffffffa051f606>] ocfs2_dismount_volume+0x1d8/0x39e [ocfs2]
       [<ffffffffa051fbe8>] ocfs2_put_super+0x88/0xf4 [ocfs2]
       [<ffffffff820dcc6d>] generic_shutdown_super+0x58/0xcc
       [<ffffffff820dcd03>] kill_block_super+0x22/0x3a
       [<ffffffffa051d671>] ocfs2_kill_sb+0x77/0x7f [ocfs2]
       [<ffffffff820dd438>] deactivate_super+0x68/0x7d
       [<ffffffff820f11ba>] mntput_no_expire+0x75/0xb0
       [<ffffffff820f172a>] sys_umount+0x2c2/0x321
       [<ffffffff8200296b>] system_call_fastpath+0x16/0x1b

other info that might help us debug this:




More information about the Ocfs2-devel mailing list