[Ocfs2-devel] [PATCH] ocfs2: check if the ocfs2 lock resource be initialized before calling ocfs2_dlm_lock
alex chen
alex.chen at huawei.com
Wed Apr 22 01:14:09 PDT 2015
On 2015/4/22 9:46, Junxiao Bi wrote:
> On 04/21/2015 10:54 AM, alex chen wrote:
>> Hi Junxiao,
>>
>> On 2015/4/16 15:28, Junxiao Bi wrote:
>>> Hi Alex,
>>>
>>> On 03/30/2015 11:22 AM, alex chen wrote:
>>>> If ocfs2 lockres has not been initialized before calling ocfs2_dlm_lock,
>>>> the lock won't be dropped and then will lead umount hung. The case is
>>>> described below:
>>>>
>>>> ocfs2_mknod
>>>> ocfs2_mknod_locked
>>>> __ocfs2_mknod_locked
>>>> ocfs2_journal_access_di
>>>> Failed because of -ENOMEM or other reasons, the inode lockres
>>>> has not been initialized yet.
>>>
>>> If failed here, is OCFS2_I(inode)->ip_inode_lockres initialized? If not
>>
>> The OCFS2_I(inode)->ip_inode_lockres is initialized as follows:
>> __ocfs2_mknod_locked
>> ocfs2_populate_inode
>> ocfs2_inode_lock_res_init
>> ocfs2_lock_res_init_common
>> So if ocfs2_journal_access_di is failed, the ip_inode_lockres will not be
>> initialized.
>> In this situation, we should not allocate a new dlm lockres through calling
>> ocfs2_dlm_lock() in __ocfs2_cluster_lock(), otherwise it will lead umount
>> hung. So we need bread __ocfs2_cluster_lock() if the inode lockres is not
>> be initialized, that is the condition
>> (!(lockres->l_flags & OCFS2_LOCK_INITIALIZED)) is TRUE.
>
> Looks good. Just didn't remind that inode_info is init once when it is
> allocated from "ocfs2_inode_cache". lockres->l_flags is init there.
>
> Thanks,
> Junxiao.
>
Thank you for your remind. If the subject is "check if lockres->l_flags be
initialized to OCFS2_LOCK_INITIALIZED before calling ocfs2_dlm_lock" will
be better.
>>
>>> how can you break __ocfs2_cluster_lock with the following condition?
>>>
>>> if (!(lockres->l_flags & OCFS2_LOCK_INITIALIZED))
>>>
>>> Thanks,
>>> Junxiao.
>>>
>>>>
>>>> iput(inode)
>>>> ocfs2_evict_inode
>>>> ocfs2_delete_inode
>>>> ocfs2_inode_lock
>>>> ocfs2_inode_lock_full_nested
>>>> __ocfs2_cluster_lock
>>>> Succeeds and allocates a new dlm lockres.
>>>> ocfs2_clear_inode
>>>> ocfs2_open_unlock
>>>> ocfs2_drop_inode_locks
>>>> ocfs2_drop_lock
>>>> Since lockres has not been initialized, the lock
>>>> can't be dropped and the lockres can't be
>>>> migrated, thus umount will hang forever.
>>>>
>>>> Signed-off-by: Alex Chen <alex.chen at huawei.com>
>>>> Reviewed-by: Joseph Qi <joseph.qi at huawei.com>
>>>> Reviewed-by: joyce.xue <xuejiufei at huawei.com>
>>>>
>>>> ---
>>>> fs/ocfs2/dlmglue.c | 5 +++++
>>>> 1 file changed, 5 insertions(+)
>>>>
>>>> diff --git a/fs/ocfs2/dlmglue.c b/fs/ocfs2/dlmglue.c
>>>> index 11849a4..8b23aa2 100644
>>>> --- a/fs/ocfs2/dlmglue.c
>>>> +++ b/fs/ocfs2/dlmglue.c
>>>> @@ -1391,6 +1391,11 @@ static int __ocfs2_cluster_lock(struct ocfs2_super *osb,
>>>> int noqueue_attempted = 0;
>>>> int dlm_locked = 0;
>>>>
>>>> + if (!(lockres->l_flags & OCFS2_LOCK_INITIALIZED)) {
>>>> + mlog_errno(-EINVAL);
>>>> + return -EINVAL;
>>>> + }
>>>> +
>>>> ocfs2_init_mask_waiter(&mw);
>>>>
>>>> if (lockres->l_ops->flags & LOCK_TYPE_USES_LVB)
>>>>
>>>
>>>
>>> .
>>>
>>
>
>
> .
>
More information about the Ocfs2-devel
mailing list