[Ocfs2-devel] [PATCH] ocfs2: Fix locking for res->tracking and dlm->tracking_list
Changwei Ge
gechangwei at live.cn
Sun Jun 24 18:07:10 PDT 2018
On 2018/6/23 7:33, Ashish Samant wrote:
>
>
> On 06/22/2018 02:25 AM, Changwei Ge wrote:
>>
>> On 2018/6/22 16:55, Joseph Qi wrote:
>>> On 18/6/22 16:50, Changwei Ge wrote:
>>>> On 2018/6/22 16:32, Joseph Qi wrote:
>>>>> On 18/6/22 07:57, Ashish Samant wrote:
>>>>>> In dlm_init_lockres() and dlm_unregister_domain() we access and
>>>>>> modify
>>>>>> res->tracking and dlm->tracking_list without holding
>>>>>> dlm->track_lock.
>>>>>> This can cause list corruptions and can end up in kernel panic.
>>>>>>
>>>>>> Fix this by locking res->tracking and dlm->tracking_list with
>>>>>> dlm->track_lock at all places.
>>>>>>
>>>>>> Signed-off-by: Ashish Samant <ashish.samant at oracle.com>
>>>>>> ---
>>>>>> fs/ocfs2/dlm/dlmdomain.c | 2 ++
>>>>>> fs/ocfs2/dlm/dlmmaster.c | 4 ++--
>>>>>> 2 files changed, 4 insertions(+), 2 deletions(-)
>>>>>>
>>>>>> diff --git a/fs/ocfs2/dlm/dlmdomain.c b/fs/ocfs2/dlm/dlmdomain.c
>>>>>> index 2acd58b..cfb1edd 100644
>>>>>> --- a/fs/ocfs2/dlm/dlmdomain.c
>>>>>> +++ b/fs/ocfs2/dlm/dlmdomain.c
>>>>>> @@ -723,6 +723,7 @@ void dlm_unregister_domain(struct dlm_ctxt *dlm)
>>>>>> mlog(0, "%s: more migration to do\n", dlm->name);
>>>>>> }
>>>>>> + spin_lock(&dlm->track_lock);
>>>>>> /* This list should be empty. If not, print remaining
>>>>>> lockres */
>>>>>> if (!list_empty(&dlm->tracking_list)) {
>>>>>> mlog(ML_ERROR, "Following lockres' are still on
>>>>>> the "
>>>>>> @@ -730,6 +731,7 @@ void dlm_unregister_domain(struct dlm_ctxt *dlm)
>>>>>> list_for_each_entry(res, &dlm->tracking_list,
>>>>>> tracking)
>>>>>> dlm_print_one_lock_resource(res);
>>>>>> }
>>>>>> + spin_unlock(&dlm->track_lock);
>>>>> The locking order should be res->spinlock > dlm->track_lock.
>>>>> Since here just want to print error message for issue tracking, I'm
>>>>> wandering if we can copy tracking list to local first.
>
> Right, for some reason, I was thinking the call is to
> __dlm_print_lock_resource() and not dlm_print_one_lock_resource(). So
> this could deadlock.
>
>>>> That won't be easy since I think the copying should also should lock
>>>> resource lock.
>>> Copy tracking list only need taking track_lock.
>>> Then access local tracking list we don't have to take it any more
>>> and then we can call dlm_print_one_lock_resource() which will take
>>> res->spinlock.
>> I thought you' want to copy lock resources as well.
>> Um, is it possible that the copied track list points to some stale lock
>> resources which are released after the copy.
> Yes dropping the track_lock can still cause the same problem. However,
> I am wondering , since this is during dlm unregister domain/ cluster
> disconnect after the dlm_thread has run, under what conditions would a
> concurrent access to the tracking_list occur at this point?
I think your assumption stands, we don't have to worry much about
concurrent access to the ::tracking_list. DLM should make sure that
after migrating all lock resources, no more lock resources should be born.
>
> Thanks,
> Ashish
>
>>
>> Thanks,
>> Changwei
>>
>>> Thanks,
>>> Joseph
>>>
>>>> Perhaps, we can remove lock resource from dlm->track_list only when
>>>> the
>>>> lock resource is released.
>>>> It brings another benefit that we can easily find which lock
>>>> resource is
>>>> leaked.
>>>>
>>>> Thanks,
>>>> Changwei
>>>>
>>>>> Thanks,
>>>>> Joseph
>>>>>
>>>>>> dlm_mark_domain_leaving(dlm);
>>>>>> dlm_leave_domain(dlm);
>>>>>> diff --git a/fs/ocfs2/dlm/dlmmaster.c b/fs/ocfs2/dlm/dlmmaster.c
>>>>>> index aaca094..826f056 100644
>>>>>> --- a/fs/ocfs2/dlm/dlmmaster.c
>>>>>> +++ b/fs/ocfs2/dlm/dlmmaster.c
>>>>>> @@ -584,9 +584,9 @@ static void dlm_init_lockres(struct dlm_ctxt
>>>>>> *dlm,
>>>>>> res->last_used = 0;
>>>>>> - spin_lock(&dlm->spinlock);
>>>>>> + spin_lock(&dlm->track_lock);
>>>>>> list_add_tail(&res->tracking, &dlm->tracking_list);
>>>>>> - spin_unlock(&dlm->spinlock);
>>>>>> + spin_unlock(&dlm->track_lock);
> Maybe we only need this to fix the issue.
Agree. Could you resend your patch?
Thanks,
Changwei
>
> Thanks,
> Ashish
>
>
>>>>>> memset(res->lvb, 0, DLM_LVB_LEN);
>>>>>> memset(res->refmap, 0, sizeof(res->refmap));
>>>>>>
>>>>> _______________________________________________
>>>>> Ocfs2-devel mailing list
>>>>> Ocfs2-devel at oss.oracle.com
>>>>> https://oss.oracle.com/mailman/listinfo/ocfs2-devel
>
More information about the Ocfs2-devel
mailing list