[Ocfs2-devel] [PATCH] ocfs2/dlm: correct the refmap on recovery master

Wengang Wang wen.gang.wang at oracle.com
Mon Jul 5 03:00:31 PDT 2010


Any comment?

On 10-06-25 09:55, Wengang Wang wrote:
> Hi,
> 
> Any comment on this?
> 
> regards,
> wengang.
> On 10-06-11 00:25, Wengang Wang wrote:
> > If the dlm recovery goes on the non-master node where purging work is going on, 
> > There could be unexpected reference left on some lockres' on recovery master.
> > That is because we migrated the lockres' to recovery master but didn't send
> > deref requests to it accordingly(was sending to the dead original master or to
> > the "UNKNOWN").
> > 
> > Fix:
> > For the lockres which is in progress of dropping reference, we don't migrate it
> > to recovery master and unhash the lockres in the purge work.
> > For those not in progress of the dropping, delay the purge work until recovery
> > finished so that it can send deref request to the correct master(recovery
> > master) later.
> > 
> > Signed-off-by: Wengang Wang <wen.gang.wang at oracle.com>
> > ---
> >  fs/ocfs2/dlm/dlmrecovery.c |   17 +++++++++++++++--
> >  fs/ocfs2/dlm/dlmthread.c   |   36 ++++++++++++++++++++++--------------
> >  2 files changed, 37 insertions(+), 16 deletions(-)
> > 
> > diff --git a/fs/ocfs2/dlm/dlmrecovery.c b/fs/ocfs2/dlm/dlmrecovery.c
> > index f8b75ce..43530ce 100644
> > --- a/fs/ocfs2/dlm/dlmrecovery.c
> > +++ b/fs/ocfs2/dlm/dlmrecovery.c
> > @@ -1997,6 +1997,8 @@ void dlm_move_lockres_to_recovery_list(struct dlm_ctxt *dlm,
> >  	struct list_head *queue;
> >  	struct dlm_lock *lock, *next;
> >  
> > +	assert_spin_locked(&dlm->spinlock);
> > +	assert_spin_locked(&res->spinlock);
> >  	res->state |= DLM_LOCK_RES_RECOVERING;
> >  	if (!list_empty(&res->recovering)) {
> >  		mlog(0,
> > @@ -2336,9 +2338,20 @@ static void dlm_do_local_recovery_cleanup(struct dlm_ctxt *dlm, u8 dead_node)
> >  
> >  				/* the wake_up for this will happen when the
> >  				 * RECOVERING flag is dropped later */
> > -				res->state &= ~DLM_LOCK_RES_DROPPING_REF;
> > +				if (res->state & DLM_LOCK_RES_DROPPING_REF) {
> > +					/*
> > +					 * don't migrate a lockres which is in
> > +					 * progress of dropping ref
> > +					 */
> > +					mlog(ML_NOTICE, "%.*s ignored for "
> > +					     "migration\n", res->lockname.len,
> > +					     res->lockname.name);
> > +					res->state &=
> > +						~DLM_LOCK_RES_DROPPING_REF;
> > +				} else
> > +					dlm_move_lockres_to_recovery_list(dlm,
> > +									  res);
> >  
> > -				dlm_move_lockres_to_recovery_list(dlm, res);
> >  			} else if (res->owner == dlm->node_num) {
> >  				dlm_free_dead_locks(dlm, res, dead_node);
> >  				__dlm_lockres_calc_usage(dlm, res);
> > diff --git a/fs/ocfs2/dlm/dlmthread.c b/fs/ocfs2/dlm/dlmthread.c
> > index d4f73ca..0771420 100644
> > --- a/fs/ocfs2/dlm/dlmthread.c
> > +++ b/fs/ocfs2/dlm/dlmthread.c
> > @@ -92,17 +92,23 @@ int __dlm_lockres_has_locks(struct dlm_lock_resource *res)
> >   * truly ready to be freed. */
> >  int __dlm_lockres_unused(struct dlm_lock_resource *res)
> >  {
> > -	if (!__dlm_lockres_has_locks(res) &&
> > -	    (list_empty(&res->dirty) && !(res->state & DLM_LOCK_RES_DIRTY))) {
> > -		/* try not to scan the bitmap unless the first two
> > -		 * conditions are already true */
> > -		int bit = find_next_bit(res->refmap, O2NM_MAX_NODES, 0);
> > -		if (bit >= O2NM_MAX_NODES) {
> > -			/* since the bit for dlm->node_num is not
> > -			 * set, inflight_locks better be zero */
> > -			BUG_ON(res->inflight_locks != 0);
> > -			return 1;
> > -		}
> > +	int bit;
> > +
> > +	if (__dlm_lockres_has_locks(res))
> > +		return 0;
> > +
> > +	if (!list_empty(&res->dirty) || res->state & DLM_LOCK_RES_DIRTY)
> > +		return 0;
> > +
> > +	if (res->state & DLM_LOCK_RES_RECOVERING)
> > +		return 0;
> > +
> > +	bit = find_next_bit(res->refmap, O2NM_MAX_NODES, 0);
> > +	if (bit >= O2NM_MAX_NODES) {
> > +		/* since the bit for dlm->node_num is not
> > +		 * set, inflight_locks better be zero */
> > +		BUG_ON(res->inflight_locks != 0);
> > +		return 1;
> >  	}
> >  	return 0;
> >  }
> > @@ -158,6 +164,8 @@ static int dlm_purge_lockres(struct dlm_ctxt *dlm,
> >  	int master;
> >  	int ret = 0;
> >  
> > +	assert_spin_locked(&dlm->spinlock);
> > +
> >  	spin_lock(&res->spinlock);
> >  	if (!__dlm_lockres_unused(res)) {
> >  		mlog(0, "%s:%.*s: tried to purge but not unused\n",
> > @@ -216,13 +224,13 @@ static int dlm_purge_lockres(struct dlm_ctxt *dlm,
> >  		     "master = %d\n", res->lockname.len, res->lockname.name,
> >  		     res, master);
> >  		list_del_init(&res->purge);
> > -		spin_unlock(&res->spinlock);
> > +		/* not the last ref */
> >  		dlm_lockres_put(res);
> >  		dlm->purge_count--;
> > -	} else
> > -		spin_unlock(&res->spinlock);
> > +	}
> >  
> >  	__dlm_unhash_lockres(res);
> > +	spin_unlock(&res->spinlock);
> >  
> >  	/* lockres is not in the hash now.  drop the flag and wake up
> >  	 * any processes waiting in dlm_get_lock_resource. */
> > -- 
> > 1.6.6.1
> > 
> > 
> > _______________________________________________
> > Ocfs2-devel mailing list
> > Ocfs2-devel at oss.oracle.com
> > http://oss.oracle.com/mailman/listinfo/ocfs2-devel



More information about the Ocfs2-devel mailing list