/* * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. * Copyright (C) 2004-2005 Red Hat, Inc. All rights reserved. * * This copyrighted material is made available to anyone wishing to use, * modify, copy, or redistribute it subject to the terms and conditions * of the GNU General Public License v.2. */ #include "lock_dlm.h" /* A lock placed on this queue is re-submitted to DLM as soon as the lock_dlm thread gets to it. */ static void queue_submit(struct gdlm_lock *lp) { struct gdlm_ls *ls = lp->ls; spin_lock(&ls->async_lock); list_add_tail(&lp->delay_list, &ls->submit); spin_unlock(&ls->async_lock); wake_up(&ls->thread_wait); } static void process_blocking(struct gdlm_lock *lp, int bast_mode) { struct gdlm_ls *ls = lp->ls; unsigned int cb = 0; switch (gdlm_make_lmstate(bast_mode)) { case LM_ST_EXCLUSIVE: cb = LM_CB_NEED_E; break; case LM_ST_DEFERRED: cb = LM_CB_NEED_D; break; case LM_ST_SHARED: cb = LM_CB_NEED_S; break; default: gdlm_assert(0, "unknown bast mode %u", lp->bast_mode); } ls->fscb(ls->fsdata, cb, &lp->lockname); } static void process_complete(struct gdlm_lock *lp) { struct gdlm_ls *ls = lp->ls; struct lm_async_cb acb; int16_t prev_mode = lp->cur; memset(&acb, 0, sizeof(acb)); if (lp->lksb.sb_status == -DLM_ECANCEL) { log_info("complete dlm cancel %x,%llx flags %lx", lp->lockname.ln_type, (unsigned long long)lp->lockname.ln_number, lp->flags); lp->req = lp->cur; acb.lc_ret |= LM_OUT_CANCELED; if (lp->cur == DLM_LOCK_IV) lp->lksb.sb_lkid = 0; goto out; } if (test_and_clear_bit(LFL_DLM_UNLOCK, &lp->flags)) { if (lp->lksb.sb_status != -DLM_EUNLOCK) { log_info("unlock sb_status %d %x,%llx flags %lx", lp->lksb.sb_status, lp->lockname.ln_type, (unsigned long long)lp->lockname.ln_number, lp->flags); return; } lp->cur = DLM_LOCK_IV; lp->req = DLM_LOCK_IV; lp->lksb.sb_lkid = 0; if (test_and_clear_bit(LFL_UNLOCK_DELETE, &lp->flags)) { gdlm_delete_lp(lp); return; } goto out; } if (lp->lksb.sb_flags & DLM_SBF_VALNOTVALID) memset(lp->lksb.sb_lvbptr, 0, GDLM_LVB_SIZE); if (lp->lksb.sb_flags & DLM_SBF_ALTMODE) { if (lp->req == DLM_LOCK_PR) lp->req = DLM_LOCK_CW; else if (lp->req == DLM_LOCK_CW) lp->req = DLM_LOCK_PR; } /* * A canceled lock request. The lock was just taken off the delayed * list and was never even submitted to dlm. */ if (test_and_clear_bit(LFL_CANCEL, &lp->flags)) { log_info("complete internal cancel %x,%llx", lp->lockname.ln_type, (unsigned long long)lp->lockname.ln_number); lp->req = lp->cur; acb.lc_ret |= LM_OUT_CANCELED; goto out; } /* * An error occured. */ if (lp->lksb.sb_status) { /* a "normal" error */ if ((lp->lksb.sb_status == -EAGAIN) && (lp->lkf & DLM_LKF_NOQUEUE)) { lp->req = lp->cur; if (lp->cur == DLM_LOCK_IV) lp->lksb.sb_lkid = 0; goto out; } /* this could only happen with cancels I think */ log_info("ast sb_status %d %x,%llx flags %lx", lp->lksb.sb_status, lp->lockname.ln_type, (unsigned long long)lp->lockname.ln_number, lp->flags); return; } /* * This is an AST for an EX->EX conversion for sync_lvb from GFS. */ if (test_and_clear_bit(LFL_SYNC_LVB, &lp->flags)) { complete(&lp->ast_wait); return; } /* * A lock has been demoted to NL because it initially completed during * BLOCK_LOCKS. Now it must be requested in the originally requested * mode. */ if (test_and_clear_bit(LFL_REREQUEST, &lp->flags)) { gdlm_assert(lp->req == DLM_LOCK_NL, "%x,%llx", lp->lockname.ln_type, (unsigned long long)lp->lockname.ln_number); gdlm_assert(lp->prev_req > DLM_LOCK_NL, "%x,%llx", lp->lockname.ln_type, (unsigned long long)lp->lockname.ln_number); lp->cur = DLM_LOCK_NL; lp->req = lp->prev_req; lp->prev_req = DLM_LOCK_IV; lp->lkf &= ~DLM_LKF_CONVDEADLK; set_bit(LFL_NOCACHE, &lp->flags); if (test_bit(DFL_BLOCK_LOCKS, &ls->flags) && !test_bit(LFL_NOBLOCK, &lp->flags)) gdlm_queue_delayed(lp); else queue_submit(lp); return; } /* * A request is granted during dlm recovery. It may be granted * because the locks of a failed node were cleared. In that case, * there may be inconsistent data beneath this lock and we must wait * for recovery to complete to use it. When gfs recovery is done this * granted lock will be converted to NL and then reacquired in this * granted state. */ if (test_bit(DFL_BLOCK_LOCKS, &ls->flags) && !test_bit(LFL_NOBLOCK, &lp->flags) && lp->req != DLM_LOCK_NL) { lp->cur = lp->req; lp->prev_req = lp->req; lp->req = DLM_LOCK_NL; lp->lkf |= DLM_LKF_CONVERT; lp->lkf &= ~DLM_LKF_CONVDEADLK; log_debug("rereq %x,%llx id %x %d,%d", lp->lockname.ln_type, (unsigned long long)lp->lockname.ln_number, lp->lksb.sb_lkid, lp->cur, lp->req); set_bit(LFL_REREQUEST, &lp->flags); queue_submit(lp); return; } /* * DLM demoted the lock to NL before it was granted so GFS must be * told it cannot cache data for this lock. */ if (lp->lksb.sb_flags & DLM_SBF_DEMOTED) set_bit(LFL_NOCACHE, &lp->flags); out: /* * This is an internal lock_dlm lock */ if (test_bit(LFL_INLOCK, &lp->flags)) { clear_bit(LFL_NOBLOCK, &lp->flags); lp->cur = lp->req; complete(&lp->ast_wait); return; } /* * Normal completion of a lock request. Tell GFS it now has the lock. */ clear_bit(LFL_NOBLOCK, &lp->flags); lp->cur = lp->req; acb.lc_name = lp->lockname; acb.lc_ret |= gdlm_make_lmstate(lp->cur); if (!test_and_clear_bit(LFL_NOCACHE, &lp->flags) && (lp->cur > DLM_LOCK_NL) && (prev_mode > DLM_LOCK_NL)) acb.lc_ret |= LM_OUT_CACHEABLE; ls->fscb(ls->fsdata, LM_CB_ASYNC, &acb); } static inline int no_work(struct gdlm_ls *ls, int blocking) { int ret; spin_lock(&ls->async_lock); ret = list_empty(&ls->complete) && list_empty(&ls->submit); if (ret && blocking) ret = list_empty(&ls->blocking); spin_unlock(&ls->async_lock); return ret; } static inline int check_drop(struct gdlm_ls *ls) { if (!ls->drop_locks_count) return 0; if (time_after(jiffies, ls->drop_time + ls->drop_locks_period * HZ)) { ls->drop_time = jiffies; if (ls->all_locks_count >= ls->drop_locks_count) return 1; } return 0; } static int gdlm_thread(void *data) { struct gdlm_ls *ls = (struct gdlm_ls *) data; struct gdlm_lock *lp = NULL; int blist = 0; uint8_t complete, blocking, submit, drop; DECLARE_WAITQUEUE(wait, current); /* Only thread1 is allowed to do blocking callbacks since gfs may wait for a completion callback within a blocking cb. */ if (current == ls->thread1) blist = 1; while (!kthread_should_stop()) { set_current_state(TASK_INTERRUPTIBLE); add_wait_queue(&ls->thread_wait, &wait); if (no_work(ls, blist)) schedule(); remove_wait_queue(&ls->thread_wait, &wait); set_current_state(TASK_RUNNING); complete = blocking = submit = drop = 0; spin_lock(&ls->async_lock); if (blist && !list_empty(&ls->blocking)) { lp = list_entry(ls->blocking.next, struct gdlm_lock, blist); list_del_init(&lp->blist); blocking = lp->bast_mode; lp->bast_mode = 0; } else if (!list_empty(&ls->complete)) { lp = list_entry(ls->complete.next, struct gdlm_lock, clist); list_del_init(&lp->clist); complete = 1; } else if (!list_empty(&ls->submit)) { lp = list_entry(ls->submit.next, struct gdlm_lock, delay_list); list_del_init(&lp->delay_list); submit = 1; } drop = check_drop(ls); spin_unlock(&ls->async_lock); if (complete) process_complete(lp); else if (blocking) process_blocking(lp, blocking); else if (submit) gdlm_do_lock(lp); if (drop) ls->fscb(ls->fsdata, LM_CB_DROPLOCKS, NULL); schedule(); } return 0; } int gdlm_init_threads(struct gdlm_ls *ls) { struct task_struct *p; int error; p = kthread_run(gdlm_thread, ls, "lock_dlm1"); error = IS_ERR(p); if (error) { log_error("can't start lock_dlm1 thread %d", error); return error; } ls->thread1 = p; p = kthread_run(gdlm_thread, ls, "lock_dlm2"); error = IS_ERR(p); if (error) { log_error("can't start lock_dlm2 thread %d", error); kthread_stop(ls->thread1); return error; } ls->thread2 = p; return 0; } void gdlm_release_threads(struct gdlm_ls *ls) { kthread_stop(ls->thread1); kthread_stop(ls->thread2); }