staging: lustre: fid: use wait_event_cmd()

Rather than open-coding a wait event loop twice,
use wait_event_cmd() to wait, dropping the spinlock
over schedule().
This does require duplicating part of the wait
condition, but that is just three tests on values that
are in registers or in cache, so the cost is small
and the increased readability is large.

Signed-off-by: NeilBrown <neilb@suse.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
This commit is contained in:
NeilBrown 2018-02-20 13:23:38 +11:00 committed by Greg Kroah-Hartman
parent daa5611c71
commit 40f5bd3501
1 changed files with 20 additions and 48 deletions

View File

@ -192,26 +192,6 @@ static int seq_client_alloc_seq(const struct lu_env *env,
return rc;
}
static int seq_fid_alloc_prep(struct lu_client_seq *seq,
wait_queue_entry_t *link)
{
if (seq->lcs_update) {
add_wait_queue(&seq->lcs_waitq, link);
set_current_state(TASK_UNINTERRUPTIBLE);
spin_unlock(&seq->lcs_lock);
schedule();
spin_lock(&seq->lcs_lock);
remove_wait_queue(&seq->lcs_waitq, link);
set_current_state(TASK_RUNNING);
return -EAGAIN;
}
++seq->lcs_update;
spin_unlock(&seq->lcs_lock);
return 0;
}
static void seq_fid_alloc_fini(struct lu_client_seq *seq)
{
LASSERT(seq->lcs_update == 1);
@ -224,32 +204,34 @@ static void seq_fid_alloc_fini(struct lu_client_seq *seq)
int seq_client_alloc_fid(const struct lu_env *env,
struct lu_client_seq *seq, struct lu_fid *fid)
{
wait_queue_entry_t link;
int rc;
LASSERT(seq);
LASSERT(fid);
init_waitqueue_entry(&link, current);
spin_lock(&seq->lcs_lock);
if (OBD_FAIL_CHECK(OBD_FAIL_SEQ_EXHAUST))
seq->lcs_fid.f_oid = seq->lcs_width;
while (1) {
wait_event_cmd(seq->lcs_waitq,
(!fid_is_zero(&seq->lcs_fid) &&
fid_oid(&seq->lcs_fid) < seq->lcs_width) ||
!seq->lcs_update,
spin_unlock(&seq->lcs_lock),
spin_lock(&seq->lcs_lock));
if (!fid_is_zero(&seq->lcs_fid) &&
fid_oid(&seq->lcs_fid) < seq->lcs_width) {
/* Just bump last allocated fid and return to caller. */
seq->lcs_fid.f_oid += 1;
rc = 0;
} else {
u64 seqnr;
if (!fid_is_zero(&seq->lcs_fid) &&
fid_oid(&seq->lcs_fid) < seq->lcs_width) {
/* Just bump last allocated fid and return to caller. */
seq->lcs_fid.f_oid += 1;
rc = 0;
break;
}
rc = seq_fid_alloc_prep(seq, &link);
if (rc)
continue;
LASSERT(seq->lcs_update == 0);
++seq->lcs_update;
spin_unlock(&seq->lcs_lock);
rc = seq_client_alloc_seq(env, seq, &seqnr);
if (rc) {
@ -274,7 +256,6 @@ int seq_client_alloc_fid(const struct lu_env *env,
rc = 1;
seq_fid_alloc_fini(seq);
break;
}
*fid = seq->lcs_fid;
@ -292,23 +273,14 @@ EXPORT_SYMBOL(seq_client_alloc_fid);
*/
void seq_client_flush(struct lu_client_seq *seq)
{
wait_queue_entry_t link;
LASSERT(seq);
init_waitqueue_entry(&link, current);
spin_lock(&seq->lcs_lock);
while (seq->lcs_update) {
add_wait_queue(&seq->lcs_waitq, &link);
set_current_state(TASK_UNINTERRUPTIBLE);
spin_unlock(&seq->lcs_lock);
schedule();
spin_lock(&seq->lcs_lock);
remove_wait_queue(&seq->lcs_waitq, &link);
set_current_state(TASK_RUNNING);
}
wait_event_cmd(seq->lcs_waitq,
!seq->lcs_update,
spin_unlock(&seq->lcs_lock),
spin_lock(&seq->lcs_lock));
fid_zero(&seq->lcs_fid);
/**