RDMA/cm: Pull duplicated code into cm_queue_work_unlock()

While unlocking a spinlock held by the caller is a disturbing pattern,
this extensively duplicated code is even worse. Pull all the duplicates
into a function and explain the purpose of the algorithm.

The on creation side call in cm_req_handler() which is different has been
micro-optimized on the basis that the work_count == -1 during creation,
remove that and just use the normal function.

Link: https://lore.kernel.org/r/20200506074701.9775-5-leon@kernel.org
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
This commit is contained in:
Jason Gunthorpe 2020-05-06 10:46:55 +03:00
parent 42113eed8f
commit e83f195aa4
1 changed files with 44 additions and 102 deletions

View File

@ -81,8 +81,11 @@ const char *__attribute_const__ ibcm_reject_msg(int reason)
EXPORT_SYMBOL(ibcm_reject_msg); EXPORT_SYMBOL(ibcm_reject_msg);
struct cm_id_private; struct cm_id_private;
struct cm_work;
static int cm_add_one(struct ib_device *device); static int cm_add_one(struct ib_device *device);
static void cm_remove_one(struct ib_device *device, void *client_data); static void cm_remove_one(struct ib_device *device, void *client_data);
static void cm_process_work(struct cm_id_private *cm_id_priv,
struct cm_work *work);
static int cm_send_sidr_rep_locked(struct cm_id_private *cm_id_priv, static int cm_send_sidr_rep_locked(struct cm_id_private *cm_id_priv,
struct ib_cm_sidr_rep_param *param); struct ib_cm_sidr_rep_param *param);
static int cm_send_dreq_locked(struct cm_id_private *cm_id_priv, static int cm_send_dreq_locked(struct cm_id_private *cm_id_priv,
@ -907,6 +910,35 @@ static void cm_free_work(struct cm_work *work)
kfree(work); kfree(work);
} }
static void cm_queue_work_unlock(struct cm_id_private *cm_id_priv,
struct cm_work *work)
{
bool immediate;
/*
* To deliver the event to the user callback we have the drop the
* spinlock, however, we need to ensure that the user callback is single
* threaded and receives events in the temporal order. If there are
* already events being processed then thread new events onto a list,
* the thread currently processing will pick them up.
*/
immediate = atomic_inc_and_test(&cm_id_priv->work_count);
if (!immediate) {
list_add_tail(&work->list, &cm_id_priv->work_list);
/*
* This routine always consumes incoming reference. Once queued
* to the work_list then a reference is held by the thread
* currently running cm_process_work() and this reference is not
* needed.
*/
cm_deref_id(cm_id_priv);
}
spin_unlock_irq(&cm_id_priv->lock);
if (immediate)
cm_process_work(cm_id_priv, work);
}
static inline int cm_convert_to_ms(int iba_time) static inline int cm_convert_to_ms(int iba_time)
{ {
/* approximate conversion to ms from 4.096us x 2^iba_time */ /* approximate conversion to ms from 4.096us x 2^iba_time */
@ -2144,9 +2176,7 @@ static int cm_req_handler(struct cm_work *work)
/* Refcount belongs to the event, pairs with cm_process_work() */ /* Refcount belongs to the event, pairs with cm_process_work() */
refcount_inc(&cm_id_priv->refcount); refcount_inc(&cm_id_priv->refcount);
atomic_inc(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
spin_unlock_irq(&cm_id_priv->lock);
cm_process_work(cm_id_priv, work);
/* /*
* Since this ID was just created and was not made visible to other MAD * Since this ID was just created and was not made visible to other MAD
* handlers until the cm_finalize_id() above we know that the * handlers until the cm_finalize_id() above we know that the
@ -2492,15 +2522,7 @@ static int cm_rep_handler(struct cm_work *work)
cm_id_priv->alt_av.timeout - 1); cm_id_priv->alt_av.timeout - 1);
ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg); ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg);
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
error: error:
@ -2511,7 +2533,6 @@ static int cm_rep_handler(struct cm_work *work)
static int cm_establish_handler(struct cm_work *work) static int cm_establish_handler(struct cm_work *work)
{ {
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
int ret;
/* See comment in cm_establish about lookup. */ /* See comment in cm_establish about lookup. */
cm_id_priv = cm_acquire_id(work->local_id, work->remote_id); cm_id_priv = cm_acquire_id(work->local_id, work->remote_id);
@ -2525,15 +2546,7 @@ static int cm_establish_handler(struct cm_work *work)
} }
ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg); ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg);
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
out: out:
cm_deref_id(cm_id_priv); cm_deref_id(cm_id_priv);
@ -2544,7 +2557,6 @@ static int cm_rtu_handler(struct cm_work *work)
{ {
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
struct cm_rtu_msg *rtu_msg; struct cm_rtu_msg *rtu_msg;
int ret;
rtu_msg = (struct cm_rtu_msg *)work->mad_recv_wc->recv_buf.mad; rtu_msg = (struct cm_rtu_msg *)work->mad_recv_wc->recv_buf.mad;
cm_id_priv = cm_acquire_id( cm_id_priv = cm_acquire_id(
@ -2567,15 +2579,7 @@ static int cm_rtu_handler(struct cm_work *work)
cm_id_priv->id.state = IB_CM_ESTABLISHED; cm_id_priv->id.state = IB_CM_ESTABLISHED;
ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg); ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg);
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
out: out:
cm_deref_id(cm_id_priv); cm_deref_id(cm_id_priv);
@ -2768,7 +2772,6 @@ static int cm_dreq_handler(struct cm_work *work)
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
struct cm_dreq_msg *dreq_msg; struct cm_dreq_msg *dreq_msg;
struct ib_mad_send_buf *msg = NULL; struct ib_mad_send_buf *msg = NULL;
int ret;
dreq_msg = (struct cm_dreq_msg *)work->mad_recv_wc->recv_buf.mad; dreq_msg = (struct cm_dreq_msg *)work->mad_recv_wc->recv_buf.mad;
cm_id_priv = cm_acquire_id( cm_id_priv = cm_acquire_id(
@ -2833,15 +2836,7 @@ static int cm_dreq_handler(struct cm_work *work)
} }
cm_id_priv->id.state = IB_CM_DREQ_RCVD; cm_id_priv->id.state = IB_CM_DREQ_RCVD;
cm_id_priv->tid = dreq_msg->hdr.tid; cm_id_priv->tid = dreq_msg->hdr.tid;
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
unlock: spin_unlock_irq(&cm_id_priv->lock); unlock: spin_unlock_irq(&cm_id_priv->lock);
@ -2853,7 +2848,6 @@ static int cm_drep_handler(struct cm_work *work)
{ {
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
struct cm_drep_msg *drep_msg; struct cm_drep_msg *drep_msg;
int ret;
drep_msg = (struct cm_drep_msg *)work->mad_recv_wc->recv_buf.mad; drep_msg = (struct cm_drep_msg *)work->mad_recv_wc->recv_buf.mad;
cm_id_priv = cm_acquire_id( cm_id_priv = cm_acquire_id(
@ -2874,15 +2868,7 @@ static int cm_drep_handler(struct cm_work *work)
cm_enter_timewait(cm_id_priv); cm_enter_timewait(cm_id_priv);
ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg); ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg);
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
out: out:
cm_deref_id(cm_id_priv); cm_deref_id(cm_id_priv);
@ -3010,7 +2996,6 @@ static int cm_rej_handler(struct cm_work *work)
{ {
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
struct cm_rej_msg *rej_msg; struct cm_rej_msg *rej_msg;
int ret;
rej_msg = (struct cm_rej_msg *)work->mad_recv_wc->recv_buf.mad; rej_msg = (struct cm_rej_msg *)work->mad_recv_wc->recv_buf.mad;
cm_id_priv = cm_acquire_rejected_id(rej_msg); cm_id_priv = cm_acquire_rejected_id(rej_msg);
@ -3059,15 +3044,7 @@ static int cm_rej_handler(struct cm_work *work)
goto out; goto out;
} }
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
out: out:
cm_deref_id(cm_id_priv); cm_deref_id(cm_id_priv);
@ -3177,7 +3154,7 @@ static int cm_mra_handler(struct cm_work *work)
{ {
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
struct cm_mra_msg *mra_msg; struct cm_mra_msg *mra_msg;
int timeout, ret; int timeout;
mra_msg = (struct cm_mra_msg *)work->mad_recv_wc->recv_buf.mad; mra_msg = (struct cm_mra_msg *)work->mad_recv_wc->recv_buf.mad;
cm_id_priv = cm_acquire_mraed_id(mra_msg); cm_id_priv = cm_acquire_mraed_id(mra_msg);
@ -3237,15 +3214,7 @@ static int cm_mra_handler(struct cm_work *work)
cm_id_priv->msg->context[1] = (void *) (unsigned long) cm_id_priv->msg->context[1] = (void *) (unsigned long)
cm_id_priv->id.state; cm_id_priv->id.state;
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
out: out:
spin_unlock_irq(&cm_id_priv->lock); spin_unlock_irq(&cm_id_priv->lock);
@ -3380,15 +3349,7 @@ static int cm_lap_handler(struct cm_work *work)
cm_id_priv->id.lap_state = IB_CM_LAP_RCVD; cm_id_priv->id.lap_state = IB_CM_LAP_RCVD;
cm_id_priv->tid = lap_msg->hdr.tid; cm_id_priv->tid = lap_msg->hdr.tid;
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
unlock: spin_unlock_irq(&cm_id_priv->lock); unlock: spin_unlock_irq(&cm_id_priv->lock);
@ -3400,7 +3361,6 @@ static int cm_apr_handler(struct cm_work *work)
{ {
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
struct cm_apr_msg *apr_msg; struct cm_apr_msg *apr_msg;
int ret;
/* Currently Alternate path messages are not supported for /* Currently Alternate path messages are not supported for
* RoCE link layer. * RoCE link layer.
@ -3435,16 +3395,7 @@ static int cm_apr_handler(struct cm_work *work)
cm_id_priv->id.lap_state = IB_CM_LAP_IDLE; cm_id_priv->id.lap_state = IB_CM_LAP_IDLE;
ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg); ib_cancel_mad(cm_id_priv->av.port->mad_agent, cm_id_priv->msg);
cm_id_priv->msg = NULL; cm_id_priv->msg = NULL;
cm_queue_work_unlock(cm_id_priv, work);
ret = atomic_inc_and_test(&cm_id_priv->work_count);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
out: out:
cm_deref_id(cm_id_priv); cm_deref_id(cm_id_priv);
@ -3455,7 +3406,6 @@ static int cm_timewait_handler(struct cm_work *work)
{ {
struct cm_timewait_info *timewait_info; struct cm_timewait_info *timewait_info;
struct cm_id_private *cm_id_priv; struct cm_id_private *cm_id_priv;
int ret;
timewait_info = container_of(work, struct cm_timewait_info, work); timewait_info = container_of(work, struct cm_timewait_info, work);
spin_lock_irq(&cm.lock); spin_lock_irq(&cm.lock);
@ -3474,15 +3424,7 @@ static int cm_timewait_handler(struct cm_work *work)
goto out; goto out;
} }
cm_id_priv->id.state = IB_CM_IDLE; cm_id_priv->id.state = IB_CM_IDLE;
ret = atomic_inc_and_test(&cm_id_priv->work_count); cm_queue_work_unlock(cm_id_priv, work);
if (!ret)
list_add_tail(&work->list, &cm_id_priv->work_list);
spin_unlock_irq(&cm_id_priv->lock);
if (ret)
cm_process_work(cm_id_priv, work);
else
cm_deref_id(cm_id_priv);
return 0; return 0;
out: out:
cm_deref_id(cm_id_priv); cm_deref_id(cm_id_priv);