mirror of https://gitee.com/openkylin/linux.git
Merge branch 'mlx4_bond_notify'
Or Gerlitz says: ==================== bonding and mlx4 fixes for the HA/LAG support and mlx4 reset flow There are two fixes to the boding + mlx4 HA/LAG support from Moni and a patch from Yishai which does further hardening of the mlx4 reset support for IB kernel ULPs. ==================== Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
commit
46857b5771
|
@ -188,6 +188,8 @@ struct ib_cq *mlx4_ib_create_cq(struct ib_device *ibdev, int entries, int vector
|
||||||
spin_lock_init(&cq->lock);
|
spin_lock_init(&cq->lock);
|
||||||
cq->resize_buf = NULL;
|
cq->resize_buf = NULL;
|
||||||
cq->resize_umem = NULL;
|
cq->resize_umem = NULL;
|
||||||
|
INIT_LIST_HEAD(&cq->send_qp_list);
|
||||||
|
INIT_LIST_HEAD(&cq->recv_qp_list);
|
||||||
|
|
||||||
if (context) {
|
if (context) {
|
||||||
struct mlx4_ib_create_cq ucmd;
|
struct mlx4_ib_create_cq ucmd;
|
||||||
|
@ -594,6 +596,55 @@ static int use_tunnel_data(struct mlx4_ib_qp *qp, struct mlx4_ib_cq *cq, struct
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void mlx4_ib_qp_sw_comp(struct mlx4_ib_qp *qp, int num_entries,
|
||||||
|
struct ib_wc *wc, int *npolled, int is_send)
|
||||||
|
{
|
||||||
|
struct mlx4_ib_wq *wq;
|
||||||
|
unsigned cur;
|
||||||
|
int i;
|
||||||
|
|
||||||
|
wq = is_send ? &qp->sq : &qp->rq;
|
||||||
|
cur = wq->head - wq->tail;
|
||||||
|
|
||||||
|
if (cur == 0)
|
||||||
|
return;
|
||||||
|
|
||||||
|
for (i = 0; i < cur && *npolled < num_entries; i++) {
|
||||||
|
wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)];
|
||||||
|
wc->status = IB_WC_WR_FLUSH_ERR;
|
||||||
|
wc->vendor_err = MLX4_CQE_SYNDROME_WR_FLUSH_ERR;
|
||||||
|
wq->tail++;
|
||||||
|
(*npolled)++;
|
||||||
|
wc->qp = &qp->ibqp;
|
||||||
|
wc++;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
static void mlx4_ib_poll_sw_comp(struct mlx4_ib_cq *cq, int num_entries,
|
||||||
|
struct ib_wc *wc, int *npolled)
|
||||||
|
{
|
||||||
|
struct mlx4_ib_qp *qp;
|
||||||
|
|
||||||
|
*npolled = 0;
|
||||||
|
/* Find uncompleted WQEs belonging to that cq and retrun
|
||||||
|
* simulated FLUSH_ERR completions
|
||||||
|
*/
|
||||||
|
list_for_each_entry(qp, &cq->send_qp_list, cq_send_list) {
|
||||||
|
mlx4_ib_qp_sw_comp(qp, num_entries, wc, npolled, 1);
|
||||||
|
if (*npolled >= num_entries)
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
list_for_each_entry(qp, &cq->recv_qp_list, cq_recv_list) {
|
||||||
|
mlx4_ib_qp_sw_comp(qp, num_entries, wc + *npolled, npolled, 0);
|
||||||
|
if (*npolled >= num_entries)
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
|
out:
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
static int mlx4_ib_poll_one(struct mlx4_ib_cq *cq,
|
static int mlx4_ib_poll_one(struct mlx4_ib_cq *cq,
|
||||||
struct mlx4_ib_qp **cur_qp,
|
struct mlx4_ib_qp **cur_qp,
|
||||||
struct ib_wc *wc)
|
struct ib_wc *wc)
|
||||||
|
@ -836,8 +887,13 @@ int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
int npolled;
|
int npolled;
|
||||||
int err = 0;
|
int err = 0;
|
||||||
|
struct mlx4_ib_dev *mdev = to_mdev(cq->ibcq.device);
|
||||||
|
|
||||||
spin_lock_irqsave(&cq->lock, flags);
|
spin_lock_irqsave(&cq->lock, flags);
|
||||||
|
if (mdev->dev->persist->state & MLX4_DEVICE_STATE_INTERNAL_ERROR) {
|
||||||
|
mlx4_ib_poll_sw_comp(cq, num_entries, wc, &npolled);
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
for (npolled = 0; npolled < num_entries; ++npolled) {
|
for (npolled = 0; npolled < num_entries; ++npolled) {
|
||||||
err = mlx4_ib_poll_one(cq, &cur_qp, wc + npolled);
|
err = mlx4_ib_poll_one(cq, &cur_qp, wc + npolled);
|
||||||
|
@ -847,6 +903,7 @@ int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
|
||||||
|
|
||||||
mlx4_cq_set_ci(&cq->mcq);
|
mlx4_cq_set_ci(&cq->mcq);
|
||||||
|
|
||||||
|
out:
|
||||||
spin_unlock_irqrestore(&cq->lock, flags);
|
spin_unlock_irqrestore(&cq->lock, flags);
|
||||||
|
|
||||||
if (err == 0 || err == -EAGAIN)
|
if (err == 0 || err == -EAGAIN)
|
||||||
|
|
|
@ -1186,6 +1186,9 @@ static struct ib_flow *mlx4_ib_create_flow(struct ib_qp *qp,
|
||||||
goto err_create_flow;
|
goto err_create_flow;
|
||||||
i++;
|
i++;
|
||||||
if (is_bonded) {
|
if (is_bonded) {
|
||||||
|
/* Application always sees one port so the mirror rule
|
||||||
|
* must be on port #2
|
||||||
|
*/
|
||||||
flow_attr->port = 2;
|
flow_attr->port = 2;
|
||||||
err = __mlx4_ib_create_flow(qp, flow_attr,
|
err = __mlx4_ib_create_flow(qp, flow_attr,
|
||||||
domain, type[j],
|
domain, type[j],
|
||||||
|
@ -1286,7 +1289,8 @@ static int mlx4_ib_mcg_attach(struct ib_qp *ibqp, union ib_gid *gid, u16 lid)
|
||||||
|
|
||||||
reg_id.mirror = 0;
|
reg_id.mirror = 0;
|
||||||
if (mlx4_is_bonded(dev)) {
|
if (mlx4_is_bonded(dev)) {
|
||||||
err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw, 2,
|
err = mlx4_multicast_attach(mdev->dev, &mqp->mqp, gid->raw,
|
||||||
|
(mqp->port == 1) ? 2 : 1,
|
||||||
!!(mqp->flags &
|
!!(mqp->flags &
|
||||||
MLX4_IB_QP_BLOCK_MULTICAST_LOOPBACK),
|
MLX4_IB_QP_BLOCK_MULTICAST_LOOPBACK),
|
||||||
prot, ®_id.mirror);
|
prot, ®_id.mirror);
|
||||||
|
@ -2304,6 +2308,8 @@ static void *mlx4_ib_add(struct mlx4_dev *dev)
|
||||||
|
|
||||||
spin_lock_init(&ibdev->sm_lock);
|
spin_lock_init(&ibdev->sm_lock);
|
||||||
mutex_init(&ibdev->cap_mask_mutex);
|
mutex_init(&ibdev->cap_mask_mutex);
|
||||||
|
INIT_LIST_HEAD(&ibdev->qp_list);
|
||||||
|
spin_lock_init(&ibdev->reset_flow_resource_lock);
|
||||||
|
|
||||||
if (ibdev->steering_support == MLX4_STEERING_MODE_DEVICE_MANAGED &&
|
if (ibdev->steering_support == MLX4_STEERING_MODE_DEVICE_MANAGED &&
|
||||||
ib_num_ports) {
|
ib_num_ports) {
|
||||||
|
@ -2618,6 +2624,67 @@ static void do_slave_init(struct mlx4_ib_dev *ibdev, int slave, int do_init)
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void mlx4_ib_handle_catas_error(struct mlx4_ib_dev *ibdev)
|
||||||
|
{
|
||||||
|
struct mlx4_ib_qp *mqp;
|
||||||
|
unsigned long flags_qp;
|
||||||
|
unsigned long flags_cq;
|
||||||
|
struct mlx4_ib_cq *send_mcq, *recv_mcq;
|
||||||
|
struct list_head cq_notify_list;
|
||||||
|
struct mlx4_cq *mcq;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
|
pr_warn("mlx4_ib_handle_catas_error was started\n");
|
||||||
|
INIT_LIST_HEAD(&cq_notify_list);
|
||||||
|
|
||||||
|
/* Go over qp list reside on that ibdev, sync with create/destroy qp.*/
|
||||||
|
spin_lock_irqsave(&ibdev->reset_flow_resource_lock, flags);
|
||||||
|
|
||||||
|
list_for_each_entry(mqp, &ibdev->qp_list, qps_list) {
|
||||||
|
spin_lock_irqsave(&mqp->sq.lock, flags_qp);
|
||||||
|
if (mqp->sq.tail != mqp->sq.head) {
|
||||||
|
send_mcq = to_mcq(mqp->ibqp.send_cq);
|
||||||
|
spin_lock_irqsave(&send_mcq->lock, flags_cq);
|
||||||
|
if (send_mcq->mcq.comp &&
|
||||||
|
mqp->ibqp.send_cq->comp_handler) {
|
||||||
|
if (!send_mcq->mcq.reset_notify_added) {
|
||||||
|
send_mcq->mcq.reset_notify_added = 1;
|
||||||
|
list_add_tail(&send_mcq->mcq.reset_notify,
|
||||||
|
&cq_notify_list);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
spin_unlock_irqrestore(&send_mcq->lock, flags_cq);
|
||||||
|
}
|
||||||
|
spin_unlock_irqrestore(&mqp->sq.lock, flags_qp);
|
||||||
|
/* Now, handle the QP's receive queue */
|
||||||
|
spin_lock_irqsave(&mqp->rq.lock, flags_qp);
|
||||||
|
/* no handling is needed for SRQ */
|
||||||
|
if (!mqp->ibqp.srq) {
|
||||||
|
if (mqp->rq.tail != mqp->rq.head) {
|
||||||
|
recv_mcq = to_mcq(mqp->ibqp.recv_cq);
|
||||||
|
spin_lock_irqsave(&recv_mcq->lock, flags_cq);
|
||||||
|
if (recv_mcq->mcq.comp &&
|
||||||
|
mqp->ibqp.recv_cq->comp_handler) {
|
||||||
|
if (!recv_mcq->mcq.reset_notify_added) {
|
||||||
|
recv_mcq->mcq.reset_notify_added = 1;
|
||||||
|
list_add_tail(&recv_mcq->mcq.reset_notify,
|
||||||
|
&cq_notify_list);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
spin_unlock_irqrestore(&recv_mcq->lock,
|
||||||
|
flags_cq);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
spin_unlock_irqrestore(&mqp->rq.lock, flags_qp);
|
||||||
|
}
|
||||||
|
|
||||||
|
list_for_each_entry(mcq, &cq_notify_list, reset_notify) {
|
||||||
|
mcq->comp(mcq);
|
||||||
|
}
|
||||||
|
spin_unlock_irqrestore(&ibdev->reset_flow_resource_lock, flags);
|
||||||
|
pr_warn("mlx4_ib_handle_catas_error ended\n");
|
||||||
|
}
|
||||||
|
|
||||||
static void handle_bonded_port_state_event(struct work_struct *work)
|
static void handle_bonded_port_state_event(struct work_struct *work)
|
||||||
{
|
{
|
||||||
struct ib_event_work *ew =
|
struct ib_event_work *ew =
|
||||||
|
@ -2697,6 +2764,7 @@ static void mlx4_ib_event(struct mlx4_dev *dev, void *ibdev_ptr,
|
||||||
case MLX4_DEV_EVENT_CATASTROPHIC_ERROR:
|
case MLX4_DEV_EVENT_CATASTROPHIC_ERROR:
|
||||||
ibdev->ib_active = false;
|
ibdev->ib_active = false;
|
||||||
ibev.event = IB_EVENT_DEVICE_FATAL;
|
ibev.event = IB_EVENT_DEVICE_FATAL;
|
||||||
|
mlx4_ib_handle_catas_error(ibdev);
|
||||||
break;
|
break;
|
||||||
|
|
||||||
case MLX4_DEV_EVENT_PORT_MGMT_CHANGE:
|
case MLX4_DEV_EVENT_PORT_MGMT_CHANGE:
|
||||||
|
|
|
@ -110,6 +110,9 @@ struct mlx4_ib_cq {
|
||||||
struct mutex resize_mutex;
|
struct mutex resize_mutex;
|
||||||
struct ib_umem *umem;
|
struct ib_umem *umem;
|
||||||
struct ib_umem *resize_umem;
|
struct ib_umem *resize_umem;
|
||||||
|
/* List of qps that it serves.*/
|
||||||
|
struct list_head send_qp_list;
|
||||||
|
struct list_head recv_qp_list;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct mlx4_ib_mr {
|
struct mlx4_ib_mr {
|
||||||
|
@ -300,6 +303,9 @@ struct mlx4_ib_qp {
|
||||||
struct mlx4_roce_smac_vlan_info pri;
|
struct mlx4_roce_smac_vlan_info pri;
|
||||||
struct mlx4_roce_smac_vlan_info alt;
|
struct mlx4_roce_smac_vlan_info alt;
|
||||||
u64 reg_id;
|
u64 reg_id;
|
||||||
|
struct list_head qps_list;
|
||||||
|
struct list_head cq_recv_list;
|
||||||
|
struct list_head cq_send_list;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct mlx4_ib_srq {
|
struct mlx4_ib_srq {
|
||||||
|
@ -535,6 +541,9 @@ struct mlx4_ib_dev {
|
||||||
/* lock when destroying qp1_proxy and getting netdev events */
|
/* lock when destroying qp1_proxy and getting netdev events */
|
||||||
struct mutex qp1_proxy_lock[MLX4_MAX_PORTS];
|
struct mutex qp1_proxy_lock[MLX4_MAX_PORTS];
|
||||||
u8 bond_next_port;
|
u8 bond_next_port;
|
||||||
|
/* protect resources needed as part of reset flow */
|
||||||
|
spinlock_t reset_flow_resource_lock;
|
||||||
|
struct list_head qp_list;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct ib_event_work {
|
struct ib_event_work {
|
||||||
|
|
|
@ -46,6 +46,11 @@
|
||||||
#include "mlx4_ib.h"
|
#include "mlx4_ib.h"
|
||||||
#include "user.h"
|
#include "user.h"
|
||||||
|
|
||||||
|
static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq,
|
||||||
|
struct mlx4_ib_cq *recv_cq);
|
||||||
|
static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq,
|
||||||
|
struct mlx4_ib_cq *recv_cq);
|
||||||
|
|
||||||
enum {
|
enum {
|
||||||
MLX4_IB_ACK_REQ_FREQ = 8,
|
MLX4_IB_ACK_REQ_FREQ = 8,
|
||||||
};
|
};
|
||||||
|
@ -618,6 +623,8 @@ static int create_qp_common(struct mlx4_ib_dev *dev, struct ib_pd *pd,
|
||||||
struct mlx4_ib_sqp *sqp;
|
struct mlx4_ib_sqp *sqp;
|
||||||
struct mlx4_ib_qp *qp;
|
struct mlx4_ib_qp *qp;
|
||||||
enum mlx4_ib_qp_type qp_type = (enum mlx4_ib_qp_type) init_attr->qp_type;
|
enum mlx4_ib_qp_type qp_type = (enum mlx4_ib_qp_type) init_attr->qp_type;
|
||||||
|
struct mlx4_ib_cq *mcq;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
/* When tunneling special qps, we use a plain UD qp */
|
/* When tunneling special qps, we use a plain UD qp */
|
||||||
if (sqpn) {
|
if (sqpn) {
|
||||||
|
@ -828,6 +835,24 @@ static int create_qp_common(struct mlx4_ib_dev *dev, struct ib_pd *pd,
|
||||||
qp->mqp.event = mlx4_ib_qp_event;
|
qp->mqp.event = mlx4_ib_qp_event;
|
||||||
if (!*caller_qp)
|
if (!*caller_qp)
|
||||||
*caller_qp = qp;
|
*caller_qp = qp;
|
||||||
|
|
||||||
|
spin_lock_irqsave(&dev->reset_flow_resource_lock, flags);
|
||||||
|
mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq),
|
||||||
|
to_mcq(init_attr->recv_cq));
|
||||||
|
/* Maintain device to QPs access, needed for further handling
|
||||||
|
* via reset flow
|
||||||
|
*/
|
||||||
|
list_add_tail(&qp->qps_list, &dev->qp_list);
|
||||||
|
/* Maintain CQ to QPs access, needed for further handling
|
||||||
|
* via reset flow
|
||||||
|
*/
|
||||||
|
mcq = to_mcq(init_attr->send_cq);
|
||||||
|
list_add_tail(&qp->cq_send_list, &mcq->send_qp_list);
|
||||||
|
mcq = to_mcq(init_attr->recv_cq);
|
||||||
|
list_add_tail(&qp->cq_recv_list, &mcq->recv_qp_list);
|
||||||
|
mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq),
|
||||||
|
to_mcq(init_attr->recv_cq));
|
||||||
|
spin_unlock_irqrestore(&dev->reset_flow_resource_lock, flags);
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err_qpn:
|
err_qpn:
|
||||||
|
@ -886,13 +911,13 @@ static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv
|
||||||
__acquires(&send_cq->lock) __acquires(&recv_cq->lock)
|
__acquires(&send_cq->lock) __acquires(&recv_cq->lock)
|
||||||
{
|
{
|
||||||
if (send_cq == recv_cq) {
|
if (send_cq == recv_cq) {
|
||||||
spin_lock_irq(&send_cq->lock);
|
spin_lock(&send_cq->lock);
|
||||||
__acquire(&recv_cq->lock);
|
__acquire(&recv_cq->lock);
|
||||||
} else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
|
} else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
|
||||||
spin_lock_irq(&send_cq->lock);
|
spin_lock(&send_cq->lock);
|
||||||
spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING);
|
spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING);
|
||||||
} else {
|
} else {
|
||||||
spin_lock_irq(&recv_cq->lock);
|
spin_lock(&recv_cq->lock);
|
||||||
spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
|
spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -902,13 +927,13 @@ static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *re
|
||||||
{
|
{
|
||||||
if (send_cq == recv_cq) {
|
if (send_cq == recv_cq) {
|
||||||
__release(&recv_cq->lock);
|
__release(&recv_cq->lock);
|
||||||
spin_unlock_irq(&send_cq->lock);
|
spin_unlock(&send_cq->lock);
|
||||||
} else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
|
} else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) {
|
||||||
spin_unlock(&recv_cq->lock);
|
spin_unlock(&recv_cq->lock);
|
||||||
spin_unlock_irq(&send_cq->lock);
|
spin_unlock(&send_cq->lock);
|
||||||
} else {
|
} else {
|
||||||
spin_unlock(&send_cq->lock);
|
spin_unlock(&send_cq->lock);
|
||||||
spin_unlock_irq(&recv_cq->lock);
|
spin_unlock(&recv_cq->lock);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -953,6 +978,7 @@ static void destroy_qp_common(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp,
|
||||||
int is_user)
|
int is_user)
|
||||||
{
|
{
|
||||||
struct mlx4_ib_cq *send_cq, *recv_cq;
|
struct mlx4_ib_cq *send_cq, *recv_cq;
|
||||||
|
unsigned long flags;
|
||||||
|
|
||||||
if (qp->state != IB_QPS_RESET) {
|
if (qp->state != IB_QPS_RESET) {
|
||||||
if (mlx4_qp_modify(dev->dev, NULL, to_mlx4_state(qp->state),
|
if (mlx4_qp_modify(dev->dev, NULL, to_mlx4_state(qp->state),
|
||||||
|
@ -984,8 +1010,13 @@ static void destroy_qp_common(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp,
|
||||||
|
|
||||||
get_cqs(qp, &send_cq, &recv_cq);
|
get_cqs(qp, &send_cq, &recv_cq);
|
||||||
|
|
||||||
|
spin_lock_irqsave(&dev->reset_flow_resource_lock, flags);
|
||||||
mlx4_ib_lock_cqs(send_cq, recv_cq);
|
mlx4_ib_lock_cqs(send_cq, recv_cq);
|
||||||
|
|
||||||
|
/* del from lists under both locks above to protect reset flow paths */
|
||||||
|
list_del(&qp->qps_list);
|
||||||
|
list_del(&qp->cq_send_list);
|
||||||
|
list_del(&qp->cq_recv_list);
|
||||||
if (!is_user) {
|
if (!is_user) {
|
||||||
__mlx4_ib_cq_clean(recv_cq, qp->mqp.qpn,
|
__mlx4_ib_cq_clean(recv_cq, qp->mqp.qpn,
|
||||||
qp->ibqp.srq ? to_msrq(qp->ibqp.srq): NULL);
|
qp->ibqp.srq ? to_msrq(qp->ibqp.srq): NULL);
|
||||||
|
@ -996,6 +1027,7 @@ static void destroy_qp_common(struct mlx4_ib_dev *dev, struct mlx4_ib_qp *qp,
|
||||||
mlx4_qp_remove(dev->dev, &qp->mqp);
|
mlx4_qp_remove(dev->dev, &qp->mqp);
|
||||||
|
|
||||||
mlx4_ib_unlock_cqs(send_cq, recv_cq);
|
mlx4_ib_unlock_cqs(send_cq, recv_cq);
|
||||||
|
spin_unlock_irqrestore(&dev->reset_flow_resource_lock, flags);
|
||||||
|
|
||||||
mlx4_qp_free(dev->dev, &qp->mqp);
|
mlx4_qp_free(dev->dev, &qp->mqp);
|
||||||
|
|
||||||
|
@ -2618,8 +2650,15 @@ int mlx4_ib_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
|
||||||
__be32 uninitialized_var(lso_hdr_sz);
|
__be32 uninitialized_var(lso_hdr_sz);
|
||||||
__be32 blh;
|
__be32 blh;
|
||||||
int i;
|
int i;
|
||||||
|
struct mlx4_ib_dev *mdev = to_mdev(ibqp->device);
|
||||||
|
|
||||||
spin_lock_irqsave(&qp->sq.lock, flags);
|
spin_lock_irqsave(&qp->sq.lock, flags);
|
||||||
|
if (mdev->dev->persist->state & MLX4_DEVICE_STATE_INTERNAL_ERROR) {
|
||||||
|
err = -EIO;
|
||||||
|
*bad_wr = wr;
|
||||||
|
nreq = 0;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
ind = qp->sq_next_wqe;
|
ind = qp->sq_next_wqe;
|
||||||
|
|
||||||
|
@ -2917,10 +2956,18 @@ int mlx4_ib_post_recv(struct ib_qp *ibqp, struct ib_recv_wr *wr,
|
||||||
int ind;
|
int ind;
|
||||||
int max_gs;
|
int max_gs;
|
||||||
int i;
|
int i;
|
||||||
|
struct mlx4_ib_dev *mdev = to_mdev(ibqp->device);
|
||||||
|
|
||||||
max_gs = qp->rq.max_gs;
|
max_gs = qp->rq.max_gs;
|
||||||
spin_lock_irqsave(&qp->rq.lock, flags);
|
spin_lock_irqsave(&qp->rq.lock, flags);
|
||||||
|
|
||||||
|
if (mdev->dev->persist->state & MLX4_DEVICE_STATE_INTERNAL_ERROR) {
|
||||||
|
err = -EIO;
|
||||||
|
*bad_wr = wr;
|
||||||
|
nreq = 0;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
ind = qp->rq.head & (qp->rq.wqe_cnt - 1);
|
ind = qp->rq.head & (qp->rq.wqe_cnt - 1);
|
||||||
|
|
||||||
for (nreq = 0; wr; ++nreq, wr = wr->next) {
|
for (nreq = 0; wr; ++nreq, wr = wr->next) {
|
||||||
|
|
|
@ -316,8 +316,15 @@ int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr,
|
||||||
int err = 0;
|
int err = 0;
|
||||||
int nreq;
|
int nreq;
|
||||||
int i;
|
int i;
|
||||||
|
struct mlx4_ib_dev *mdev = to_mdev(ibsrq->device);
|
||||||
|
|
||||||
spin_lock_irqsave(&srq->lock, flags);
|
spin_lock_irqsave(&srq->lock, flags);
|
||||||
|
if (mdev->dev->persist->state & MLX4_DEVICE_STATE_INTERNAL_ERROR) {
|
||||||
|
err = -EIO;
|
||||||
|
*bad_wr = wr;
|
||||||
|
nreq = 0;
|
||||||
|
goto out;
|
||||||
|
}
|
||||||
|
|
||||||
for (nreq = 0; wr; ++nreq, wr = wr->next) {
|
for (nreq = 0; wr; ++nreq, wr = wr->next) {
|
||||||
if (unlikely(wr->num_sge > srq->msrq.max_gs)) {
|
if (unlikely(wr->num_sge > srq->msrq.max_gs)) {
|
||||||
|
@ -362,6 +369,7 @@ int mlx4_ib_post_srq_recv(struct ib_srq *ibsrq, struct ib_recv_wr *wr,
|
||||||
|
|
||||||
*srq->db.db = cpu_to_be32(srq->wqe_ctr);
|
*srq->db.db = cpu_to_be32(srq->wqe_ctr);
|
||||||
}
|
}
|
||||||
|
out:
|
||||||
|
|
||||||
spin_unlock_irqrestore(&srq->lock, flags);
|
spin_unlock_irqrestore(&srq->lock, flags);
|
||||||
|
|
||||||
|
|
|
@ -1196,18 +1196,11 @@ static void bond_fill_ifslave(struct slave *slave, struct ifslave *info)
|
||||||
info->link_failure_count = slave->link_failure_count;
|
info->link_failure_count = slave->link_failure_count;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bond_netdev_notify(struct slave *slave, struct net_device *dev)
|
static void bond_netdev_notify(struct net_device *dev,
|
||||||
|
struct netdev_bonding_info *info)
|
||||||
{
|
{
|
||||||
struct bonding *bond = slave->bond;
|
|
||||||
struct netdev_bonding_info bonding_info;
|
|
||||||
|
|
||||||
rtnl_lock();
|
rtnl_lock();
|
||||||
/* make sure that slave is still valid */
|
netdev_bonding_info_change(dev, info);
|
||||||
if (dev->priv_flags & IFF_BONDING) {
|
|
||||||
bond_fill_ifslave(slave, &bonding_info.slave);
|
|
||||||
bond_fill_ifbond(bond, &bonding_info.master);
|
|
||||||
netdev_bonding_info_change(slave->dev, &bonding_info);
|
|
||||||
}
|
|
||||||
rtnl_unlock();
|
rtnl_unlock();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1216,25 +1209,26 @@ static void bond_netdev_notify_work(struct work_struct *_work)
|
||||||
struct netdev_notify_work *w =
|
struct netdev_notify_work *w =
|
||||||
container_of(_work, struct netdev_notify_work, work.work);
|
container_of(_work, struct netdev_notify_work, work.work);
|
||||||
|
|
||||||
bond_netdev_notify(w->slave, w->dev);
|
bond_netdev_notify(w->dev, &w->bonding_info);
|
||||||
dev_put(w->dev);
|
dev_put(w->dev);
|
||||||
|
kfree(w);
|
||||||
}
|
}
|
||||||
|
|
||||||
void bond_queue_slave_event(struct slave *slave)
|
void bond_queue_slave_event(struct slave *slave)
|
||||||
{
|
{
|
||||||
|
struct bonding *bond = slave->bond;
|
||||||
struct netdev_notify_work *nnw = kzalloc(sizeof(*nnw), GFP_ATOMIC);
|
struct netdev_notify_work *nnw = kzalloc(sizeof(*nnw), GFP_ATOMIC);
|
||||||
|
|
||||||
if (!nnw)
|
if (!nnw)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
INIT_DELAYED_WORK(&nnw->work, bond_netdev_notify_work);
|
dev_hold(slave->dev);
|
||||||
nnw->slave = slave;
|
|
||||||
nnw->dev = slave->dev;
|
nnw->dev = slave->dev;
|
||||||
|
bond_fill_ifslave(slave, &nnw->bonding_info.slave);
|
||||||
|
bond_fill_ifbond(bond, &nnw->bonding_info.master);
|
||||||
|
INIT_DELAYED_WORK(&nnw->work, bond_netdev_notify_work);
|
||||||
|
|
||||||
if (queue_delayed_work(slave->bond->wq, &nnw->work, 0))
|
queue_delayed_work(slave->bond->wq, &nnw->work, 0);
|
||||||
dev_hold(slave->dev);
|
|
||||||
else
|
|
||||||
kfree(nnw);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* enslave device <slave> to bond device <master> */
|
/* enslave device <slave> to bond device <master> */
|
||||||
|
|
|
@ -689,6 +689,8 @@ struct mlx4_cq {
|
||||||
void (*comp)(struct mlx4_cq *);
|
void (*comp)(struct mlx4_cq *);
|
||||||
void *priv;
|
void *priv;
|
||||||
} tasklet_ctx;
|
} tasklet_ctx;
|
||||||
|
int reset_notify_added;
|
||||||
|
struct list_head reset_notify;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct mlx4_qp {
|
struct mlx4_qp {
|
||||||
|
|
|
@ -152,8 +152,8 @@ struct bond_parm_tbl {
|
||||||
|
|
||||||
struct netdev_notify_work {
|
struct netdev_notify_work {
|
||||||
struct delayed_work work;
|
struct delayed_work work;
|
||||||
struct slave *slave;
|
|
||||||
struct net_device *dev;
|
struct net_device *dev;
|
||||||
|
struct netdev_bonding_info bonding_info;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct slave {
|
struct slave {
|
||||||
|
|
Loading…
Reference in New Issue