RDMA: Add a dedicated CQ resource tracker function

In order to avoid double multiplexing of the resource when it is a CQ, add
a dedicated callback function.

Link: https://lore.kernel.org/r/20200623113043.1228482-6-leon@kernel.org
Signed-off-by: Maor Gottlieb <maorg@mellanox.com>
Signed-off-by: Leon Romanovsky <leonro@mellanox.com>
Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
This commit is contained in:
Maor Gottlieb 2020-06-23 14:30:37 +03:00 committed by Jason Gunthorpe
parent f443452900
commit 9e2a187a93
9 changed files with 12 additions and 22 deletions

View File

@ -2617,6 +2617,7 @@ void ib_set_device_ops(struct ib_device *dev, const struct ib_device_ops *ops)
SET_DEVICE_OP(dev_ops, drain_rq); SET_DEVICE_OP(dev_ops, drain_rq);
SET_DEVICE_OP(dev_ops, drain_sq); SET_DEVICE_OP(dev_ops, drain_sq);
SET_DEVICE_OP(dev_ops, enable_driver); SET_DEVICE_OP(dev_ops, enable_driver);
SET_DEVICE_OP(dev_ops, fill_res_cq_entry);
SET_DEVICE_OP(dev_ops, fill_res_entry); SET_DEVICE_OP(dev_ops, fill_res_entry);
SET_DEVICE_OP(dev_ops, fill_res_mr_entry); SET_DEVICE_OP(dev_ops, fill_res_mr_entry);
SET_DEVICE_OP(dev_ops, fill_stat_mr_entry); SET_DEVICE_OP(dev_ops, fill_stat_mr_entry);

View File

@ -598,9 +598,8 @@ static int fill_res_cq_entry(struct sk_buff *msg, bool has_cap_net_admin,
if (fill_res_name_pid(msg, res)) if (fill_res_name_pid(msg, res))
goto err; goto err;
if (fill_res_entry(dev, msg, res)) if (dev->ops.fill_res_cq_entry)
goto err; return dev->ops.fill_res_cq_entry(msg, cq);
return 0; return 0;
err: return -EMSGSIZE; err: return -EMSGSIZE;

View File

@ -1056,6 +1056,7 @@ struct c4iw_wr_wait *c4iw_alloc_wr_wait(gfp_t gfp);
typedef int c4iw_restrack_func(struct sk_buff *msg, typedef int c4iw_restrack_func(struct sk_buff *msg,
struct rdma_restrack_entry *res); struct rdma_restrack_entry *res);
int c4iw_fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr); int c4iw_fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr);
int c4iw_fill_res_cq_entry(struct sk_buff *msg, struct ib_cq *ibcq);
extern c4iw_restrack_func *c4iw_restrack_funcs[RDMA_RESTRACK_MAX]; extern c4iw_restrack_func *c4iw_restrack_funcs[RDMA_RESTRACK_MAX];
#endif #endif

View File

@ -485,6 +485,7 @@ static const struct ib_device_ops c4iw_dev_ops = {
.destroy_cq = c4iw_destroy_cq, .destroy_cq = c4iw_destroy_cq,
.destroy_qp = c4iw_destroy_qp, .destroy_qp = c4iw_destroy_qp,
.destroy_srq = c4iw_destroy_srq, .destroy_srq = c4iw_destroy_srq,
.fill_res_cq_entry = c4iw_fill_res_cq_entry,
.fill_res_entry = fill_res_entry, .fill_res_entry = fill_res_entry,
.fill_res_mr_entry = c4iw_fill_res_mr_entry, .fill_res_mr_entry = c4iw_fill_res_mr_entry,
.get_dev_fw_str = get_dev_fw_str, .get_dev_fw_str = get_dev_fw_str,

View File

@ -372,10 +372,8 @@ static int fill_swcqes(struct sk_buff *msg, struct t4_cq *cq,
return -EMSGSIZE; return -EMSGSIZE;
} }
static int fill_res_cq_entry(struct sk_buff *msg, int c4iw_fill_res_cq_entry(struct sk_buff *msg, struct ib_cq *ibcq)
struct rdma_restrack_entry *res)
{ {
struct ib_cq *ibcq = container_of(res, struct ib_cq, res);
struct c4iw_cq *chp = to_c4iw_cq(ibcq); struct c4iw_cq *chp = to_c4iw_cq(ibcq);
struct nlattr *table_attr; struct nlattr *table_attr;
struct t4_cqe hwcqes[2]; struct t4_cqe hwcqes[2];
@ -494,5 +492,4 @@ int c4iw_fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ibmr)
c4iw_restrack_func *c4iw_restrack_funcs[RDMA_RESTRACK_MAX] = { c4iw_restrack_func *c4iw_restrack_funcs[RDMA_RESTRACK_MAX] = {
[RDMA_RESTRACK_QP] = fill_res_qp_entry, [RDMA_RESTRACK_QP] = fill_res_qp_entry,
[RDMA_RESTRACK_CM_ID] = fill_res_ep_entry, [RDMA_RESTRACK_CM_ID] = fill_res_ep_entry,
[RDMA_RESTRACK_CQ] = fill_res_cq_entry,
}; };

View File

@ -1266,6 +1266,6 @@ void hns_roce_handle_device_err(struct hns_roce_dev *hr_dev);
int hns_roce_init(struct hns_roce_dev *hr_dev); int hns_roce_init(struct hns_roce_dev *hr_dev);
void hns_roce_exit(struct hns_roce_dev *hr_dev); void hns_roce_exit(struct hns_roce_dev *hr_dev);
int hns_roce_fill_res_entry(struct sk_buff *msg, int hns_roce_fill_res_cq_entry(struct sk_buff *msg,
struct rdma_restrack_entry *res); struct ib_cq *ib_cq);
#endif /* _HNS_ROCE_DEVICE_H */ #endif /* _HNS_ROCE_DEVICE_H */

View File

@ -428,7 +428,7 @@ static const struct ib_device_ops hns_roce_dev_ops = {
.destroy_ah = hns_roce_destroy_ah, .destroy_ah = hns_roce_destroy_ah,
.destroy_cq = hns_roce_destroy_cq, .destroy_cq = hns_roce_destroy_cq,
.disassociate_ucontext = hns_roce_disassociate_ucontext, .disassociate_ucontext = hns_roce_disassociate_ucontext,
.fill_res_entry = hns_roce_fill_res_entry, .fill_res_cq_entry = hns_roce_fill_res_cq_entry,
.get_dma_mr = hns_roce_get_dma_mr, .get_dma_mr = hns_roce_get_dma_mr,
.get_link_layer = hns_roce_get_link_layer, .get_link_layer = hns_roce_get_link_layer,
.get_port_immutable = hns_roce_port_immutable, .get_port_immutable = hns_roce_port_immutable,

View File

@ -76,10 +76,9 @@ static int hns_roce_fill_cq(struct sk_buff *msg,
return -EMSGSIZE; return -EMSGSIZE;
} }
static int hns_roce_fill_res_cq_entry(struct sk_buff *msg, int hns_roce_fill_res_cq_entry(struct sk_buff *msg,
struct rdma_restrack_entry *res) struct ib_cq *ib_cq)
{ {
struct ib_cq *ib_cq = container_of(res, struct ib_cq, res);
struct hns_roce_dev *hr_dev = to_hr_dev(ib_cq->device); struct hns_roce_dev *hr_dev = to_hr_dev(ib_cq->device);
struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq); struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq);
struct hns_roce_v2_cq_context *context; struct hns_roce_v2_cq_context *context;
@ -119,12 +118,3 @@ static int hns_roce_fill_res_cq_entry(struct sk_buff *msg,
kfree(context); kfree(context);
return ret; return ret;
} }
int hns_roce_fill_res_entry(struct sk_buff *msg,
struct rdma_restrack_entry *res)
{
if (res->type == RDMA_RESTRACK_CQ)
return hns_roce_fill_res_cq_entry(msg, res);
return 0;
}

View File

@ -2584,6 +2584,7 @@ struct ib_device_ops {
int (*fill_res_entry)(struct sk_buff *msg, int (*fill_res_entry)(struct sk_buff *msg,
struct rdma_restrack_entry *entry); struct rdma_restrack_entry *entry);
int (*fill_res_mr_entry)(struct sk_buff *msg, struct ib_mr *ibmr); int (*fill_res_mr_entry)(struct sk_buff *msg, struct ib_mr *ibmr);
int (*fill_res_cq_entry)(struct sk_buff *msg, struct ib_cq *ibcq);
/* Device lifecycle callbacks */ /* Device lifecycle callbacks */
/* /*