mirror of
https://github.com/torvalds/linux
synced 2024-10-07 20:05:15 +00:00
RDMA/hns: Optimize qp doorbell allocation flow
Encapsulate the kernel qp doorbell allocation related code into 2 functions: alloc_qp_db() and free_qp_db(). Link: https://lore.kernel.org/r/1582526258-13825-8-git-send-email-liweihang@huawei.com Signed-off-by: Xi Wang <wangxi11@huawei.com> Signed-off-by: Weihang Li <liweihang@huawei.com> Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
This commit is contained in:
parent
b37c413997
commit
cfec045b82
|
@ -843,7 +843,7 @@ static int alloc_qp_buf(struct hns_roce_dev *hr_dev, struct hns_roce_qp *hr_qp,
|
|||
}
|
||||
|
||||
if (udata) {
|
||||
hr_qp->umem = ib_umem_get(udata, addr, hr_qp->buff_size, 0);
|
||||
hr_qp->umem = ib_umem_get(ibdev, addr, hr_qp->buff_size, 0);
|
||||
if (IS_ERR(hr_qp->umem)) {
|
||||
ret = PTR_ERR(hr_qp->umem);
|
||||
goto err_inline;
|
||||
|
@ -895,6 +895,114 @@ static void free_qp_buf(struct hns_roce_dev *hr_dev, struct hns_roce_qp *hr_qp)
|
|||
free_rq_inline_buf(hr_qp);
|
||||
}
|
||||
|
||||
static inline bool user_qp_has_sdb(struct hns_roce_dev *hr_dev,
|
||||
struct ib_qp_init_attr *init_attr,
|
||||
struct ib_udata *udata,
|
||||
struct hns_roce_ib_create_qp_resp *resp,
|
||||
struct hns_roce_ib_create_qp *ucmd)
|
||||
{
|
||||
return ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_SQ_RECORD_DB) &&
|
||||
udata->outlen >= offsetofend(typeof(*resp), cap_flags) &&
|
||||
hns_roce_qp_has_sq(init_attr) &&
|
||||
udata->inlen >= offsetofend(typeof(*ucmd), sdb_addr));
|
||||
}
|
||||
|
||||
static inline bool user_qp_has_rdb(struct hns_roce_dev *hr_dev,
|
||||
struct ib_qp_init_attr *init_attr,
|
||||
struct ib_udata *udata,
|
||||
struct hns_roce_ib_create_qp_resp *resp)
|
||||
{
|
||||
return ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) &&
|
||||
udata->outlen >= offsetofend(typeof(*resp), cap_flags) &&
|
||||
hns_roce_qp_has_rq(init_attr));
|
||||
}
|
||||
|
||||
static inline bool kernel_qp_has_rdb(struct hns_roce_dev *hr_dev,
|
||||
struct ib_qp_init_attr *init_attr)
|
||||
{
|
||||
return ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) &&
|
||||
hns_roce_qp_has_rq(init_attr));
|
||||
}
|
||||
|
||||
static int alloc_qp_db(struct hns_roce_dev *hr_dev, struct hns_roce_qp *hr_qp,
|
||||
struct ib_qp_init_attr *init_attr,
|
||||
struct ib_udata *udata,
|
||||
struct hns_roce_ib_create_qp *ucmd,
|
||||
struct hns_roce_ib_create_qp_resp *resp)
|
||||
{
|
||||
struct hns_roce_ucontext *uctx = rdma_udata_to_drv_context(
|
||||
udata, struct hns_roce_ucontext, ibucontext);
|
||||
struct ib_device *ibdev = &hr_dev->ib_dev;
|
||||
int ret;
|
||||
|
||||
if (udata) {
|
||||
if (user_qp_has_sdb(hr_dev, init_attr, udata, resp, ucmd)) {
|
||||
ret = hns_roce_db_map_user(uctx, udata, ucmd->sdb_addr,
|
||||
&hr_qp->sdb);
|
||||
if (ret) {
|
||||
ibdev_err(ibdev,
|
||||
"Failed to map user SQ doorbell\n");
|
||||
goto err_out;
|
||||
}
|
||||
hr_qp->sdb_en = 1;
|
||||
resp->cap_flags |= HNS_ROCE_SUPPORT_SQ_RECORD_DB;
|
||||
}
|
||||
|
||||
if (user_qp_has_rdb(hr_dev, init_attr, udata, resp)) {
|
||||
ret = hns_roce_db_map_user(uctx, udata, ucmd->db_addr,
|
||||
&hr_qp->rdb);
|
||||
if (ret) {
|
||||
ibdev_err(ibdev,
|
||||
"Failed to map user RQ doorbell\n");
|
||||
goto err_sdb;
|
||||
}
|
||||
hr_qp->rdb_en = 1;
|
||||
resp->cap_flags |= HNS_ROCE_SUPPORT_RQ_RECORD_DB;
|
||||
}
|
||||
} else {
|
||||
/* QP doorbell register address */
|
||||
hr_qp->sq.db_reg_l = hr_dev->reg_base + hr_dev->sdb_offset +
|
||||
DB_REG_OFFSET * hr_dev->priv_uar.index;
|
||||
hr_qp->rq.db_reg_l = hr_dev->reg_base + hr_dev->odb_offset +
|
||||
DB_REG_OFFSET * hr_dev->priv_uar.index;
|
||||
|
||||
if (kernel_qp_has_rdb(hr_dev, init_attr)) {
|
||||
ret = hns_roce_alloc_db(hr_dev, &hr_qp->rdb, 0);
|
||||
if (ret) {
|
||||
ibdev_err(ibdev,
|
||||
"Failed to alloc kernel RQ doorbell\n");
|
||||
goto err_out;
|
||||
}
|
||||
*hr_qp->rdb.db_record = 0;
|
||||
hr_qp->rdb_en = 1;
|
||||
}
|
||||
}
|
||||
|
||||
return 0;
|
||||
err_sdb:
|
||||
if (udata && hr_qp->sdb_en)
|
||||
hns_roce_db_unmap_user(uctx, &hr_qp->sdb);
|
||||
err_out:
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void free_qp_db(struct hns_roce_dev *hr_dev, struct hns_roce_qp *hr_qp,
|
||||
struct ib_udata *udata)
|
||||
{
|
||||
struct hns_roce_ucontext *uctx = rdma_udata_to_drv_context(
|
||||
udata, struct hns_roce_ucontext, ibucontext);
|
||||
|
||||
if (udata) {
|
||||
if (hr_qp->rdb_en)
|
||||
hns_roce_db_unmap_user(uctx, &hr_qp->rdb);
|
||||
if (hr_qp->sdb_en)
|
||||
hns_roce_db_unmap_user(uctx, &hr_qp->sdb);
|
||||
} else {
|
||||
if (hr_qp->rdb_en)
|
||||
hns_roce_free_db(hr_dev, &hr_qp->rdb);
|
||||
}
|
||||
}
|
||||
|
||||
static int alloc_kernel_wrid(struct hns_roce_dev *hr_dev,
|
||||
struct hns_roce_qp *hr_qp)
|
||||
{
|
||||
|
@ -991,11 +1099,9 @@ static int hns_roce_create_qp_common(struct hns_roce_dev *hr_dev,
|
|||
struct ib_udata *udata,
|
||||
struct hns_roce_qp *hr_qp)
|
||||
{
|
||||
struct device *dev = hr_dev->dev;
|
||||
struct hns_roce_ib_create_qp ucmd;
|
||||
struct hns_roce_ib_create_qp_resp resp = {};
|
||||
struct hns_roce_ucontext *uctx = rdma_udata_to_drv_context(
|
||||
udata, struct hns_roce_ucontext, ibucontext);
|
||||
struct ib_device *ibdev = &hr_dev->ib_dev;
|
||||
struct hns_roce_ib_create_qp ucmd;
|
||||
int ret;
|
||||
|
||||
mutex_init(&hr_qp->mutex);
|
||||
|
@ -1007,95 +1113,55 @@ static int hns_roce_create_qp_common(struct hns_roce_dev *hr_dev,
|
|||
|
||||
ret = set_qp_param(hr_dev, hr_qp, init_attr, udata, &ucmd);
|
||||
if (ret) {
|
||||
ibdev_err(&hr_dev->ib_dev, "Failed to set QP param\n");
|
||||
ibdev_err(ibdev, "Failed to set QP param\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
if (udata) {
|
||||
if ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_SQ_RECORD_DB) &&
|
||||
(udata->inlen >= sizeof(ucmd)) &&
|
||||
(udata->outlen >= sizeof(resp)) &&
|
||||
hns_roce_qp_has_sq(init_attr)) {
|
||||
ret = hns_roce_db_map_user(uctx, udata, ucmd.sdb_addr,
|
||||
&hr_qp->sdb);
|
||||
if (ret) {
|
||||
dev_err(dev, "sq record doorbell map failed!\n");
|
||||
goto err_out;
|
||||
}
|
||||
|
||||
/* indicate kernel supports sq record db */
|
||||
resp.cap_flags |= HNS_ROCE_SUPPORT_SQ_RECORD_DB;
|
||||
hr_qp->sdb_en = 1;
|
||||
}
|
||||
|
||||
if ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) &&
|
||||
(udata->outlen >= sizeof(resp)) &&
|
||||
hns_roce_qp_has_rq(init_attr)) {
|
||||
ret = hns_roce_db_map_user(uctx, udata, ucmd.db_addr,
|
||||
&hr_qp->rdb);
|
||||
if (ret) {
|
||||
dev_err(dev, "rq record doorbell map failed!\n");
|
||||
goto err_sq_dbmap;
|
||||
}
|
||||
|
||||
/* indicate kernel supports rq record db */
|
||||
resp.cap_flags |= HNS_ROCE_SUPPORT_RQ_RECORD_DB;
|
||||
hr_qp->rdb_en = 1;
|
||||
}
|
||||
} else {
|
||||
/* QP doorbell register address */
|
||||
hr_qp->sq.db_reg_l = hr_dev->reg_base + hr_dev->sdb_offset +
|
||||
DB_REG_OFFSET * hr_dev->priv_uar.index;
|
||||
hr_qp->rq.db_reg_l = hr_dev->reg_base + hr_dev->odb_offset +
|
||||
DB_REG_OFFSET * hr_dev->priv_uar.index;
|
||||
|
||||
if ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) &&
|
||||
hns_roce_qp_has_rq(init_attr)) {
|
||||
ret = hns_roce_alloc_db(hr_dev, &hr_qp->rdb, 0);
|
||||
if (ret) {
|
||||
dev_err(dev, "rq record doorbell alloc failed!\n");
|
||||
goto err_out;
|
||||
}
|
||||
*hr_qp->rdb.db_record = 0;
|
||||
hr_qp->rdb_en = 1;
|
||||
}
|
||||
|
||||
if (!udata) {
|
||||
ret = alloc_kernel_wrid(hr_dev, hr_qp);
|
||||
if (ret) {
|
||||
ibdev_err(&hr_dev->ib_dev, "Failed to alloc wrid\n");
|
||||
goto err_db;
|
||||
ibdev_err(ibdev, "Failed to alloc wrid\n");
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
ret = alloc_qp_db(hr_dev, hr_qp, init_attr, udata, &ucmd, &resp);
|
||||
if (ret) {
|
||||
ibdev_err(ibdev, "Failed to alloc QP doorbell\n");
|
||||
goto err_wrid;
|
||||
}
|
||||
|
||||
ret = alloc_qp_buf(hr_dev, hr_qp, init_attr, udata, ucmd.buf_addr);
|
||||
if (ret) {
|
||||
ibdev_err(&hr_dev->ib_dev, "Failed to alloc QP buffer\n");
|
||||
ibdev_err(ibdev, "Failed to alloc QP buffer\n");
|
||||
goto err_db;
|
||||
}
|
||||
|
||||
ret = alloc_qpn(hr_dev, hr_qp);
|
||||
if (ret) {
|
||||
ibdev_err(&hr_dev->ib_dev, "Failed to alloc QPN\n");
|
||||
ibdev_err(ibdev, "Failed to alloc QPN\n");
|
||||
goto err_buf;
|
||||
}
|
||||
|
||||
ret = alloc_qpc(hr_dev, hr_qp);
|
||||
if (ret) {
|
||||
ibdev_err(&hr_dev->ib_dev, "Failed to alloc QP context\n");
|
||||
ibdev_err(ibdev, "Failed to alloc QP context\n");
|
||||
goto err_qpn;
|
||||
}
|
||||
|
||||
ret = hns_roce_qp_store(hr_dev, hr_qp, init_attr);
|
||||
if (ret) {
|
||||
ibdev_err(&hr_dev->ib_dev, "Failed to store QP\n");
|
||||
ibdev_err(ibdev, "Failed to store QP\n");
|
||||
goto err_qpc;
|
||||
}
|
||||
|
||||
if (udata) {
|
||||
ret = ib_copy_to_udata(udata, &resp,
|
||||
min(udata->outlen, sizeof(resp)));
|
||||
if (ret)
|
||||
if (ret) {
|
||||
ibdev_err(ibdev, "copy qp resp failed!\n");
|
||||
goto err_store;
|
||||
}
|
||||
}
|
||||
|
||||
if (hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_QP_FLOW_CTRL) {
|
||||
|
@ -1119,30 +1185,10 @@ static int hns_roce_create_qp_common(struct hns_roce_dev *hr_dev,
|
|||
free_qpn(hr_dev, hr_qp);
|
||||
err_buf:
|
||||
free_qp_buf(hr_dev, hr_qp);
|
||||
|
||||
free_kernel_wrid(hr_dev, hr_qp);
|
||||
|
||||
if (udata) {
|
||||
if ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB) &&
|
||||
(udata->outlen >= sizeof(resp)) &&
|
||||
hns_roce_qp_has_rq(init_attr))
|
||||
hns_roce_db_unmap_user(uctx, &hr_qp->rdb);
|
||||
}
|
||||
|
||||
err_sq_dbmap:
|
||||
if (udata)
|
||||
if ((hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_SQ_RECORD_DB) &&
|
||||
(udata->inlen >= sizeof(ucmd)) &&
|
||||
(udata->outlen >= sizeof(resp)) &&
|
||||
hns_roce_qp_has_sq(init_attr))
|
||||
hns_roce_db_unmap_user(uctx, &hr_qp->sdb);
|
||||
|
||||
err_db:
|
||||
if (!udata && hns_roce_qp_has_rq(init_attr) &&
|
||||
(hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_RECORD_DB))
|
||||
hns_roce_free_db(hr_dev, &hr_qp->rdb);
|
||||
|
||||
err_out:
|
||||
free_qp_db(hr_dev, hr_qp, udata);
|
||||
err_wrid:
|
||||
free_kernel_wrid(hr_dev, hr_qp);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@ -1157,23 +1203,7 @@ void hns_roce_qp_destroy(struct hns_roce_dev *hr_dev, struct hns_roce_qp *hr_qp,
|
|||
free_qpn(hr_dev, hr_qp);
|
||||
free_qp_buf(hr_dev, hr_qp);
|
||||
free_kernel_wrid(hr_dev, hr_qp);
|
||||
|
||||
if (udata) {
|
||||
struct hns_roce_ucontext *context =
|
||||
rdma_udata_to_drv_context(
|
||||
udata,
|
||||
struct hns_roce_ucontext,
|
||||
ibucontext);
|
||||
|
||||
if (hr_qp->sq.wqe_cnt && (hr_qp->sdb_en == 1))
|
||||
hns_roce_db_unmap_user(context, &hr_qp->sdb);
|
||||
|
||||
if (hr_qp->rq.wqe_cnt && (hr_qp->rdb_en == 1))
|
||||
hns_roce_db_unmap_user(context, &hr_qp->rdb);
|
||||
} else {
|
||||
if (hr_qp->rq.wqe_cnt)
|
||||
hns_roce_free_db(hr_dev, &hr_qp->rdb);
|
||||
}
|
||||
free_qp_db(hr_dev, hr_qp, udata);
|
||||
|
||||
kfree(hr_qp);
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue