[PATCH rdma-next 1/4] RDMA/bnxt_re: Simplify bnxt_re_init_depth() callers and implementation
From: Leon Romanovsky
Date: Wed Mar 18 2026 - 06:10:39 EST
From: Leon Romanovsky <leonro@xxxxxxxxxx>
All callers of bnxt_re_init_depth() compute the minimum between its return
value and another internal variable, often mixing variable types in the
process. Clean this up by making the logic simpler and more readable.
Signed-off-by: Leon Romanovsky <leonro@xxxxxxxxxx>
---
drivers/infiniband/hw/bnxt_re/ib_verbs.c | 81 ++++++++++++++------------------
drivers/infiniband/hw/bnxt_re/ib_verbs.h | 9 ++--
2 files changed, 42 insertions(+), 48 deletions(-)
diff --git a/drivers/infiniband/hw/bnxt_re/ib_verbs.c b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
index 182128ee4f242..40ac546f113bc 100644
--- a/drivers/infiniband/hw/bnxt_re/ib_verbs.c
+++ b/drivers/infiniband/hw/bnxt_re/ib_verbs.c
@@ -1442,7 +1442,6 @@ static int bnxt_re_init_rq_attr(struct bnxt_re_qp *qp,
struct bnxt_qplib_qp *qplqp;
struct bnxt_re_dev *rdev;
struct bnxt_qplib_q *rq;
- int entries;
rdev = qp->rdev;
qplqp = &qp->qplib_qp;
@@ -1465,8 +1464,9 @@ static int bnxt_re_init_rq_attr(struct bnxt_re_qp *qp,
/* Allocate 1 more than what's provided so posting max doesn't
* mean empty.
*/
- entries = bnxt_re_init_depth(init_attr->cap.max_recv_wr + 1, uctx);
- rq->max_wqe = min_t(u32, entries, dev_attr->max_qp_wqes + 1);
+ rq->max_wqe = bnxt_re_init_depth(init_attr->cap.max_recv_wr + 1,
+ dev_attr->max_qp_wqes + 1,
+ uctx);
rq->max_sw_wqe = rq->max_wqe;
rq->q_full_delta = 0;
rq->sg_info.pgsize = PAGE_SIZE;
@@ -1504,7 +1504,6 @@ static int bnxt_re_init_sq_attr(struct bnxt_re_qp *qp,
struct bnxt_re_dev *rdev;
struct bnxt_qplib_q *sq;
int diff = 0;
- int entries;
int rc;
rdev = qp->rdev;
@@ -1513,7 +1512,6 @@ static int bnxt_re_init_sq_attr(struct bnxt_re_qp *qp,
dev_attr = rdev->dev_attr;
sq->max_sge = init_attr->cap.max_send_sge;
- entries = init_attr->cap.max_send_wr;
if (uctx && qplqp->wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE) {
sq->max_wqe = ureq->sq_slots;
sq->max_sw_wqe = ureq->sq_slots;
@@ -1529,10 +1527,11 @@ static int bnxt_re_init_sq_attr(struct bnxt_re_qp *qp,
return rc;
/* Allocate 128 + 1 more than what's provided */
- diff = (qplqp->wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE) ?
- 0 : BNXT_QPLIB_RESERVED_QP_WRS;
- entries = bnxt_re_init_depth(entries + diff + 1, uctx);
- sq->max_wqe = min_t(u32, entries, dev_attr->max_qp_wqes + diff + 1);
+ if (qplqp->wqe_mode != BNXT_QPLIB_WQE_MODE_VARIABLE)
+ diff = BNXT_QPLIB_RESERVED_QP_WRS;
+ sq->max_wqe = bnxt_re_init_depth(
+ init_attr->cap.max_send_wr + diff + 1,
+ dev_attr->max_qp_wqes + diff + 1, uctx);
if (qplqp->wqe_mode == BNXT_QPLIB_WQE_MODE_VARIABLE)
sq->max_sw_wqe = bnxt_qplib_get_depth(sq, qplqp->wqe_mode, true);
else
@@ -1559,16 +1558,15 @@ static void bnxt_re_adjust_gsi_sq_attr(struct bnxt_re_qp *qp,
struct bnxt_qplib_dev_attr *dev_attr;
struct bnxt_qplib_qp *qplqp;
struct bnxt_re_dev *rdev;
- int entries;
rdev = qp->rdev;
qplqp = &qp->qplib_qp;
dev_attr = rdev->dev_attr;
if (!bnxt_qplib_is_chip_gen_p5_p7(rdev->chip_ctx)) {
- entries = bnxt_re_init_depth(init_attr->cap.max_send_wr + 1, uctx);
- qplqp->sq.max_wqe = min_t(u32, entries,
- dev_attr->max_qp_wqes + 1);
+ qplqp->sq.max_wqe =
+ bnxt_re_init_depth(init_attr->cap.max_send_wr + 1,
+ dev_attr->max_qp_wqes + 1, uctx);
qplqp->sq.q_full_delta = qplqp->sq.max_wqe -
init_attr->cap.max_send_wr;
qplqp->sq.max_sge++; /* Need one extra sge to put UD header */
@@ -2086,7 +2084,7 @@ int bnxt_re_create_srq(struct ib_srq *ib_srq,
struct bnxt_re_pd *pd;
struct ib_pd *ib_pd;
u32 active_srqs;
- int rc, entries;
+ int rc;
ib_pd = ib_srq->pd;
pd = container_of(ib_pd, struct bnxt_re_pd, ib_pd);
@@ -2112,10 +2110,9 @@ int bnxt_re_create_srq(struct ib_srq *ib_srq,
/* Allocate 1 more than what's provided so posting max doesn't
* mean empty
*/
- entries = bnxt_re_init_depth(srq_init_attr->attr.max_wr + 1, uctx);
- if (entries > dev_attr->max_srq_wqes + 1)
- entries = dev_attr->max_srq_wqes + 1;
- srq->qplib_srq.max_wqe = entries;
+ srq->qplib_srq.max_wqe =
+ bnxt_re_init_depth(srq_init_attr->attr.max_wr + 1,
+ dev_attr->max_srq_wqes + 1, uctx);
srq->qplib_srq.max_sge = srq_init_attr->attr.max_sge;
/* 128 byte wqe size for SRQ . So use max sges */
@@ -2296,7 +2293,7 @@ int bnxt_re_modify_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr,
struct bnxt_re_dev *rdev = qp->rdev;
struct bnxt_qplib_dev_attr *dev_attr = rdev->dev_attr;
enum ib_qp_state curr_qp_state, new_qp_state;
- int rc, entries;
+ int rc;
unsigned int flags;
u8 nw_type;
@@ -2510,9 +2507,9 @@ int bnxt_re_modify_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr,
"Create QP failed - max exceeded");
return -EINVAL;
}
- entries = bnxt_re_init_depth(qp_attr->cap.max_send_wr, uctx);
- qp->qplib_qp.sq.max_wqe = min_t(u32, entries,
- dev_attr->max_qp_wqes + 1);
+ qp->qplib_qp.sq.max_wqe =
+ bnxt_re_init_depth(qp_attr->cap.max_send_wr,
+ dev_attr->max_qp_wqes + 1, uctx);
qp->qplib_qp.sq.q_full_delta = qp->qplib_qp.sq.max_wqe -
qp_attr->cap.max_send_wr;
/*
@@ -2523,9 +2520,9 @@ int bnxt_re_modify_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr,
qp->qplib_qp.sq.q_full_delta -= 1;
qp->qplib_qp.sq.max_sge = qp_attr->cap.max_send_sge;
if (qp->qplib_qp.rq.max_wqe) {
- entries = bnxt_re_init_depth(qp_attr->cap.max_recv_wr, uctx);
- qp->qplib_qp.rq.max_wqe =
- min_t(u32, entries, dev_attr->max_qp_wqes + 1);
+ qp->qplib_qp.rq.max_wqe = bnxt_re_init_depth(
+ qp_attr->cap.max_recv_wr,
+ dev_attr->max_qp_wqes + 1, uctx);
qp->qplib_qp.rq.max_sw_wqe = qp->qplib_qp.rq.max_wqe;
qp->qplib_qp.rq.q_full_delta = qp->qplib_qp.rq.max_wqe -
qp_attr->cap.max_recv_wr;
@@ -3381,8 +3378,8 @@ int bnxt_re_create_user_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *att
struct bnxt_re_cq_resp resp = {};
struct bnxt_re_cq_req req;
int cqe = attr->cqe;
- int rc, entries;
- u32 active_cqs;
+ int rc;
+ u32 active_cqs, entries;
if (attr->flags)
return -EOPNOTSUPP;
@@ -3397,17 +3394,16 @@ int bnxt_re_create_user_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *att
cctx = rdev->chip_ctx;
cq->qplib_cq.cq_handle = (u64)(unsigned long)(&cq->qplib_cq);
- entries = bnxt_re_init_depth(cqe + 1, uctx);
- if (entries > dev_attr->max_cq_wqes + 1)
- entries = dev_attr->max_cq_wqes + 1;
-
rc = ib_copy_validate_udata_in_cm(udata, req, cq_handle,
BNXT_RE_CQ_FIXED_NUM_CQE_ENABLE);
if (rc)
return rc;
if (req.comp_mask & BNXT_RE_CQ_FIXED_NUM_CQE_ENABLE)
- entries = cqe;
+ entries = attr->cqe;
+ else
+ entries = bnxt_re_init_depth(attr->cqe + 1,
+ dev_attr->max_cq_wqes + 1, uctx);
if (!ibcq->umem) {
ibcq->umem = ib_umem_get(&rdev->ibdev, req.cq_va,
@@ -3480,7 +3476,7 @@ int bnxt_re_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
rdma_udata_to_drv_context(udata, struct bnxt_re_ucontext, ib_uctx);
struct bnxt_qplib_dev_attr *dev_attr = rdev->dev_attr;
int cqe = attr->cqe;
- int rc, entries;
+ int rc;
u32 active_cqs;
if (udata)
@@ -3498,11 +3494,8 @@ int bnxt_re_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
cq->rdev = rdev;
cq->qplib_cq.cq_handle = (u64)(unsigned long)(&cq->qplib_cq);
- entries = bnxt_re_init_depth(cqe + 1, uctx);
- if (entries > dev_attr->max_cq_wqes + 1)
- entries = dev_attr->max_cq_wqes + 1;
-
- cq->max_cql = min_t(u32, entries, MAX_CQL_PER_POLL);
+ cq->max_cql = bnxt_re_init_depth(attr->cqe + 1,
+ dev_attr->max_cq_wqes + 1, uctx);
cq->cql = kcalloc(cq->max_cql, sizeof(struct bnxt_qplib_cqe),
GFP_KERNEL);
if (!cq->cql)
@@ -3511,7 +3504,7 @@ int bnxt_re_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
cq->qplib_cq.sg_info.pgsize = SZ_4K;
cq->qplib_cq.sg_info.pgshft = __builtin_ctz(SZ_4K);
cq->qplib_cq.dpi = &rdev->dpi_privileged;
- cq->qplib_cq.max_wqe = entries;
+ cq->qplib_cq.max_wqe = cq->max_cql;
cq->qplib_cq.coalescing = &rdev->cq_coalescing;
cq->qplib_cq.nq = bnxt_re_get_nq(rdev);
cq->qplib_cq.cnq_hw_ring_id = cq->qplib_cq.nq->ring_id;
@@ -3522,7 +3515,7 @@ int bnxt_re_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
goto fail;
}
- cq->ib_cq.cqe = entries;
+ cq->ib_cq.cqe = cq->max_cql;
cq->cq_period = cq->qplib_cq.period;
active_cqs = atomic_inc_return(&rdev->stats.res.cq_count);
if (active_cqs > rdev->stats.res.cq_watermark)
@@ -3560,7 +3553,8 @@ int bnxt_re_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
struct bnxt_re_resize_cq_req req;
struct bnxt_re_dev *rdev;
struct bnxt_re_cq *cq;
- int rc, entries;
+ int rc;
+ u32 entries;
cq = container_of(ibcq, struct bnxt_re_cq, ib_cq);
rdev = cq->rdev;
@@ -3584,10 +3578,7 @@ int bnxt_re_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata)
}
uctx = rdma_udata_to_drv_context(udata, struct bnxt_re_ucontext, ib_uctx);
- entries = bnxt_re_init_depth(cqe + 1, uctx);
- if (entries > dev_attr->max_cq_wqes + 1)
- entries = dev_attr->max_cq_wqes + 1;
-
+ entries = bnxt_re_init_depth(cqe + 1, dev_attr->max_cq_wqes + 1, uctx);
/* uverbs consumer */
rc = ib_copy_validate_udata_in(udata, req, cq_va);
if (rc)
diff --git a/drivers/infiniband/hw/bnxt_re/ib_verbs.h b/drivers/infiniband/hw/bnxt_re/ib_verbs.h
index 3d02c16f54b61..dfe790ef42d75 100644
--- a/drivers/infiniband/hw/bnxt_re/ib_verbs.h
+++ b/drivers/infiniband/hw/bnxt_re/ib_verbs.h
@@ -190,10 +190,13 @@ enum {
BNXT_RE_UCNTX_CAP_VAR_WQE_ENABLED = 0x2ULL,
};
-static inline u32 bnxt_re_init_depth(u32 ent, struct bnxt_re_ucontext *uctx)
+static inline u32 bnxt_re_init_depth(u32 ent, u32 max,
+ struct bnxt_re_ucontext *uctx)
{
- return uctx ? (uctx->cmask & BNXT_RE_UCNTX_CAP_POW2_DISABLED) ?
- ent : roundup_pow_of_two(ent) : ent;
+ if (uctx && !(uctx->cmask & BNXT_RE_UCNTX_CAP_POW2_DISABLED))
+ return min(roundup_pow_of_two(ent), max);
+
+ return ent;
}
static inline bool bnxt_re_is_var_size_supported(struct bnxt_re_dev *rdev,
--
2.53.0