From nobody Thu Apr 2 22:12:12 2026 Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 83D613587A2; Fri, 13 Feb 2026 11:00:11 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1770980411; cv=none; b=CtRbzMxBOJeaKAP+p2zWuLUq/d0eE7SJKWw8mS4Sq+V0YqJjWxpEXLNE2x8++8QwpGFv+k0FMtt7aP7gn9vUam8q6xu52cNfEcDadcHYVoa286yVq+5Q0bSt6gXzifAh8V6uLRCruEN3hCw4ZYI/mH6RY8fN6BQ4HSgRntRHHZM= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1770980411; c=relaxed/simple; bh=s78Zi7UXsBRgLKBAZ5n7fQmfZPeu1Dtkh9uqA8RaVzw=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version:Content-Type; b=WXSK668oICNKuOS1XnJ5YhXTzLec0CnJjo5syTXVbr21TfXRUJtEn6Rw53CCIcRmcXhSpYSkW0ZVUIe4dKIBbVq0mH68gHcDxobzsRMu1ZbwYUWbW3XAN6rhwJicsBIkdKsDiaxFFmaqdOh7Aw2W8p2tAZ8DmZ4MQCQUIYyGpcg= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=XEGQDWfk; arc=none smtp.client-ip=10.30.226.201 Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="XEGQDWfk" Received: by smtp.kernel.org (Postfix) with ESMTPSA id 96048C116C6; Fri, 13 Feb 2026 11:00:10 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1770980411; bh=s78Zi7UXsBRgLKBAZ5n7fQmfZPeu1Dtkh9uqA8RaVzw=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=XEGQDWfkcO/J4lqcTytbrozdtN2QImw6fX8TX2+VVw+kgh7Ul6sUjT3/+tn20G+xZ fGKgofiPyxNhood8LIOuMLXzh6K7QpUVj3vzDaEttTjAbEv38H+jVR7vqBWaEr6GPb NymVNsDTkKf+SacdhWRGWN5XKCwIS83otzgDKZBVjqoyRf3slF+8eWmLvqlCDHJwTF X8Rr+e7IF8oUYxc8Ocpdx1FcD9oiyCK6qYeho3amuxruHBXlyFsYepuIgti+IPy47a ZduKb9WTwNUsX5ytKvTPOMYS472HqhXZoXNYYEH0VpcAQUNt8slETwCACophh7VagA lBlOoAg2OE6Xw== From: Leon Romanovsky To: Jason Gunthorpe , Leon Romanovsky , Selvin Xavier , Kalesh AP , Potnuri Bharat Teja , Michael Margolin , Gal Pressman , Yossi Leybovich , Cheng Xu , Kai Shen , Chengchang Tang , Junxian Huang , Abhijit Gangurde , Allen Hubbe , Krzysztof Czurylo , Tatyana Nikolova , Long Li , Konstantin Taranov , Yishai Hadas , Michal Kalderon , Bryan Tan , Vishnu Dasa , Broadcom internal kernel review list , Christian Benvenuti , Nelson Escobar , Dennis Dalessandro , Bernard Metzler , Zhu Yanjun Cc: linux-kernel@vger.kernel.org, linux-rdma@vger.kernel.org, linux-hyperv@vger.kernel.org Subject: [PATCH rdma-next 21/50] RDMA/vmw_pvrdma: Provide a modern CQ creation interface Date: Fri, 13 Feb 2026 12:57:57 +0200 Message-ID: <20260213-refactor-umem-v1-21-f3be85847922@nvidia.com> X-Mailer: git-send-email 2.53.0 In-Reply-To: <20260213-refactor-umem-v1-0-f3be85847922@nvidia.com> References: <20260213-refactor-umem-v1-0-f3be85847922@nvidia.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" X-Mailer: b4 0.15-dev-47773 Content-Transfer-Encoding: quoted-printable From: Leon Romanovsky The uverbs CQ creation UAPI allows users to supply their own umem for a CQ. Update vmw_pvrdma to support this workflow while preserving support for cre= ating umem through the legacy interface. Signed-off-by: Leon Romanovsky --- drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c | 171 ++++++++++++++++----= ---- drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c | 1 + drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h | 3 + 3 files changed, 121 insertions(+), 54 deletions(-) diff --git a/drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c b/drivers/infinib= and/hw/vmw_pvrdma/pvrdma_cq.c index b3df6eb9b8ef..c43c363565c1 100644 --- a/drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c +++ b/drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c @@ -90,16 +90,9 @@ int pvrdma_req_notify_cq(struct ib_cq *ibcq, return has_data; } =20 -/** - * pvrdma_create_cq - create completion queue - * @ibcq: Allocated CQ - * @attr: completion queue attributes - * @attrs: bundle - * - * @return: 0 on success - */ -int pvrdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *att= r, - struct uverbs_attr_bundle *attrs) +int pvrdma_create_user_cq(struct ib_cq *ibcq, + const struct ib_cq_init_attr *attr, + struct uverbs_attr_bundle *attrs) { struct ib_udata *udata =3D &attrs->driver_udata; struct ib_device *ibdev =3D ibcq->device; @@ -123,58 +116,48 @@ int pvrdma_create_cq(struct ib_cq *ibcq, const struct= ib_cq_init_attr *attr, if (attr->flags) return -EOPNOTSUPP; =20 - entries =3D roundup_pow_of_two(entries); - if (entries < 1 || entries > dev->dsr->caps.max_cqe) + if (attr->cqe > dev->dsr->caps.max_cqe) return -EINVAL; =20 + entries =3D roundup_pow_of_two(entries); + if (!atomic_add_unless(&dev->num_cqs, 1, dev->dsr->caps.max_cq)) return -ENOMEM; =20 cq->ibcq.cqe =3D entries; - cq->is_kernel =3D !udata; - - if (!cq->is_kernel) { - if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { - ret =3D -EFAULT; - goto err_cq; - } - - cq->umem =3D ib_umem_get(ibdev, ucmd.buf_addr, ucmd.buf_size, - IB_ACCESS_LOCAL_WRITE); - if (IS_ERR(cq->umem)) { - ret =3D PTR_ERR(cq->umem); - goto err_cq; - } + cq->is_kernel =3D false; =20 - npages =3D ib_umem_num_dma_blocks(cq->umem, PAGE_SIZE); - } else { - /* One extra page for shared ring state */ - npages =3D 1 + (entries * sizeof(struct pvrdma_cqe) + - PAGE_SIZE - 1) / PAGE_SIZE; + if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { + ret =3D -EFAULT; + goto err_cq; + } =20 - /* Skip header page. */ - cq->offset =3D PAGE_SIZE; + if (!ibcq->umem) + ibcq->umem =3D ib_umem_get(ibdev, ucmd.buf_addr, ucmd.buf_size, + IB_ACCESS_LOCAL_WRITE); + if (IS_ERR(ibcq->umem)) { + ret =3D PTR_ERR(ibcq->umem); + goto err_cq; } =20 + npages =3D ib_umem_num_dma_blocks(cq->umem, PAGE_SIZE); + if (npages < 0 || npages > PVRDMA_PAGE_DIR_MAX_PAGES) { dev_warn(&dev->pdev->dev, "overflow pages in completion queue\n"); ret =3D -EINVAL; - goto err_umem; + goto err_cq; } =20 - ret =3D pvrdma_page_dir_init(dev, &cq->pdir, npages, cq->is_kernel); + ret =3D pvrdma_page_dir_init(dev, &cq->pdir, npages, false); if (ret) { dev_warn(&dev->pdev->dev, "could not allocate page directory\n"); - goto err_umem; + goto err_cq; } =20 /* Ring state is always the first page. Set in library for user cq. */ - if (cq->is_kernel) - cq->ring_state =3D cq->pdir.pages[0]; - else - pvrdma_page_dir_insert_umem(&cq->pdir, cq->umem, 0); + pvrdma_page_dir_insert_umem(&cq->pdir, cq->umem, 0); =20 refcount_set(&cq->refcnt, 1); init_completion(&cq->free); @@ -183,7 +166,7 @@ int pvrdma_create_cq(struct ib_cq *ibcq, const struct i= b_cq_init_attr *attr, memset(cmd, 0, sizeof(*cmd)); cmd->hdr.cmd =3D PVRDMA_CMD_CREATE_CQ; cmd->nchunks =3D npages; - cmd->ctx_handle =3D context ? context->ctx_handle : 0; + cmd->ctx_handle =3D context->ctx_handle; cmd->cqe =3D entries; cmd->pdir_dma =3D cq->pdir.dir_dma; ret =3D pvrdma_cmd_post(dev, &req, &rsp, PVRDMA_CMD_CREATE_CQ_RESP); @@ -200,24 +183,106 @@ int pvrdma_create_cq(struct ib_cq *ibcq, const struc= t ib_cq_init_attr *attr, dev->cq_tbl[cq->cq_handle % dev->dsr->caps.max_cq] =3D cq; spin_unlock_irqrestore(&dev->cq_tbl_lock, flags); =20 - if (!cq->is_kernel) { - cq->uar =3D &context->uar; + cq->uar =3D &context->uar; =20 - /* Copy udata back. */ - if (ib_copy_to_udata(udata, &cq_resp, sizeof(cq_resp))) { - dev_warn(&dev->pdev->dev, - "failed to copy back udata\n"); - pvrdma_destroy_cq(&cq->ibcq, udata); - return -EINVAL; - } + /* Copy udata back. */ + if (ib_copy_to_udata(udata, &cq_resp, sizeof(cq_resp))) { + dev_warn(&dev->pdev->dev, + "failed to copy back udata\n"); + pvrdma_destroy_cq(&cq->ibcq, udata); + return -EINVAL; } =20 return 0; =20 err_page_dir: pvrdma_page_dir_cleanup(dev, &cq->pdir); -err_umem: - ib_umem_release(cq->umem); +err_cq: + atomic_dec(&dev->num_cqs); + return ret; +} + +int pvrdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *att= r, + struct uverbs_attr_bundle *attrs) +{ + struct ib_device *ibdev =3D ibcq->device; + int entries =3D attr->cqe; + struct pvrdma_dev *dev =3D to_vdev(ibdev); + struct pvrdma_cq *cq =3D to_vcq(ibcq); + int ret; + int npages; + unsigned long flags; + union pvrdma_cmd_req req; + union pvrdma_cmd_resp rsp; + struct pvrdma_cmd_create_cq *cmd =3D &req.create_cq; + struct pvrdma_cmd_create_cq_resp *resp =3D &rsp.create_cq_resp; + + BUILD_BUG_ON(sizeof(struct pvrdma_cqe) !=3D 64); + + if (attr->flags) + return -EOPNOTSUPP; + + if (attr->cqe > dev->dsr->caps.max_cqe) + return -EINVAL; + entries =3D roundup_pow_of_two(entries); + + if (!atomic_add_unless(&dev->num_cqs, 1, dev->dsr->caps.max_cq)) + return -ENOMEM; + + cq->ibcq.cqe =3D entries; + cq->is_kernel =3D true; + + /* One extra page for shared ring state */ + npages =3D 1 + (entries * sizeof(struct pvrdma_cqe) + + PAGE_SIZE - 1) / PAGE_SIZE; + + /* Skip header page. */ + cq->offset =3D PAGE_SIZE; + + if (npages < 0 || npages > PVRDMA_PAGE_DIR_MAX_PAGES) { + dev_warn(&dev->pdev->dev, + "overflow pages in completion queue\n"); + ret =3D -EINVAL; + goto err_cq; + } + + ret =3D pvrdma_page_dir_init(dev, &cq->pdir, npages, true); + if (ret) { + dev_warn(&dev->pdev->dev, + "could not allocate page directory\n"); + goto err_cq; + } + + /* Ring state is always the first page. Set in library for user cq. */ + cq->ring_state =3D cq->pdir.pages[0]; + + refcount_set(&cq->refcnt, 1); + init_completion(&cq->free); + spin_lock_init(&cq->cq_lock); + + memset(cmd, 0, sizeof(*cmd)); + cmd->hdr.cmd =3D PVRDMA_CMD_CREATE_CQ; + cmd->nchunks =3D npages; + cmd->ctx_handle =3D 0; + cmd->cqe =3D entries; + cmd->pdir_dma =3D cq->pdir.dir_dma; + ret =3D pvrdma_cmd_post(dev, &req, &rsp, PVRDMA_CMD_CREATE_CQ_RESP); + if (ret < 0) { + dev_warn(&dev->pdev->dev, + "could not create completion queue, error: %d\n", ret); + goto err_page_dir; + } + + cq->ibcq.cqe =3D resp->cqe; + cq->cq_handle =3D resp->cq_handle; + spin_lock_irqsave(&dev->cq_tbl_lock, flags); + dev->cq_tbl[cq->cq_handle % dev->dsr->caps.max_cq] =3D cq; + spin_unlock_irqrestore(&dev->cq_tbl_lock, flags); + + return 0; + +err_page_dir: + pvrdma_page_dir_cleanup(dev, &cq->pdir); err_cq: atomic_dec(&dev->num_cqs); return ret; @@ -229,8 +294,6 @@ static void pvrdma_free_cq(struct pvrdma_dev *dev, stru= ct pvrdma_cq *cq) complete(&cq->free); wait_for_completion(&cq->free); =20 - ib_umem_release(cq->umem); - pvrdma_page_dir_cleanup(dev, &cq->pdir); } =20 diff --git a/drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c b/drivers/infin= iband/hw/vmw_pvrdma/pvrdma_main.c index 1664d1d7d969..3f5b94a1e517 100644 --- a/drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c +++ b/drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c @@ -194,6 +194,7 @@ static const struct ib_device_ops pvrdma_dev_ops =3D { .alloc_ucontext =3D pvrdma_alloc_ucontext, .create_ah =3D pvrdma_create_ah, .create_cq =3D pvrdma_create_cq, + .create_user_cq =3D pvrdma_create_user_cq, .create_qp =3D pvrdma_create_qp, .dealloc_pd =3D pvrdma_dealloc_pd, .dealloc_ucontext =3D pvrdma_dealloc_ucontext, diff --git a/drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h b/drivers/infi= niband/hw/vmw_pvrdma/pvrdma_verbs.h index 603e5a9311eb..18910d336744 100644 --- a/drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h +++ b/drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h @@ -375,6 +375,9 @@ int pvrdma_map_mr_sg(struct ib_mr *ibmr, struct scatter= list *sg, int sg_nents, unsigned int *sg_offset); int pvrdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *att= r, struct uverbs_attr_bundle *attrs); +int pvrdma_create_user_cq(struct ib_cq *ibcq, + const struct ib_cq_init_attr *attr, + struct uverbs_attr_bundle *attrs); int pvrdma_destroy_cq(struct ib_cq *cq, struct ib_udata *udata); int pvrdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc); int pvrdma_req_notify_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags); --=20 2.52.0