From: Ajay Sharma <sharmaajay@microsoft.com>
Query the adapter capabilities to expose to
other clients and VF. This checks against
the user supplied values and protects against
overflows.
Signed-off-by: Ajay Sharma <sharmaajay@microsoft.com>
---
drivers/infiniband/hw/mana/device.c | 4 ++
drivers/infiniband/hw/mana/main.c | 66 +++++++++++++++++++++++++---
drivers/infiniband/hw/mana/mana_ib.h | 53 +++++++++++++++++++++-
3 files changed, 115 insertions(+), 8 deletions(-)
diff --git a/drivers/infiniband/hw/mana/device.c b/drivers/infiniband/hw/mana/device.c
index 4077e440657a..e15da43c73a0 100644
--- a/drivers/infiniband/hw/mana/device.c
+++ b/drivers/infiniband/hw/mana/device.c
@@ -97,6 +97,10 @@ static int mana_ib_probe(struct auxiliary_device *adev,
goto free_error_eq;
}
+ ret = mana_ib_query_adapter_caps(mib_dev);
+ if (ret)
+ ibdev_dbg(&mib_dev->ib_dev, "Failed to get caps, use defaults");
+
ret = ib_register_device(&mib_dev->ib_dev, "mana_%d",
mdev->gdma_context->dev);
if (ret)
diff --git a/drivers/infiniband/hw/mana/main.c b/drivers/infiniband/hw/mana/main.c
index 1b1a8670d0fa..512815e1e64d 100644
--- a/drivers/infiniband/hw/mana/main.c
+++ b/drivers/infiniband/hw/mana/main.c
@@ -469,21 +469,27 @@ int mana_ib_get_port_immutable(struct ib_device *ibdev, u32 port_num,
int mana_ib_query_device(struct ib_device *ibdev, struct ib_device_attr *props,
struct ib_udata *uhw)
{
+ struct mana_ib_dev *mib_dev = container_of(ibdev,
+ struct mana_ib_dev, ib_dev);
+
props->max_qp = MANA_MAX_NUM_QUEUES;
props->max_qp_wr = MAX_SEND_BUFFERS_PER_QUEUE;
-
- /*
- * max_cqe could be potentially much bigger.
- * As this version of driver only support RAW QP, set it to the same
- * value as max_qp_wr
- */
props->max_cqe = MAX_SEND_BUFFERS_PER_QUEUE;
-
props->max_mr_size = MANA_IB_MAX_MR_SIZE;
props->max_mr = MANA_IB_MAX_MR;
props->max_send_sge = MAX_TX_WQE_SGL_ENTRIES;
props->max_recv_sge = MAX_RX_WQE_SGL_ENTRIES;
+ /* If the Management SW is updated and supports adapter creation */
+ if (mib_dev->adapter_handle) {
+ props->max_qp = mib_dev->adapter_caps.max_qp_count;
+ props->max_qp_wr = mib_dev->adapter_caps.max_requester_sq_size;
+ props->max_cqe = mib_dev->adapter_caps.max_requester_sq_size;
+ props->max_mr = mib_dev->adapter_caps.max_mr_count;
+ props->max_send_sge = mib_dev->adapter_caps.max_send_wqe_size;
+ props->max_recv_sge = mib_dev->adapter_caps.max_recv_wqe_size;
+ }
+
return 0;
}
@@ -599,3 +605,49 @@ int mana_ib_create_error_eq(struct mana_ib_dev *mib_dev)
return 0;
}
+
+static void assign_caps(struct mana_ib_adapter_caps *caps,
+ struct mana_ib_query_adapter_caps_resp *resp)
+{
+ caps->max_sq_id = resp->max_sq_id;
+ caps->max_rq_id = resp->max_rq_id;
+ caps->max_cq_id = resp->max_cq_id;
+ caps->max_qp_count = resp->max_qp_count;
+ caps->max_cq_count = resp->max_cq_count;
+ caps->max_mr_count = resp->max_mr_count;
+ caps->max_pd_count = resp->max_pd_count;
+ caps->max_inbound_read_limit = resp->max_inbound_read_limit;
+ caps->max_outbound_read_limit = resp->max_outbound_read_limit;
+ caps->mw_count = resp->mw_count;
+ caps->max_srq_count = resp->max_srq_count;
+ caps->max_requester_sq_size = resp->max_requester_sq_size;
+ caps->max_responder_sq_size = resp->max_responder_sq_size;
+ caps->max_requester_rq_size = resp->max_requester_rq_size;
+ caps->max_responder_rq_size = resp->max_responder_rq_size;
+ caps->max_send_wqe_size = resp->max_send_wqe_size;
+ caps->max_recv_wqe_size = resp->max_recv_wqe_size;
+ caps->max_inline_data_size = resp->max_inline_data_size;
+}
+
+int mana_ib_query_adapter_caps(struct mana_ib_dev *mib_dev)
+{
+ struct mana_ib_query_adapter_caps_resp resp = {};
+ struct mana_ib_query_adapter_caps_req req = {};
+ int err;
+
+ mana_gd_init_req_hdr(&req.hdr, MANA_IB_GET_ADAPTER_CAP, sizeof(req),
+ sizeof(resp));
+ req.hdr.resp.msg_version = MANA_IB__GET_ADAPTER_CAP_RESPONSE_V3;
+ req.hdr.dev_id = mib_dev->gc->mana_ib.dev_id;
+
+ err = mana_gd_send_request(mib_dev->gc, sizeof(req), &req,
+ sizeof(resp), &resp);
+
+ if (err) {
+ ibdev_err(&mib_dev->ib_dev, "Failed to query adapter caps err %d", err);
+ return err;
+ }
+
+ assign_caps(&mib_dev->adapter_caps, &resp);
+ return 0;
+}
diff --git a/drivers/infiniband/hw/mana/mana_ib.h b/drivers/infiniband/hw/mana/mana_ib.h
index 8a652bccd978..1044358230d3 100644
--- a/drivers/infiniband/hw/mana/mana_ib.h
+++ b/drivers/infiniband/hw/mana/mana_ib.h
@@ -20,19 +20,41 @@
/* MANA doesn't have any limit for MR size */
#define MANA_IB_MAX_MR_SIZE U64_MAX
-
+#define MANA_IB__GET_ADAPTER_CAP_RESPONSE_V3 3
/*
* The hardware limit of number of MRs is greater than maximum number of MRs
* that can possibly represent in 24 bits
*/
#define MANA_IB_MAX_MR 0xFFFFFFu
+struct mana_ib_adapter_caps {
+ u32 max_sq_id;
+ u32 max_rq_id;
+ u32 max_cq_id;
+ u32 max_qp_count;
+ u32 max_cq_count;
+ u32 max_mr_count;
+ u32 max_pd_count;
+ u32 max_inbound_read_limit;
+ u32 max_outbound_read_limit;
+ u32 mw_count;
+ u32 max_srq_count;
+ u32 max_requester_sq_size;
+ u32 max_responder_sq_size;
+ u32 max_requester_rq_size;
+ u32 max_responder_rq_size;
+ u32 max_send_wqe_size;
+ u32 max_recv_wqe_size;
+ u32 max_inline_data_size;
+};
+
struct mana_ib_dev {
struct ib_device ib_dev;
struct gdma_dev *gdma_dev;
struct gdma_context *gc;
struct gdma_queue *fatal_err_eq;
mana_handle_t adapter_handle;
+ struct mana_ib_adapter_caps adapter_caps;
};
struct mana_ib_wq {
@@ -96,6 +118,7 @@ struct mana_ib_rwq_ind_table {
};
enum mana_ib_command_code {
+ MANA_IB_GET_ADAPTER_CAP = 0x30001,
MANA_IB_CREATE_ADAPTER = 0x30002,
MANA_IB_DESTROY_ADAPTER = 0x30003,
};
@@ -120,6 +143,32 @@ struct mana_ib_destroy_adapter_resp {
struct gdma_resp_hdr hdr;
}; /* HW Data */
+struct mana_ib_query_adapter_caps_req {
+ struct gdma_req_hdr hdr;
+}; /*HW Data */
+
+struct mana_ib_query_adapter_caps_resp {
+ struct gdma_resp_hdr hdr;
+ u32 max_sq_id;
+ u32 max_rq_id;
+ u32 max_cq_id;
+ u32 max_qp_count;
+ u32 max_cq_count;
+ u32 max_mr_count;
+ u32 max_pd_count;
+ u32 max_inbound_read_limit;
+ u32 max_outbound_read_limit;
+ u32 mw_count;
+ u32 max_srq_count;
+ u32 max_requester_sq_size;
+ u32 max_responder_sq_size;
+ u32 max_requester_rq_size;
+ u32 max_responder_rq_size;
+ u32 max_send_wqe_size;
+ u32 max_recv_wqe_size;
+ u32 max_inline_data_size;
+}; /* HW Data */
+
int mana_ib_gd_create_dma_region(struct mana_ib_dev *mib_dev,
struct ib_umem *umem,
mana_handle_t *gdma_region);
@@ -194,4 +243,6 @@ int mana_ib_create_adapter(struct mana_ib_dev *mib_dev);
int mana_ib_destroy_adapter(struct mana_ib_dev *mib_dev);
+int mana_ib_query_adapter_caps(struct mana_ib_dev *mib_dev);
+
#endif
--
2.25.1
> Subject: [Patch v3 4/4] RDMA/mana_ib : Query adapter capabilities
>
> From: Ajay Sharma <sharmaajay@microsoft.com>
>
> Query the adapter capabilities to expose to other clients and VF. This checks
> against the user supplied values and protects against overflows.
>
> Signed-off-by: Ajay Sharma <sharmaajay@microsoft.com>
> ---
> drivers/infiniband/hw/mana/device.c | 4 ++
> drivers/infiniband/hw/mana/main.c | 66 +++++++++++++++++++++++++-
> --
> drivers/infiniband/hw/mana/mana_ib.h | 53 +++++++++++++++++++++-
> 3 files changed, 115 insertions(+), 8 deletions(-)
>
> diff --git a/drivers/infiniband/hw/mana/device.c
> b/drivers/infiniband/hw/mana/device.c
> index 4077e440657a..e15da43c73a0 100644
> --- a/drivers/infiniband/hw/mana/device.c
> +++ b/drivers/infiniband/hw/mana/device.c
> @@ -97,6 +97,10 @@ static int mana_ib_probe(struct auxiliary_device *adev,
> goto free_error_eq;
> }
>
> + ret = mana_ib_query_adapter_caps(mib_dev);
> + if (ret)
> + ibdev_dbg(&mib_dev->ib_dev, "Failed to get caps, use
> defaults");
There is an ibdev_err() in mana_ib_query_adapter_caps(), how about merging this message with that?
And you can remove the return value of mana_ib_query_adapter_caps(), since it doesn't do something meaningful.
> +
> ret = ib_register_device(&mib_dev->ib_dev, "mana_%d",
> mdev->gdma_context->dev);
> if (ret)
> diff --git a/drivers/infiniband/hw/mana/main.c
> b/drivers/infiniband/hw/mana/main.c
> index 1b1a8670d0fa..512815e1e64d 100644
> --- a/drivers/infiniband/hw/mana/main.c
> +++ b/drivers/infiniband/hw/mana/main.c
> @@ -469,21 +469,27 @@ int mana_ib_get_port_immutable(struct ib_device
> *ibdev, u32 port_num, int mana_ib_query_device(struct ib_device *ibdev,
> struct ib_device_attr *props,
> struct ib_udata *uhw)
> {
> + struct mana_ib_dev *mib_dev = container_of(ibdev,
> + struct mana_ib_dev, ib_dev);
> +
> props->max_qp = MANA_MAX_NUM_QUEUES;
> props->max_qp_wr = MAX_SEND_BUFFERS_PER_QUEUE;
> -
> - /*
> - * max_cqe could be potentially much bigger.
> - * As this version of driver only support RAW QP, set it to the same
> - * value as max_qp_wr
> - */
> props->max_cqe = MAX_SEND_BUFFERS_PER_QUEUE;
> -
> props->max_mr_size = MANA_IB_MAX_MR_SIZE;
> props->max_mr = MANA_IB_MAX_MR;
> props->max_send_sge = MAX_TX_WQE_SGL_ENTRIES;
> props->max_recv_sge = MAX_RX_WQE_SGL_ENTRIES;
>
> + /* If the Management SW is updated and supports adapter creation */
> + if (mib_dev->adapter_handle) {
Does this mean mana_ib_query_adapter_caps() was a success?
> + props->max_qp = mib_dev->adapter_caps.max_qp_count;
> + props->max_qp_wr = mib_dev-
> >adapter_caps.max_requester_sq_size;
> + props->max_cqe = mib_dev-
> >adapter_caps.max_requester_sq_size;
> + props->max_mr = mib_dev->adapter_caps.max_mr_count;
> + props->max_send_sge = mib_dev-
> >adapter_caps.max_send_wqe_size;
> + props->max_recv_sge = mib_dev-
> >adapter_caps.max_recv_wqe_size;
> + }
> +
> return 0;
> }
>
> @@ -599,3 +605,49 @@ int mana_ib_create_error_eq(struct mana_ib_dev
> *mib_dev)
>
> return 0;
> }
> +
> +static void assign_caps(struct mana_ib_adapter_caps *caps,
> + struct mana_ib_query_adapter_caps_resp *resp) {
> + caps->max_sq_id = resp->max_sq_id;
> + caps->max_rq_id = resp->max_rq_id;
> + caps->max_cq_id = resp->max_cq_id;
> + caps->max_qp_count = resp->max_qp_count;
> + caps->max_cq_count = resp->max_cq_count;
> + caps->max_mr_count = resp->max_mr_count;
> + caps->max_pd_count = resp->max_pd_count;
> + caps->max_inbound_read_limit = resp->max_inbound_read_limit;
> + caps->max_outbound_read_limit = resp->max_outbound_read_limit;
> + caps->mw_count = resp->mw_count;
> + caps->max_srq_count = resp->max_srq_count;
> + caps->max_requester_sq_size = resp->max_requester_sq_size;
> + caps->max_responder_sq_size = resp->max_responder_sq_size;
> + caps->max_requester_rq_size = resp->max_requester_rq_size;
> + caps->max_responder_rq_size = resp->max_responder_rq_size;
> + caps->max_send_wqe_size = resp->max_send_wqe_size;
> + caps->max_recv_wqe_size = resp->max_recv_wqe_size;
> + caps->max_inline_data_size = resp->max_inline_data_size; }
> +
> +int mana_ib_query_adapter_caps(struct mana_ib_dev *mib_dev) {
> + struct mana_ib_query_adapter_caps_resp resp = {};
> + struct mana_ib_query_adapter_caps_req req = {};
> + int err;
> +
> + mana_gd_init_req_hdr(&req.hdr, MANA_IB_GET_ADAPTER_CAP,
> sizeof(req),
> + sizeof(resp));
> + req.hdr.resp.msg_version =
> MANA_IB__GET_ADAPTER_CAP_RESPONSE_V3;
> + req.hdr.dev_id = mib_dev->gc->mana_ib.dev_id;
> +
> + err = mana_gd_send_request(mib_dev->gc, sizeof(req), &req,
> + sizeof(resp), &resp);
> +
> + if (err) {
> + ibdev_err(&mib_dev->ib_dev, "Failed to query adapter caps
> err %d", err);
> + return err;
> + }
> +
> + assign_caps(&mib_dev->adapter_caps, &resp);
> + return 0;
> +}
> diff --git a/drivers/infiniband/hw/mana/mana_ib.h
> b/drivers/infiniband/hw/mana/mana_ib.h
> index 8a652bccd978..1044358230d3 100644
> --- a/drivers/infiniband/hw/mana/mana_ib.h
> +++ b/drivers/infiniband/hw/mana/mana_ib.h
> @@ -20,19 +20,41 @@
>
> /* MANA doesn't have any limit for MR size */
> #define MANA_IB_MAX_MR_SIZE U64_MAX
> -
> +#define MANA_IB__GET_ADAPTER_CAP_RESPONSE_V3 3
This value is used in GDMA header of the request message? If so, define GDMA_MESSAGE_V3 in "include/net/mana/gdma.h".
> -----Original Message-----
> From: sharmaajay@linuxonhyperv.com <sharmaajay@linuxonhyperv.com>
> Sent: Wednesday, July 26, 2023 3:08 PM
> To: Jason Gunthorpe <jgg@ziepe.ca>; Leon Romanovsky <leon@kernel.org>;
> Dexuan Cui <decui@microsoft.com>; Wei Liu <wei.liu@kernel.org>; David S.
> Miller <davem@davemloft.net>; Eric Dumazet <edumazet@google.com>;
> Jakub Kicinski <kuba@kernel.org>; Paolo Abeni <pabeni@redhat.com>
> Cc: linux-rdma@vger.kernel.org; linux-hyperv@vger.kernel.org;
> netdev@vger.kernel.org; linux-kernel@vger.kernel.org; Ajay Sharma
> <sharmaajay@microsoft.com>
> Subject: [EXTERNAL] [Patch v3 4/4] RDMA/mana_ib : Query adapter
> capabilities
>
> From: Ajay Sharma <sharmaajay@microsoft.com>
>
> Query the adapter capabilities to expose to other clients and VF. This checks
> against the user supplied values and protects against overflows.
>
> Signed-off-by: Ajay Sharma <sharmaajay@microsoft.com>
> ---
> drivers/infiniband/hw/mana/device.c | 4 ++
> drivers/infiniband/hw/mana/main.c | 66 +++++++++++++++++++++++++---
> drivers/infiniband/hw/mana/mana_ib.h | 53 +++++++++++++++++++++-
> 3 files changed, 115 insertions(+), 8 deletions(-)
>
> diff --git a/drivers/infiniband/hw/mana/device.c
> b/drivers/infiniband/hw/mana/device.c
> index 4077e440657a..e15da43c73a0 100644
> --- a/drivers/infiniband/hw/mana/device.c
> +++ b/drivers/infiniband/hw/mana/device.c
> @@ -97,6 +97,10 @@ static int mana_ib_probe(struct auxiliary_device *adev,
> goto free_error_eq;
> }
>
> + ret = mana_ib_query_adapter_caps(mib_dev);
> + if (ret)
> + ibdev_dbg(&mib_dev->ib_dev, "Failed to get caps, use
> defaults");
> +
> ret = ib_register_device(&mib_dev->ib_dev, "mana_%d",
> mdev->gdma_context->dev);
> if (ret)
> diff --git a/drivers/infiniband/hw/mana/main.c
> b/drivers/infiniband/hw/mana/main.c
> index 1b1a8670d0fa..512815e1e64d 100644
> --- a/drivers/infiniband/hw/mana/main.c
> +++ b/drivers/infiniband/hw/mana/main.c
> @@ -469,21 +469,27 @@ int mana_ib_get_port_immutable(struct ib_device
> *ibdev, u32 port_num, int mana_ib_query_device(struct ib_device *ibdev,
> struct ib_device_attr *props,
> struct ib_udata *uhw)
> {
> + struct mana_ib_dev *mib_dev = container_of(ibdev,
> + struct mana_ib_dev, ib_dev);
> +
> props->max_qp = MANA_MAX_NUM_QUEUES;
> props->max_qp_wr = MAX_SEND_BUFFERS_PER_QUEUE;
> -
> - /*
> - * max_cqe could be potentially much bigger.
> - * As this version of driver only support RAW QP, set it to the same
> - * value as max_qp_wr
> - */
> props->max_cqe = MAX_SEND_BUFFERS_PER_QUEUE;
> -
> props->max_mr_size = MANA_IB_MAX_MR_SIZE;
> props->max_mr = MANA_IB_MAX_MR;
> props->max_send_sge = MAX_TX_WQE_SGL_ENTRIES;
> props->max_recv_sge = MAX_RX_WQE_SGL_ENTRIES;
>
> + /* If the Management SW is updated and supports adapter creation */
> + if (mib_dev->adapter_handle) {
> + props->max_qp = mib_dev->adapter_caps.max_qp_count;
> + props->max_qp_wr = mib_dev-
> >adapter_caps.max_requester_sq_size;
> + props->max_cqe = mib_dev-
> >adapter_caps.max_requester_sq_size;
> + props->max_mr = mib_dev->adapter_caps.max_mr_count;
> + props->max_send_sge = mib_dev-
> >adapter_caps.max_send_wqe_size;
> + props->max_recv_sge = mib_dev-
> >adapter_caps.max_recv_wqe_size;
> + }
> +
> return 0;
> }
>
> @@ -599,3 +605,49 @@ int mana_ib_create_error_eq(struct mana_ib_dev
> *mib_dev)
>
> return 0;
> }
> +
> +static void assign_caps(struct mana_ib_adapter_caps *caps,
> + struct mana_ib_query_adapter_caps_resp *resp) {
> + caps->max_sq_id = resp->max_sq_id;
> + caps->max_rq_id = resp->max_rq_id;
> + caps->max_cq_id = resp->max_cq_id;
> + caps->max_qp_count = resp->max_qp_count;
> + caps->max_cq_count = resp->max_cq_count;
> + caps->max_mr_count = resp->max_mr_count;
> + caps->max_pd_count = resp->max_pd_count;
> + caps->max_inbound_read_limit = resp->max_inbound_read_limit;
> + caps->max_outbound_read_limit = resp->max_outbound_read_limit;
> + caps->mw_count = resp->mw_count;
> + caps->max_srq_count = resp->max_srq_count;
> + caps->max_requester_sq_size = resp->max_requester_sq_size;
> + caps->max_responder_sq_size = resp->max_responder_sq_size;
> + caps->max_requester_rq_size = resp->max_requester_rq_size;
> + caps->max_responder_rq_size = resp->max_responder_rq_size;
> + caps->max_send_wqe_size = resp->max_send_wqe_size;
> + caps->max_recv_wqe_size = resp->max_recv_wqe_size;
> + caps->max_inline_data_size = resp->max_inline_data_size; }
> +
> +int mana_ib_query_adapter_caps(struct mana_ib_dev *mib_dev) {
> + struct mana_ib_query_adapter_caps_resp resp = {};
> + struct mana_ib_query_adapter_caps_req req = {};
> + int err;
> +
> + mana_gd_init_req_hdr(&req.hdr, MANA_IB_GET_ADAPTER_CAP,
> sizeof(req),
> + sizeof(resp));
> + req.hdr.resp.msg_version =
> MANA_IB__GET_ADAPTER_CAP_RESPONSE_V3;
> + req.hdr.dev_id = mib_dev->gc->mana_ib.dev_id;
> +
> + err = mana_gd_send_request(mib_dev->gc, sizeof(req), &req,
> + sizeof(resp), &resp);
> +
> + if (err) {
> + ibdev_err(&mib_dev->ib_dev, "Failed to query adapter caps
> err %d", err);
> + return err;
> + }
> +
> + assign_caps(&mib_dev->adapter_caps, &resp);
> + return 0;
> +}
> diff --git a/drivers/infiniband/hw/mana/mana_ib.h
> b/drivers/infiniband/hw/mana/mana_ib.h
> index 8a652bccd978..1044358230d3 100644
> --- a/drivers/infiniband/hw/mana/mana_ib.h
> +++ b/drivers/infiniband/hw/mana/mana_ib.h
> @@ -20,19 +20,41 @@
>
> /* MANA doesn't have any limit for MR size */
> #define MANA_IB_MAX_MR_SIZE U64_MAX
> -
> +#define MANA_IB__GET_ADAPTER_CAP_RESPONSE_V3 3
> /*
> * The hardware limit of number of MRs is greater than maximum number of
> MRs
> * that can possibly represent in 24 bits
> */
> #define MANA_IB_MAX_MR 0xFFFFFFu
>
> +struct mana_ib_adapter_caps {
> + u32 max_sq_id;
> + u32 max_rq_id;
> + u32 max_cq_id;
> + u32 max_qp_count;
> + u32 max_cq_count;
> + u32 max_mr_count;
> + u32 max_pd_count;
> + u32 max_inbound_read_limit;
> + u32 max_outbound_read_limit;
> + u32 mw_count;
> + u32 max_srq_count;
> + u32 max_requester_sq_size;
> + u32 max_responder_sq_size;
> + u32 max_requester_rq_size;
> + u32 max_responder_rq_size;
> + u32 max_send_wqe_size;
> + u32 max_recv_wqe_size;
> + u32 max_inline_data_size;
> +};
> +
> struct mana_ib_dev {
> struct ib_device ib_dev;
> struct gdma_dev *gdma_dev;
> struct gdma_context *gc;
> struct gdma_queue *fatal_err_eq;
> mana_handle_t adapter_handle;
> + struct mana_ib_adapter_caps adapter_caps;
> };
>
> struct mana_ib_wq {
> @@ -96,6 +118,7 @@ struct mana_ib_rwq_ind_table { };
>
> enum mana_ib_command_code {
> + MANA_IB_GET_ADAPTER_CAP = 0x30001,
> MANA_IB_CREATE_ADAPTER = 0x30002,
> MANA_IB_DESTROY_ADAPTER = 0x30003,
> };
> @@ -120,6 +143,32 @@ struct mana_ib_destroy_adapter_resp {
> struct gdma_resp_hdr hdr;
> }; /* HW Data */
>
> +struct mana_ib_query_adapter_caps_req {
> + struct gdma_req_hdr hdr;
> +}; /*HW Data */
> +
> +struct mana_ib_query_adapter_caps_resp {
> + struct gdma_resp_hdr hdr;
> + u32 max_sq_id;
> + u32 max_rq_id;
> + u32 max_cq_id;
> + u32 max_qp_count;
> + u32 max_cq_count;
> + u32 max_mr_count;
> + u32 max_pd_count;
> + u32 max_inbound_read_limit;
> + u32 max_outbound_read_limit;
> + u32 mw_count;
> + u32 max_srq_count;
> + u32 max_requester_sq_size;
> + u32 max_responder_sq_size;
> + u32 max_requester_rq_size;
> + u32 max_responder_rq_size;
> + u32 max_send_wqe_size;
> + u32 max_recv_wqe_size;
> + u32 max_inline_data_size;
> +}; /* HW Data */
> +
> int mana_ib_gd_create_dma_region(struct mana_ib_dev *mib_dev,
> struct ib_umem *umem,
> mana_handle_t *gdma_region);
> @@ -194,4 +243,6 @@ int mana_ib_create_adapter(struct mana_ib_dev
> *mib_dev);
>
> int mana_ib_destroy_adapter(struct mana_ib_dev *mib_dev);
>
> +int mana_ib_query_adapter_caps(struct mana_ib_dev *mib_dev);
> +
> #endif
> --
> 2.25.1
© 2016 - 2026 Red Hat, Inc.