nvme/rdma: Add poller statistics

New statistics include number of poller calls,
number of idle polls and total number of completions.
These statistics allow to estimate % of idle polls
and the number of completions per poll.

Since nvme_rdma_cq_process_completions function
returns number of completed NVMF requests and each
NVMF request consumes 2 RDMA completions (send+recv),
this function was extended to return the number of
RDMA completions.

Signed-off-by: Alexey Marchuk <alexeymar@mellanox.com>
Change-Id: Ifdc1e2e467f645adb5d66d39ff2a379e161fbd77
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/6298
Community-CI: Broadcom CI
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com>
Reviewed-by: Jim Harris <james.r.harris@intel.com>
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
This commit is contained in:
Alexey Marchuk 2020-12-04 09:58:17 +03:00 committed by Tomasz Zawadzki
parent 527f406b6b
commit 50569293ef

View File

@ -149,6 +149,10 @@ struct nvme_rdma_destroyed_qpair {
}; };
struct nvme_rdma_poller_stats { struct nvme_rdma_poller_stats {
uint64_t polls;
uint64_t idle_polls;
uint64_t queued_requests;
uint64_t completions;
struct spdk_rdma_qp_stats rdma_stats; struct spdk_rdma_qp_stats rdma_stats;
}; };
@ -1897,7 +1901,10 @@ nvme_rdma_qpair_submit_request(struct spdk_nvme_qpair *qpair,
assert(req != NULL); assert(req != NULL);
rdma_req = nvme_rdma_req_get(rqpair); rdma_req = nvme_rdma_req_get(rqpair);
if (!rdma_req) { if (spdk_unlikely(!rdma_req)) {
if (rqpair->poller) {
rqpair->poller->stats.queued_requests++;
}
/* Inform the upper layer to try again later. */ /* Inform the upper layer to try again later. */
return -EAGAIN; return -EAGAIN;
} }
@ -2029,7 +2036,8 @@ nvme_rdma_conditional_fail_qpair(struct nvme_rdma_qpair *rqpair, struct nvme_rdm
static int static int
nvme_rdma_cq_process_completions(struct ibv_cq *cq, uint32_t batch_size, nvme_rdma_cq_process_completions(struct ibv_cq *cq, uint32_t batch_size,
struct nvme_rdma_poll_group *group, struct nvme_rdma_poll_group *group,
struct nvme_rdma_qpair *rdma_qpair) struct nvme_rdma_qpair *rdma_qpair,
uint64_t *rdma_completions)
{ {
struct ibv_wc wc[MAX_COMPLETIONS_PER_POLL]; struct ibv_wc wc[MAX_COMPLETIONS_PER_POLL];
struct nvme_rdma_qpair *rqpair; struct nvme_rdma_qpair *rqpair;
@ -2132,6 +2140,8 @@ nvme_rdma_cq_process_completions(struct ibv_cq *cq, uint32_t batch_size,
} }
} }
*rdma_completions += rc;
if (completion_rc) { if (completion_rc) {
return completion_rc; return completion_rc;
} }
@ -2153,6 +2163,7 @@ nvme_rdma_qpair_process_completions(struct spdk_nvme_qpair *qpair,
int rc = 0, batch_size; int rc = 0, batch_size;
struct ibv_cq *cq; struct ibv_cq *cq;
struct nvme_rdma_ctrlr *rctrlr; struct nvme_rdma_ctrlr *rctrlr;
uint64_t rdma_completions = 0;
/* /*
* This is used during the connection phase. It's possible that we are still reaping error completions * This is used during the connection phase. It's possible that we are still reaping error completions
@ -2186,7 +2197,7 @@ nvme_rdma_qpair_process_completions(struct spdk_nvme_qpair *qpair,
rqpair->num_completions = 0; rqpair->num_completions = 0;
do { do {
batch_size = spdk_min((max_completions - rqpair->num_completions), MAX_COMPLETIONS_PER_POLL); batch_size = spdk_min((max_completions - rqpair->num_completions), MAX_COMPLETIONS_PER_POLL);
rc = nvme_rdma_cq_process_completions(cq, batch_size, NULL, rqpair); rc = nvme_rdma_cq_process_completions(cq, batch_size, NULL, rqpair, &rdma_completions);
if (rc == 0) { if (rc == 0) {
break; break;
@ -2530,6 +2541,7 @@ nvme_rdma_poll_group_process_completions(struct spdk_nvme_transport_poll_group *
uint64_t completions_allowed = 0; uint64_t completions_allowed = 0;
uint64_t completions_per_poller = 0; uint64_t completions_per_poller = 0;
uint64_t poller_completions = 0; uint64_t poller_completions = 0;
uint64_t rdma_completions;
if (completions_per_qpair == 0) { if (completions_per_qpair == 0) {
@ -2559,12 +2571,16 @@ nvme_rdma_poll_group_process_completions(struct spdk_nvme_transport_poll_group *
STAILQ_FOREACH(poller, &group->pollers, link) { STAILQ_FOREACH(poller, &group->pollers, link) {
poller_completions = 0; poller_completions = 0;
rdma_completions = 0;
do { do {
poller->stats.polls++;
batch_size = spdk_min((completions_per_poller - poller_completions), MAX_COMPLETIONS_PER_POLL); batch_size = spdk_min((completions_per_poller - poller_completions), MAX_COMPLETIONS_PER_POLL);
rc = nvme_rdma_cq_process_completions(poller->cq, batch_size, group, NULL); rc = nvme_rdma_cq_process_completions(poller->cq, batch_size, group, NULL, &rdma_completions);
if (rc <= 0) { if (rc <= 0) {
if (rc == -ECANCELED) { if (rc == -ECANCELED) {
return -EIO; return -EIO;
} else if (rc == 0) {
poller->stats.idle_polls++;
} }
break; break;
} }
@ -2572,6 +2588,7 @@ nvme_rdma_poll_group_process_completions(struct spdk_nvme_transport_poll_group *
poller_completions += rc; poller_completions += rc;
} while (poller_completions < completions_per_poller); } while (poller_completions < completions_per_poller);
total_completions += poller_completions; total_completions += poller_completions;
poller->stats.completions += rdma_completions;
} }
STAILQ_FOREACH_SAFE(qpair, &tgroup->connected_qpairs, poll_group_stailq, tmp_qpair) { STAILQ_FOREACH_SAFE(qpair, &tgroup->connected_qpairs, poll_group_stailq, tmp_qpair) {