nvmf/vfio-user: use a list for connected queue pairs

We can avoid extra iteration to null queue pairs by using a list.

Change-Id: Idebd5a396959372004c33f8b21c555032af67aae
Signed-off-by: Changpeng Liu <changpeng.liu@intel.com>
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/9350
Community-CI: Mellanox Build Bot
Community-CI: Broadcom CI <spdk-ci.pdl@broadcom.com>
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
Reviewed-by: Jim Harris <james.r.harris@intel.com>
This commit is contained in:
Changpeng Liu 2021-08-25 18:11:41 +08:00 committed by Keith Lucas
parent ae131ad0c9
commit 3cc3dba342

View File

@ -149,7 +149,10 @@ struct nvmf_vfio_user_qpair {
struct spdk_nvme_cmd create_io_sq_cmd; struct spdk_nvme_cmd create_io_sq_cmd;
TAILQ_HEAD(, nvmf_vfio_user_req) reqs; TAILQ_HEAD(, nvmf_vfio_user_req) reqs;
/* Poll group entry */
TAILQ_ENTRY(nvmf_vfio_user_qpair) link; TAILQ_ENTRY(nvmf_vfio_user_qpair) link;
/* Connected queue pair entry */
TAILQ_ENTRY(nvmf_vfio_user_qpair) tailq;
}; };
struct nvmf_vfio_user_poll_group { struct nvmf_vfio_user_poll_group {
@ -161,8 +164,8 @@ struct nvmf_vfio_user_ctrlr {
struct nvmf_vfio_user_endpoint *endpoint; struct nvmf_vfio_user_endpoint *endpoint;
struct nvmf_vfio_user_transport *transport; struct nvmf_vfio_user_transport *transport;
/* Number of connected queue pairs */ /* Connected queue pairs list */
uint32_t num_connected_qps; TAILQ_HEAD(, nvmf_vfio_user_qpair) connected_qps;
struct spdk_thread *thread; struct spdk_thread *thread;
struct spdk_poller *vfu_ctx_poller; struct spdk_poller *vfu_ctx_poller;
@ -1427,7 +1430,7 @@ memory_region_add_cb(vfu_ctx_t *vfu_ctx, vfu_dma_info_t *info)
struct nvmf_vfio_user_endpoint *endpoint = vfu_get_private(vfu_ctx); struct nvmf_vfio_user_endpoint *endpoint = vfu_get_private(vfu_ctx);
struct nvmf_vfio_user_ctrlr *ctrlr; struct nvmf_vfio_user_ctrlr *ctrlr;
struct nvmf_vfio_user_qpair *qpair; struct nvmf_vfio_user_qpair *qpair;
int i, ret; int ret;
/* /*
* We're not interested in any DMA regions that aren't mappable (we don't * We're not interested in any DMA regions that aren't mappable (we don't
@ -1468,12 +1471,8 @@ memory_region_add_cb(vfu_ctx_t *vfu_ctx, vfu_dma_info_t *info)
} }
} }
for (i = 0; i < NVMF_VFIO_USER_DEFAULT_MAX_QPAIRS_PER_CTRLR; i++) { pthread_mutex_lock(&endpoint->lock);
qpair = ctrlr->qp[i]; TAILQ_FOREACH(qpair, &ctrlr->connected_qps, tailq) {
if (qpair == NULL) {
continue;
}
if (qpair->state != VFIO_USER_QPAIR_INACTIVE) { if (qpair->state != VFIO_USER_QPAIR_INACTIVE) {
continue; continue;
} }
@ -1483,19 +1482,19 @@ memory_region_add_cb(vfu_ctx_t *vfu_ctx, vfu_dma_info_t *info)
continue; continue;
} }
qpair->state = VFIO_USER_QPAIR_ACTIVE; qpair->state = VFIO_USER_QPAIR_ACTIVE;
SPDK_DEBUGLOG(nvmf_vfio, "Remap QP %u successfully\n", i); SPDK_DEBUGLOG(nvmf_vfio, "Remap QP %u successfully\n", qpair->qpair.qid);
} }
pthread_mutex_unlock(&endpoint->lock);
} }
static int static int
memory_region_remove_cb(vfu_ctx_t *vfu_ctx, vfu_dma_info_t *info) memory_region_remove_cb(vfu_ctx_t *vfu_ctx, vfu_dma_info_t *info)
{ {
struct nvmf_vfio_user_endpoint *endpoint = vfu_get_private(vfu_ctx); struct nvmf_vfio_user_endpoint *endpoint = vfu_get_private(vfu_ctx);
struct nvmf_vfio_user_ctrlr *ctrlr; struct nvmf_vfio_user_ctrlr *ctrlr;
struct nvmf_vfio_user_qpair *qpair; struct nvmf_vfio_user_qpair *qpair;
void *map_start, *map_end; void *map_start, *map_end;
int i, ret; int ret = 0;
if (!info->vaddr) { if (!info->vaddr) {
return 0; return 0;
@ -1531,18 +1530,16 @@ memory_region_remove_cb(vfu_ctx_t *vfu_ctx, vfu_dma_info_t *info)
map_start = info->mapping.iov_base; map_start = info->mapping.iov_base;
map_end = info->mapping.iov_base + info->mapping.iov_len; map_end = info->mapping.iov_base + info->mapping.iov_len;
for (i = 0; i < NVMF_VFIO_USER_DEFAULT_MAX_QPAIRS_PER_CTRLR; i++) {
qpair = ctrlr->qp[i];
if (qpair == NULL) {
continue;
}
pthread_mutex_lock(&endpoint->lock);
TAILQ_FOREACH(qpair, &ctrlr->connected_qps, tailq) {
if ((qpair->cq.addr >= map_start && qpair->cq.addr < map_end) || if ((qpair->cq.addr >= map_start && qpair->cq.addr < map_end) ||
(qpair->sq.addr >= map_start && qpair->sq.addr < map_end)) { (qpair->sq.addr >= map_start && qpair->sq.addr < map_end)) {
unmap_qp(qpair); unmap_qp(qpair);
qpair->state = VFIO_USER_QPAIR_INACTIVE; qpair->state = VFIO_USER_QPAIR_INACTIVE;
} }
} }
pthread_mutex_unlock(&endpoint->lock);
return 0; return 0;
} }
@ -1984,6 +1981,7 @@ nvmf_vfio_user_create_ctrlr(struct nvmf_vfio_user_transport *transport,
ctrlr->transport = transport; ctrlr->transport = transport;
ctrlr->endpoint = endpoint; ctrlr->endpoint = endpoint;
ctrlr->doorbells = endpoint->doorbells; ctrlr->doorbells = endpoint->doorbells;
TAILQ_INIT(&ctrlr->connected_qps);
/* Then, construct an admin queue pair */ /* Then, construct an admin queue pair */
err = init_qp(ctrlr, &transport->transport, NVMF_VFIO_USER_DEFAULT_AQ_DEPTH, 0); err = init_qp(ctrlr, &transport->transport, NVMF_VFIO_USER_DEFAULT_AQ_DEPTH, 0);
@ -2262,7 +2260,7 @@ vfio_user_qpair_disconnect_cb(void *ctx)
return; return;
} }
if (!ctrlr->num_connected_qps) { if (TAILQ_EMPTY(&ctrlr->connected_qps)) {
endpoint->ctrlr = NULL; endpoint->ctrlr = NULL;
free_ctrlr(ctrlr); free_ctrlr(ctrlr);
pthread_mutex_unlock(&endpoint->lock); pthread_mutex_unlock(&endpoint->lock);
@ -2274,7 +2272,6 @@ vfio_user_qpair_disconnect_cb(void *ctx)
static int static int
vfio_user_destroy_ctrlr(struct nvmf_vfio_user_ctrlr *ctrlr) vfio_user_destroy_ctrlr(struct nvmf_vfio_user_ctrlr *ctrlr)
{ {
uint32_t i;
struct nvmf_vfio_user_qpair *qpair; struct nvmf_vfio_user_qpair *qpair;
struct nvmf_vfio_user_endpoint *endpoint; struct nvmf_vfio_user_endpoint *endpoint;
@ -2284,21 +2281,17 @@ vfio_user_destroy_ctrlr(struct nvmf_vfio_user_ctrlr *ctrlr)
assert(endpoint != NULL); assert(endpoint != NULL);
pthread_mutex_lock(&endpoint->lock); pthread_mutex_lock(&endpoint->lock);
if (ctrlr->num_connected_qps == 0) { if (TAILQ_EMPTY(&ctrlr->connected_qps)) {
endpoint->ctrlr = NULL; endpoint->ctrlr = NULL;
free_ctrlr(ctrlr); free_ctrlr(ctrlr);
pthread_mutex_unlock(&endpoint->lock); pthread_mutex_unlock(&endpoint->lock);
return 0; return 0;
} }
pthread_mutex_unlock(&endpoint->lock);
for (i = 0; i < NVMF_VFIO_USER_DEFAULT_MAX_QPAIRS_PER_CTRLR; i++) { TAILQ_FOREACH(qpair, &ctrlr->connected_qps, tailq) {
qpair = ctrlr->qp[i];
if (qpair == NULL) {
continue;
}
spdk_nvmf_qpair_disconnect(&qpair->qpair, vfio_user_qpair_disconnect_cb, endpoint); spdk_nvmf_qpair_disconnect(&qpair->qpair, vfio_user_qpair_disconnect_cb, endpoint);
} }
pthread_mutex_unlock(&endpoint->lock);
return 0; return 0;
} }
@ -2372,7 +2365,7 @@ handle_queue_connect_rsp(struct nvmf_vfio_user_req *req, void *cb_arg)
post_completion(ctrlr, &ctrlr->qp[0]->cq, 0, 0, post_completion(ctrlr, &ctrlr->qp[0]->cq, 0, 0,
qpair->create_io_sq_cmd.cid, SPDK_NVME_SC_SUCCESS, SPDK_NVME_SCT_GENERIC); qpair->create_io_sq_cmd.cid, SPDK_NVME_SC_SUCCESS, SPDK_NVME_SCT_GENERIC);
} }
ctrlr->num_connected_qps++; TAILQ_INSERT_TAIL(&ctrlr->connected_qps, qpair, tailq);
pthread_mutex_unlock(&endpoint->lock); pthread_mutex_unlock(&endpoint->lock);
free(req->req.data); free(req->req.data);
@ -2465,8 +2458,7 @@ nvmf_vfio_user_poll_group_remove(struct spdk_nvmf_transport_poll_group *group,
TAILQ_REMOVE(&vu_group->qps, vu_qpair, link); TAILQ_REMOVE(&vu_group->qps, vu_qpair, link);
pthread_mutex_lock(&endpoint->lock); pthread_mutex_lock(&endpoint->lock);
assert(vu_ctrlr->num_connected_qps); TAILQ_REMOVE(&vu_ctrlr->connected_qps, vu_qpair, tailq);
vu_ctrlr->num_connected_qps--;
pthread_mutex_unlock(&endpoint->lock); pthread_mutex_unlock(&endpoint->lock);
return 0; return 0;