lib/nvmf: remove spdk prefix from static functions in rdma.c

Signed-off-by: Seth Howell <seth.howell@intel.com>
Change-Id: I651924850083db8b6a84bdee4c9efd1103551daf
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/2287
Reviewed-by: Jim Harris <james.r.harris@intel.com>
Reviewed-by: Aleksey Marchuk <alexeymar@mellanox.com>
Reviewed-by: Darek Stojaczyk <dariusz.stojaczyk@intel.com>
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Community-CI: Mellanox Build Bot
This commit is contained in:
Seth Howell 2020-05-09 15:00:50 -07:00 committed by Jim Harris
parent e569a5785d
commit 55d8d943a8
2 changed files with 223 additions and 223 deletions

File diff suppressed because it is too large Load Diff

View File

@ -214,7 +214,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
sgl->keyed.length = rtransport.transport.opts.io_unit_size / 2; sgl->keyed.length = rtransport.transport.opts.io_unit_size / 2;
device.map = (void *)0x0; device.map = (void *)0x0;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
CU_ASSERT(rdma_req.req.length == rtransport.transport.opts.io_unit_size / 2); CU_ASSERT(rdma_req.req.length == rtransport.transport.opts.io_unit_size / 2);
@ -230,7 +230,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
/* Part 2: simple I/O, one SGL larger than the transport io unit size (equal to the max io size) */ /* Part 2: simple I/O, one SGL larger than the transport io unit size (equal to the max io size) */
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
sgl->keyed.length = rtransport.transport.opts.io_unit_size * RDMA_UT_UNITS_IN_MAX_IO; sgl->keyed.length = rtransport.transport.opts.io_unit_size * RDMA_UT_UNITS_IN_MAX_IO;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -248,7 +248,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
/* Part 3: simple I/O one SGL larger than the transport max io size */ /* Part 3: simple I/O one SGL larger than the transport max io size */
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
sgl->keyed.length = rtransport.transport.opts.max_io_size * 2; sgl->keyed.length = rtransport.transport.opts.max_io_size * 2;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == -1); CU_ASSERT(rc == -1);
@ -256,7 +256,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
MOCK_SET(spdk_mempool_get, NULL); MOCK_SET(spdk_mempool_get, NULL);
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
sgl->keyed.length = rtransport.transport.opts.io_unit_size * RDMA_UT_UNITS_IN_MAX_IO; sgl->keyed.length = rtransport.transport.opts.io_unit_size * RDMA_UT_UNITS_IN_MAX_IO;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == false); CU_ASSERT(rdma_req.req.data_from_pool == false);
@ -276,7 +276,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
sgl->address = 0; sgl->address = 0;
sgl->unkeyed.length = rtransport.transport.opts.in_capsule_data_size; sgl->unkeyed.length = rtransport.transport.opts.in_capsule_data_size;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data == (void *)0xDDDD); CU_ASSERT(rdma_req.req.data == (void *)0xDDDD);
@ -287,7 +287,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
sgl->address = rtransport.transport.opts.in_capsule_data_size; sgl->address = rtransport.transport.opts.in_capsule_data_size;
sgl->unkeyed.length = rtransport.transport.opts.in_capsule_data_size; sgl->unkeyed.length = rtransport.transport.opts.in_capsule_data_size;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == -1); CU_ASSERT(rc == -1);
@ -295,7 +295,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
sgl->address = 0; sgl->address = 0;
sgl->unkeyed.length = rtransport.transport.opts.in_capsule_data_size * 2; sgl->unkeyed.length = rtransport.transport.opts.in_capsule_data_size * 2;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == -1); CU_ASSERT(rc == -1);
@ -317,7 +317,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
sgl_desc[i].keyed.key = 0x44; sgl_desc[i].keyed.key = 0x44;
} }
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -342,7 +342,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
sgl_desc[i].keyed.key = 0x44; sgl_desc[i].keyed.key = 0x44;
} }
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -372,7 +372,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
sgl_desc[1].address = 0x4000 + rtransport.transport.opts.io_unit_size * 15 + sgl_desc[1].address = 0x4000 + rtransport.transport.opts.io_unit_size * 15 +
rtransport.transport.opts.io_unit_size / 2; rtransport.transport.opts.io_unit_size / 2;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -411,7 +411,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
MOCK_SET(spdk_mempool_get, (void *)0x2000); MOCK_SET(spdk_mempool_get, (void *)0x2000);
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
sgl->keyed.length = rtransport.transport.opts.io_unit_size * 4; sgl->keyed.length = rtransport.transport.opts.io_unit_size * 4;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
SPDK_CU_ASSERT_FATAL(rc == 0); SPDK_CU_ASSERT_FATAL(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -432,7 +432,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
/* part 2: now that we have used the buffers from the cache, try again. We should get mempool buffers. */ /* part 2: now that we have used the buffers from the cache, try again. We should get mempool buffers. */
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
SPDK_CU_ASSERT_FATAL(rc == 0); SPDK_CU_ASSERT_FATAL(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -457,7 +457,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
STAILQ_INSERT_TAIL(&group.group.buf_cache, &bufs[i], link); STAILQ_INSERT_TAIL(&group.group.buf_cache, &bufs[i], link);
} }
reset_nvmf_rdma_request(&rdma_req); reset_nvmf_rdma_request(&rdma_req);
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
SPDK_CU_ASSERT_FATAL(rc == 0); SPDK_CU_ASSERT_FATAL(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -489,7 +489,7 @@ test_spdk_nvmf_rdma_request_parse_sgl(void)
g_mr_size = rtransport.transport.opts.io_unit_size / 4; g_mr_size = rtransport.transport.opts.io_unit_size / 4;
g_mr_next_size = rtransport.transport.opts.io_unit_size / 2; g_mr_next_size = rtransport.transport.opts.io_unit_size / 2;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
SPDK_CU_ASSERT_FATAL(rc == 0); SPDK_CU_ASSERT_FATAL(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
CU_ASSERT(rdma_req.req.length == rtransport.transport.opts.io_unit_size / 2); CU_ASSERT(rdma_req.req.length == rtransport.transport.opts.io_unit_size / 2);
@ -636,13 +636,13 @@ test_spdk_nvmf_rdma_request_process(void)
rdma_req = create_req(&rqpair, rdma_recv); rdma_req = create_req(&rqpair, rdma_recv);
rqpair.current_recv_depth = 1; rqpair.current_recv_depth = 1;
/* NEW -> EXECUTING */ /* NEW -> EXECUTING */
progress = spdk_nvmf_rdma_request_process(&rtransport, rdma_req); progress = nvmf_rdma_request_process(&rtransport, rdma_req);
CU_ASSERT(progress == true); CU_ASSERT(progress == true);
CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_EXECUTING); CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_EXECUTING);
CU_ASSERT(rdma_req->req.xfer == SPDK_NVME_DATA_CONTROLLER_TO_HOST); CU_ASSERT(rdma_req->req.xfer == SPDK_NVME_DATA_CONTROLLER_TO_HOST);
/* EXECUTED -> TRANSFERRING_C2H */ /* EXECUTED -> TRANSFERRING_C2H */
rdma_req->state = RDMA_REQUEST_STATE_EXECUTED; rdma_req->state = RDMA_REQUEST_STATE_EXECUTED;
progress = spdk_nvmf_rdma_request_process(&rtransport, rdma_req); progress = nvmf_rdma_request_process(&rtransport, rdma_req);
CU_ASSERT(progress == true); CU_ASSERT(progress == true);
CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_TRANSFERRING_CONTROLLER_TO_HOST); CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_TRANSFERRING_CONTROLLER_TO_HOST);
CU_ASSERT(rdma_req->recv == NULL); CU_ASSERT(rdma_req->recv == NULL);
@ -652,7 +652,7 @@ test_spdk_nvmf_rdma_request_process(void)
CU_ASSERT(resources.recvs_to_post.last == &rdma_recv->wr); CU_ASSERT(resources.recvs_to_post.last == &rdma_recv->wr);
/* COMPLETED -> FREE */ /* COMPLETED -> FREE */
rdma_req->state = RDMA_REQUEST_STATE_COMPLETED; rdma_req->state = RDMA_REQUEST_STATE_COMPLETED;
progress = spdk_nvmf_rdma_request_process(&rtransport, rdma_req); progress = nvmf_rdma_request_process(&rtransport, rdma_req);
CU_ASSERT(progress == true); CU_ASSERT(progress == true);
CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_FREE); CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_FREE);
@ -666,7 +666,7 @@ test_spdk_nvmf_rdma_request_process(void)
rdma_req = create_req(&rqpair, rdma_recv); rdma_req = create_req(&rqpair, rdma_recv);
rqpair.current_recv_depth = 1; rqpair.current_recv_depth = 1;
/* NEW -> TRANSFERRING_H2C */ /* NEW -> TRANSFERRING_H2C */
progress = spdk_nvmf_rdma_request_process(&rtransport, rdma_req); progress = nvmf_rdma_request_process(&rtransport, rdma_req);
CU_ASSERT(progress == true); CU_ASSERT(progress == true);
CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_TRANSFERRING_HOST_TO_CONTROLLER); CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_TRANSFERRING_HOST_TO_CONTROLLER);
CU_ASSERT(rdma_req->req.xfer == SPDK_NVME_DATA_HOST_TO_CONTROLLER); CU_ASSERT(rdma_req->req.xfer == SPDK_NVME_DATA_HOST_TO_CONTROLLER);
@ -676,12 +676,12 @@ test_spdk_nvmf_rdma_request_process(void)
STAILQ_INIT(&poller.qpairs_pending_send); STAILQ_INIT(&poller.qpairs_pending_send);
/* READY_TO_EXECUTE -> EXECUTING */ /* READY_TO_EXECUTE -> EXECUTING */
rdma_req->state = RDMA_REQUEST_STATE_READY_TO_EXECUTE; rdma_req->state = RDMA_REQUEST_STATE_READY_TO_EXECUTE;
progress = spdk_nvmf_rdma_request_process(&rtransport, rdma_req); progress = nvmf_rdma_request_process(&rtransport, rdma_req);
CU_ASSERT(progress == true); CU_ASSERT(progress == true);
CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_EXECUTING); CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_EXECUTING);
/* EXECUTED -> COMPLETING */ /* EXECUTED -> COMPLETING */
rdma_req->state = RDMA_REQUEST_STATE_EXECUTED; rdma_req->state = RDMA_REQUEST_STATE_EXECUTED;
progress = spdk_nvmf_rdma_request_process(&rtransport, rdma_req); progress = nvmf_rdma_request_process(&rtransport, rdma_req);
CU_ASSERT(progress == true); CU_ASSERT(progress == true);
CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_COMPLETING); CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_COMPLETING);
CU_ASSERT(rdma_req->recv == NULL); CU_ASSERT(rdma_req->recv == NULL);
@ -691,7 +691,7 @@ test_spdk_nvmf_rdma_request_process(void)
CU_ASSERT(resources.recvs_to_post.last == &rdma_recv->wr); CU_ASSERT(resources.recvs_to_post.last == &rdma_recv->wr);
/* COMPLETED -> FREE */ /* COMPLETED -> FREE */
rdma_req->state = RDMA_REQUEST_STATE_COMPLETED; rdma_req->state = RDMA_REQUEST_STATE_COMPLETED;
progress = spdk_nvmf_rdma_request_process(&rtransport, rdma_req); progress = nvmf_rdma_request_process(&rtransport, rdma_req);
CU_ASSERT(progress == true); CU_ASSERT(progress == true);
CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_FREE); CU_ASSERT(rdma_req->state == RDMA_REQUEST_STATE_FREE);
@ -711,7 +711,7 @@ test_spdk_nvmf_rdma_request_process(void)
/* WRITE 1: NEW -> TRANSFERRING_H2C */ /* WRITE 1: NEW -> TRANSFERRING_H2C */
rqpair.current_recv_depth = 1; rqpair.current_recv_depth = 1;
spdk_nvmf_rdma_request_process(&rtransport, req1); nvmf_rdma_request_process(&rtransport, req1);
CU_ASSERT(req1->state == RDMA_REQUEST_STATE_TRANSFERRING_HOST_TO_CONTROLLER); CU_ASSERT(req1->state == RDMA_REQUEST_STATE_TRANSFERRING_HOST_TO_CONTROLLER);
/* WRITE 1 is the first in batching list */ /* WRITE 1 is the first in batching list */
CU_ASSERT(rqpair.sends_to_post.first == &req1->data.wr); CU_ASSERT(rqpair.sends_to_post.first == &req1->data.wr);
@ -719,7 +719,7 @@ test_spdk_nvmf_rdma_request_process(void)
/* WRITE 2: NEW -> TRANSFERRING_H2C */ /* WRITE 2: NEW -> TRANSFERRING_H2C */
rqpair.current_recv_depth = 2; rqpair.current_recv_depth = 2;
spdk_nvmf_rdma_request_process(&rtransport, req2); nvmf_rdma_request_process(&rtransport, req2);
CU_ASSERT(req2->state == RDMA_REQUEST_STATE_TRANSFERRING_HOST_TO_CONTROLLER); CU_ASSERT(req2->state == RDMA_REQUEST_STATE_TRANSFERRING_HOST_TO_CONTROLLER);
/* WRITE 2 is now also in the batching list */ /* WRITE 2 is now also in the batching list */
CU_ASSERT(rqpair.sends_to_post.first->next == &req2->data.wr); CU_ASSERT(rqpair.sends_to_post.first->next == &req2->data.wr);
@ -732,11 +732,11 @@ test_spdk_nvmf_rdma_request_process(void)
/* WRITE 1 completes before WRITE 2 has finished RDMA reading */ /* WRITE 1 completes before WRITE 2 has finished RDMA reading */
/* WRITE 1: READY_TO_EXECUTE -> EXECUTING */ /* WRITE 1: READY_TO_EXECUTE -> EXECUTING */
req1->state = RDMA_REQUEST_STATE_READY_TO_EXECUTE; req1->state = RDMA_REQUEST_STATE_READY_TO_EXECUTE;
spdk_nvmf_rdma_request_process(&rtransport, req1); nvmf_rdma_request_process(&rtransport, req1);
CU_ASSERT(req1->state == RDMA_REQUEST_STATE_EXECUTING); CU_ASSERT(req1->state == RDMA_REQUEST_STATE_EXECUTING);
/* WRITE 1: EXECUTED -> COMPLETING */ /* WRITE 1: EXECUTED -> COMPLETING */
req1->state = RDMA_REQUEST_STATE_EXECUTED; req1->state = RDMA_REQUEST_STATE_EXECUTED;
spdk_nvmf_rdma_request_process(&rtransport, req1); nvmf_rdma_request_process(&rtransport, req1);
CU_ASSERT(req1->state == RDMA_REQUEST_STATE_COMPLETING); CU_ASSERT(req1->state == RDMA_REQUEST_STATE_COMPLETING);
CU_ASSERT(rqpair.sends_to_post.first == &req1->rsp.wr); CU_ASSERT(rqpair.sends_to_post.first == &req1->rsp.wr);
CU_ASSERT(rqpair.sends_to_post.last == &req1->rsp.wr); CU_ASSERT(rqpair.sends_to_post.last == &req1->rsp.wr);
@ -744,18 +744,18 @@ test_spdk_nvmf_rdma_request_process(void)
STAILQ_INIT(&poller.qpairs_pending_send); STAILQ_INIT(&poller.qpairs_pending_send);
/* WRITE 1: COMPLETED -> FREE */ /* WRITE 1: COMPLETED -> FREE */
req1->state = RDMA_REQUEST_STATE_COMPLETED; req1->state = RDMA_REQUEST_STATE_COMPLETED;
spdk_nvmf_rdma_request_process(&rtransport, req1); nvmf_rdma_request_process(&rtransport, req1);
CU_ASSERT(req1->state == RDMA_REQUEST_STATE_FREE); CU_ASSERT(req1->state == RDMA_REQUEST_STATE_FREE);
/* Now WRITE 2 has finished reading and completes */ /* Now WRITE 2 has finished reading and completes */
/* WRITE 2: COMPLETED -> FREE */ /* WRITE 2: COMPLETED -> FREE */
/* WRITE 2: READY_TO_EXECUTE -> EXECUTING */ /* WRITE 2: READY_TO_EXECUTE -> EXECUTING */
req2->state = RDMA_REQUEST_STATE_READY_TO_EXECUTE; req2->state = RDMA_REQUEST_STATE_READY_TO_EXECUTE;
spdk_nvmf_rdma_request_process(&rtransport, req2); nvmf_rdma_request_process(&rtransport, req2);
CU_ASSERT(req2->state == RDMA_REQUEST_STATE_EXECUTING); CU_ASSERT(req2->state == RDMA_REQUEST_STATE_EXECUTING);
/* WRITE 1: EXECUTED -> COMPLETING */ /* WRITE 1: EXECUTED -> COMPLETING */
req2->state = RDMA_REQUEST_STATE_EXECUTED; req2->state = RDMA_REQUEST_STATE_EXECUTED;
spdk_nvmf_rdma_request_process(&rtransport, req2); nvmf_rdma_request_process(&rtransport, req2);
CU_ASSERT(req2->state == RDMA_REQUEST_STATE_COMPLETING); CU_ASSERT(req2->state == RDMA_REQUEST_STATE_COMPLETING);
CU_ASSERT(rqpair.sends_to_post.first == &req2->rsp.wr); CU_ASSERT(rqpair.sends_to_post.first == &req2->rsp.wr);
CU_ASSERT(rqpair.sends_to_post.last == &req2->rsp.wr); CU_ASSERT(rqpair.sends_to_post.last == &req2->rsp.wr);
@ -763,7 +763,7 @@ test_spdk_nvmf_rdma_request_process(void)
STAILQ_INIT(&poller.qpairs_pending_send); STAILQ_INIT(&poller.qpairs_pending_send);
/* WRITE 1: COMPLETED -> FREE */ /* WRITE 1: COMPLETED -> FREE */
req2->state = RDMA_REQUEST_STATE_COMPLETED; req2->state = RDMA_REQUEST_STATE_COMPLETED;
spdk_nvmf_rdma_request_process(&rtransport, req2); nvmf_rdma_request_process(&rtransport, req2);
CU_ASSERT(req2->state == RDMA_REQUEST_STATE_FREE); CU_ASSERT(req2->state == RDMA_REQUEST_STATE_FREE);
free_recv(recv1); free_recv(recv1);
@ -780,7 +780,7 @@ test_spdk_nvmf_rdma_request_process(void)
#define TEST_GROUPS_COUNT 5 #define TEST_GROUPS_COUNT 5
static void static void
test_spdk_nvmf_rdma_get_optimal_poll_group(void) test_nvmf_rdma_get_optimal_poll_group(void)
{ {
struct spdk_nvmf_rdma_transport rtransport = {}; struct spdk_nvmf_rdma_transport rtransport = {};
struct spdk_nvmf_transport *transport = &rtransport.transport; struct spdk_nvmf_transport *transport = &rtransport.transport;
@ -795,7 +795,7 @@ test_spdk_nvmf_rdma_get_optimal_poll_group(void)
TAILQ_INIT(&rtransport.poll_groups); TAILQ_INIT(&rtransport.poll_groups);
for (i = 0; i < TEST_GROUPS_COUNT; i++) { for (i = 0; i < TEST_GROUPS_COUNT; i++) {
groups[i] = spdk_nvmf_rdma_poll_group_create(transport); groups[i] = nvmf_rdma_poll_group_create(transport);
CU_ASSERT(groups[i] != NULL); CU_ASSERT(groups[i] != NULL);
rgroups[i] = SPDK_CONTAINEROF(groups[i], struct spdk_nvmf_rdma_poll_group, group); rgroups[i] = SPDK_CONTAINEROF(groups[i], struct spdk_nvmf_rdma_poll_group, group);
groups[i]->transport = transport; groups[i]->transport = transport;
@ -806,13 +806,13 @@ test_spdk_nvmf_rdma_get_optimal_poll_group(void)
/* Emulate connection of %TEST_GROUPS_COUNT% initiators - each creates 1 admin and 1 io qp */ /* Emulate connection of %TEST_GROUPS_COUNT% initiators - each creates 1 admin and 1 io qp */
for (i = 0; i < TEST_GROUPS_COUNT; i++) { for (i = 0; i < TEST_GROUPS_COUNT; i++) {
rqpair.qpair.qid = 0; rqpair.qpair.qid = 0;
result = spdk_nvmf_rdma_get_optimal_poll_group(&rqpair.qpair); result = nvmf_rdma_get_optimal_poll_group(&rqpair.qpair);
CU_ASSERT(result == groups[i]); CU_ASSERT(result == groups[i]);
CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[(i + 1) % TEST_GROUPS_COUNT]); CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[(i + 1) % TEST_GROUPS_COUNT]);
CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[i]); CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[i]);
rqpair.qpair.qid = 1; rqpair.qpair.qid = 1;
result = spdk_nvmf_rdma_get_optimal_poll_group(&rqpair.qpair); result = nvmf_rdma_get_optimal_poll_group(&rqpair.qpair);
CU_ASSERT(result == groups[i]); CU_ASSERT(result == groups[i]);
CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[(i + 1) % TEST_GROUPS_COUNT]); CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[(i + 1) % TEST_GROUPS_COUNT]);
CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[(i + 1) % TEST_GROUPS_COUNT]); CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[(i + 1) % TEST_GROUPS_COUNT]);
@ -820,7 +820,7 @@ test_spdk_nvmf_rdma_get_optimal_poll_group(void)
/* wrap around, admin/io pg point to the first pg /* wrap around, admin/io pg point to the first pg
Destroy all poll groups except of the last one */ Destroy all poll groups except of the last one */
for (i = 0; i < TEST_GROUPS_COUNT - 1; i++) { for (i = 0; i < TEST_GROUPS_COUNT - 1; i++) {
spdk_nvmf_rdma_poll_group_destroy(groups[i]); nvmf_rdma_poll_group_destroy(groups[i]);
CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[i + 1]); CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[i + 1]);
CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[i + 1]); CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[i + 1]);
} }
@ -830,29 +830,29 @@ test_spdk_nvmf_rdma_get_optimal_poll_group(void)
/* Check that pointers to the next admin/io poll groups are not changed */ /* Check that pointers to the next admin/io poll groups are not changed */
rqpair.qpair.qid = 0; rqpair.qpair.qid = 0;
result = spdk_nvmf_rdma_get_optimal_poll_group(&rqpair.qpair); result = nvmf_rdma_get_optimal_poll_group(&rqpair.qpair);
CU_ASSERT(result == groups[TEST_GROUPS_COUNT - 1]); CU_ASSERT(result == groups[TEST_GROUPS_COUNT - 1]);
CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[TEST_GROUPS_COUNT - 1]); CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[TEST_GROUPS_COUNT - 1]);
CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[TEST_GROUPS_COUNT - 1]); CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[TEST_GROUPS_COUNT - 1]);
rqpair.qpair.qid = 1; rqpair.qpair.qid = 1;
result = spdk_nvmf_rdma_get_optimal_poll_group(&rqpair.qpair); result = nvmf_rdma_get_optimal_poll_group(&rqpair.qpair);
CU_ASSERT(result == groups[TEST_GROUPS_COUNT - 1]); CU_ASSERT(result == groups[TEST_GROUPS_COUNT - 1]);
CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[TEST_GROUPS_COUNT - 1]); CU_ASSERT(rtransport.conn_sched.next_admin_pg == rgroups[TEST_GROUPS_COUNT - 1]);
CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[TEST_GROUPS_COUNT - 1]); CU_ASSERT(rtransport.conn_sched.next_io_pg == rgroups[TEST_GROUPS_COUNT - 1]);
/* Remove the last poll group, check that pointers are NULL */ /* Remove the last poll group, check that pointers are NULL */
spdk_nvmf_rdma_poll_group_destroy(groups[TEST_GROUPS_COUNT - 1]); nvmf_rdma_poll_group_destroy(groups[TEST_GROUPS_COUNT - 1]);
CU_ASSERT(rtransport.conn_sched.next_admin_pg == NULL); CU_ASSERT(rtransport.conn_sched.next_admin_pg == NULL);
CU_ASSERT(rtransport.conn_sched.next_io_pg == NULL); CU_ASSERT(rtransport.conn_sched.next_io_pg == NULL);
/* Request optimal poll group, result must be NULL */ /* Request optimal poll group, result must be NULL */
rqpair.qpair.qid = 0; rqpair.qpair.qid = 0;
result = spdk_nvmf_rdma_get_optimal_poll_group(&rqpair.qpair); result = nvmf_rdma_get_optimal_poll_group(&rqpair.qpair);
CU_ASSERT(result == NULL); CU_ASSERT(result == NULL);
rqpair.qpair.qid = 1; rqpair.qpair.qid = 1;
result = spdk_nvmf_rdma_get_optimal_poll_group(&rqpair.qpair); result = nvmf_rdma_get_optimal_poll_group(&rqpair.qpair);
CU_ASSERT(result == NULL); CU_ASSERT(result == NULL);
pthread_mutex_destroy(&rtransport.lock); pthread_mutex_destroy(&rtransport.lock);
@ -924,7 +924,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = data_bs * 8; rtransport.transport.opts.io_unit_size = data_bs * 8;
sgl->keyed.length = data_bs * 4; sgl->keyed.length = data_bs * 4;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -954,7 +954,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = data_bs * 4; rtransport.transport.opts.io_unit_size = data_bs * 4;
sgl->keyed.length = data_bs * 4; sgl->keyed.length = data_bs * 4;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -991,7 +991,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = data_bs; rtransport.transport.opts.io_unit_size = data_bs;
sgl->keyed.length = data_bs; sgl->keyed.length = data_bs;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -1026,7 +1026,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = (data_bs + md_size) * 4; rtransport.transport.opts.io_unit_size = (data_bs + md_size) * 4;
sgl->keyed.length = data_bs * 4; sgl->keyed.length = data_bs * 4;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -1056,7 +1056,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = (data_bs + md_size) * 2; rtransport.transport.opts.io_unit_size = (data_bs + md_size) * 2;
sgl->keyed.length = data_bs * 4; sgl->keyed.length = data_bs * 4;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -1089,7 +1089,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = data_bs * 4; rtransport.transport.opts.io_unit_size = data_bs * 4;
sgl->keyed.length = data_bs * 6; sgl->keyed.length = data_bs * 6;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -1137,7 +1137,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = data_bs * 16; rtransport.transport.opts.io_unit_size = data_bs * 16;
sgl->keyed.length = data_bs * 16; sgl->keyed.length = data_bs * 16;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -1164,7 +1164,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
rtransport.transport.opts.io_unit_size = 516; rtransport.transport.opts.io_unit_size = 516;
sgl->keyed.length = data_bs * 2; sgl->keyed.length = data_bs * 2;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -1199,7 +1199,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
g_mr_size = data_bs; g_mr_size = data_bs;
g_mr_next_size = rtransport.transport.opts.io_unit_size; g_mr_next_size = rtransport.transport.opts.io_unit_size;
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
SPDK_CU_ASSERT_FATAL(rc == 0); SPDK_CU_ASSERT_FATAL(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
CU_ASSERT(rdma_req.req.length == rtransport.transport.opts.io_unit_size / 2); CU_ASSERT(rdma_req.req.length == rtransport.transport.opts.io_unit_size / 2);
@ -1248,7 +1248,7 @@ test_spdk_nvmf_rdma_request_parse_sgl_with_md(void)
sgl_desc[i].keyed.key = 0x44; sgl_desc[i].keyed.key = 0x44;
} }
rc = spdk_nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req); rc = nvmf_rdma_request_parse_sgl(&rtransport, &device, &rdma_req);
CU_ASSERT(rc == 0); CU_ASSERT(rc == 0);
CU_ASSERT(rdma_req.req.data_from_pool == true); CU_ASSERT(rdma_req.req.data_from_pool == true);
@ -1289,7 +1289,7 @@ int main(int argc, char **argv)
CU_ADD_TEST(suite, test_spdk_nvmf_rdma_request_parse_sgl); CU_ADD_TEST(suite, test_spdk_nvmf_rdma_request_parse_sgl);
CU_ADD_TEST(suite, test_spdk_nvmf_rdma_request_process); CU_ADD_TEST(suite, test_spdk_nvmf_rdma_request_process);
CU_ADD_TEST(suite, test_spdk_nvmf_rdma_get_optimal_poll_group); CU_ADD_TEST(suite, test_nvmf_rdma_get_optimal_poll_group);
CU_ADD_TEST(suite, test_spdk_nvmf_rdma_request_parse_sgl_with_md); CU_ADD_TEST(suite, test_spdk_nvmf_rdma_request_parse_sgl_with_md);
CU_basic_set_mode(CU_BRM_VERBOSE); CU_basic_set_mode(CU_BRM_VERBOSE);