lib/nvmf: Use the accelerated engine to compute the data digest.

This patch is used to leverage accelerated engine to compute
the data digest in the following case:

1 DIF is not used.
2 The data to compute is aligned with size 4, i.e, %4 = 0.

Signed-off-by: Ziye Yang <ziye.yang@intel.com>
Change-Id: I51fb6e3ab04391062b244cba6e249c8e20d3180f
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/6014
Community-CI: Broadcom CI
Community-CI: Mellanox Build Bot
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Changpeng Liu <changpeng.liu@intel.com>
Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com>
Reviewed-by: Aleksey Marchuk <alexeymar@mellanox.com>
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
This commit is contained in:
Ziye Yang 2021-01-21 19:49:41 +08:00 committed by Tomasz Zawadzki
parent 88754353c0
commit be57e2a6ec
2 changed files with 47 additions and 14 deletions

View File

@ -102,6 +102,7 @@ struct nvme_tcp_pdu {
bool has_hdgst; bool has_hdgst;
bool ddgst_enable; bool ddgst_enable;
uint32_t header_digest_crc32; uint32_t header_digest_crc32;
uint32_t data_digest_crc32;
uint8_t data_digest[SPDK_NVME_TCP_DIGEST_LEN]; uint8_t data_digest[SPDK_NVME_TCP_DIGEST_LEN];
uint8_t ch_valid_bytes; uint8_t ch_valid_bytes;

View File

@ -815,13 +815,6 @@ _tcp_write_pdu(struct nvme_tcp_pdu *pdu)
uint32_t mapped_length = 0; uint32_t mapped_length = 0;
ssize_t rc; ssize_t rc;
struct spdk_nvmf_tcp_qpair *tqpair = pdu->qpair; struct spdk_nvmf_tcp_qpair *tqpair = pdu->qpair;
uint32_t crc32c;
/* Data Digest */
if (pdu->data_len > 0 && g_nvme_tcp_ddgst[pdu->hdr.common.pdu_type] && tqpair->host_ddgst_enable) {
crc32c = nvme_tcp_pdu_calc_data_digest(pdu);
MAKE_DIGEST_WORD(pdu->data_digest, crc32c);
}
pdu->sock_req.iovcnt = nvme_tcp_build_iovs(pdu->iov, SPDK_COUNTOF(pdu->iov), pdu, pdu->sock_req.iovcnt = nvme_tcp_build_iovs(pdu->iov, SPDK_COUNTOF(pdu->iov), pdu,
tqpair->host_hdgst_enable, tqpair->host_ddgst_enable, tqpair->host_hdgst_enable, tqpair->host_ddgst_enable,
@ -843,21 +836,60 @@ _tcp_write_pdu(struct nvme_tcp_pdu *pdu)
} }
static void static void
header_crc32_accel_done(void *cb_arg, int status) data_crc32_accel_done(void *cb_arg, int status)
{ {
struct nvme_tcp_pdu *pdu = cb_arg; struct nvme_tcp_pdu *pdu = cb_arg;
pdu->header_digest_crc32 = pdu->header_digest_crc32 ^ SPDK_CRC32C_XOR;
MAKE_DIGEST_WORD((uint8_t *)pdu->hdr.raw + pdu->hdr.common.hlen, pdu->header_digest_crc32);
if (spdk_unlikely(status)) { if (spdk_unlikely(status)) {
SPDK_ERRLOG("Failed to finish the crc32 work\n"); SPDK_ERRLOG("Failed to compute the data digest for pdu =%p\n", pdu);
_pdu_write_done(pdu, status); _pdu_write_done(pdu, status);
return; return;
} }
pdu->data_digest_crc32 ^= SPDK_CRC32C_XOR;
MAKE_DIGEST_WORD(pdu->data_digest, pdu->data_digest_crc32);
_tcp_write_pdu(pdu); _tcp_write_pdu(pdu);
} }
static void
pdu_data_crc32_compute(struct nvme_tcp_pdu *pdu)
{
struct spdk_nvmf_tcp_qpair *tqpair = pdu->qpair;
uint32_t crc32c;
/* Data Digest */
if (pdu->data_len > 0 && g_nvme_tcp_ddgst[pdu->hdr.common.pdu_type] && tqpair->host_ddgst_enable) {
/* Only suport this limitated case for the first step */
if (spdk_likely(!pdu->dif_ctx && (pdu->data_len % SPDK_NVME_TCP_DIGEST_ALIGNMENT == 0))) {
spdk_accel_submit_crc32cv(tqpair->accel_channel, &pdu->data_digest_crc32,
pdu->data_iov, pdu->data_iovcnt, 0, data_crc32_accel_done, pdu);
return;
}
crc32c = nvme_tcp_pdu_calc_data_digest(pdu);
MAKE_DIGEST_WORD(pdu->data_digest, crc32c);
}
_tcp_write_pdu(pdu);
}
static void
header_crc32_accel_done(void *cb_arg, int status)
{
struct nvme_tcp_pdu *pdu = cb_arg;
pdu->header_digest_crc32 ^= SPDK_CRC32C_XOR;
MAKE_DIGEST_WORD((uint8_t *)pdu->hdr.raw + pdu->hdr.common.hlen, pdu->header_digest_crc32);
if (spdk_unlikely(status)) {
SPDK_ERRLOG("Failed to compute header digest on pdu=%p\n", pdu);
_pdu_write_done(pdu, status);
return;
}
pdu_data_crc32_compute(pdu);
}
static void static void
nvmf_tcp_qpair_write_pdu(struct spdk_nvmf_tcp_qpair *tqpair, nvmf_tcp_qpair_write_pdu(struct spdk_nvmf_tcp_qpair *tqpair,
struct nvme_tcp_pdu *pdu, struct nvme_tcp_pdu *pdu,
@ -883,7 +915,7 @@ nvmf_tcp_qpair_write_pdu(struct spdk_nvmf_tcp_qpair *tqpair,
return; return;
} }
_tcp_write_pdu(pdu); pdu_data_crc32_compute(pdu);
} }
static int static int
@ -1705,12 +1737,12 @@ nvmf_tcp_icreq_handle(struct spdk_nvmf_tcp_transport *ttransport,
/* Not fatal. */ /* Not fatal. */
} }
if (tqpair->host_hdgst_enable) { if (tqpair->host_hdgst_enable || tqpair->host_ddgst_enable) {
tqpair->accel_channel = spdk_accel_engine_get_io_channel(); tqpair->accel_channel = spdk_accel_engine_get_io_channel();
if (spdk_unlikely(!tqpair->accel_channel)) { if (spdk_unlikely(!tqpair->accel_channel)) {
fes = SPDK_NVME_TCP_TERM_REQ_FES_HDGST_ERROR; fes = SPDK_NVME_TCP_TERM_REQ_FES_HDGST_ERROR;
error_offset = offsetof(struct spdk_nvme_tcp_ic_req, dgst); error_offset = offsetof(struct spdk_nvme_tcp_ic_req, dgst);
SPDK_ERRLOG("Unabled to get accel_channel for tqpair=%p, failed to enable header digest\n", SPDK_ERRLOG("Unabled to get accel_channel for tqpair=%p, failed to enable digest for header or data\n",
tqpair); tqpair);
goto end; goto end;
} }