The rdma buffer for stripping DIF metadata is added. CPU strips the DIF metadata and copies it to the rdma buffer, improving the rdma write bandwith. The network bandwidth during 4KB random read test is increased from 79 Gbps to 99 Gbps, the IOPS is increased from 2075K to 2637K. Fixes issue #2418 Signed-off-by: Chunsong Feng <fengchunsong@huawei.com> Change-Id: If1c31256f0390f31d396812fa33cd650bf52b336 Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/11861 Reviewed-by: Shuhei Matsumoto <smatsumoto@nvidia.com> Reviewed-by: Jim Harris <james.r.harris@intel.com> Reviewed-by: Aleksey Marchuk <alexeymar@mellanox.com> Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
850 lines
24 KiB
C
850 lines
24 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright (c) Intel Corporation. All rights reserved.
|
|
* Copyright (c) 2018-2019, 2021 Mellanox Technologies LTD. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "spdk/stdinc.h"
|
|
|
|
#include "nvmf_internal.h"
|
|
#include "transport.h"
|
|
|
|
#include "spdk/config.h"
|
|
#include "spdk/log.h"
|
|
#include "spdk/nvmf.h"
|
|
#include "spdk/nvmf_transport.h"
|
|
#include "spdk/queue.h"
|
|
#include "spdk/util.h"
|
|
#include "spdk_internal/usdt.h"
|
|
|
|
#define MAX_MEMPOOL_NAME_LENGTH 40
|
|
#define NVMF_TRANSPORT_DEFAULT_ASSOCIATION_TIMEOUT_IN_MS 120000
|
|
|
|
struct nvmf_transport_ops_list_element {
|
|
struct spdk_nvmf_transport_ops ops;
|
|
TAILQ_ENTRY(nvmf_transport_ops_list_element) link;
|
|
};
|
|
|
|
TAILQ_HEAD(nvmf_transport_ops_list, nvmf_transport_ops_list_element)
|
|
g_spdk_nvmf_transport_ops = TAILQ_HEAD_INITIALIZER(g_spdk_nvmf_transport_ops);
|
|
|
|
static inline const struct spdk_nvmf_transport_ops *
|
|
nvmf_get_transport_ops(const char *transport_name)
|
|
{
|
|
struct nvmf_transport_ops_list_element *ops;
|
|
TAILQ_FOREACH(ops, &g_spdk_nvmf_transport_ops, link) {
|
|
if (strcasecmp(transport_name, ops->ops.name) == 0) {
|
|
return &ops->ops;
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
void
|
|
spdk_nvmf_transport_register(const struct spdk_nvmf_transport_ops *ops)
|
|
{
|
|
struct nvmf_transport_ops_list_element *new_ops;
|
|
|
|
if (nvmf_get_transport_ops(ops->name) != NULL) {
|
|
SPDK_ERRLOG("Double registering nvmf transport type %s.\n", ops->name);
|
|
assert(false);
|
|
return;
|
|
}
|
|
|
|
new_ops = calloc(1, sizeof(*new_ops));
|
|
if (new_ops == NULL) {
|
|
SPDK_ERRLOG("Unable to allocate memory to register new transport type %s.\n", ops->name);
|
|
assert(false);
|
|
return;
|
|
}
|
|
|
|
new_ops->ops = *ops;
|
|
|
|
TAILQ_INSERT_TAIL(&g_spdk_nvmf_transport_ops, new_ops, link);
|
|
}
|
|
|
|
const struct spdk_nvmf_transport_opts *
|
|
spdk_nvmf_get_transport_opts(struct spdk_nvmf_transport *transport)
|
|
{
|
|
return &transport->opts;
|
|
}
|
|
|
|
void
|
|
nvmf_transport_dump_opts(struct spdk_nvmf_transport *transport, struct spdk_json_write_ctx *w,
|
|
bool named)
|
|
{
|
|
const struct spdk_nvmf_transport_opts *opts = spdk_nvmf_get_transport_opts(transport);
|
|
|
|
named ? spdk_json_write_named_object_begin(w, "params") : spdk_json_write_object_begin(w);
|
|
|
|
spdk_json_write_named_string(w, "trtype", spdk_nvmf_get_transport_name(transport));
|
|
spdk_json_write_named_uint32(w, "max_queue_depth", opts->max_queue_depth);
|
|
spdk_json_write_named_uint32(w, "max_io_qpairs_per_ctrlr", opts->max_qpairs_per_ctrlr - 1);
|
|
spdk_json_write_named_uint32(w, "in_capsule_data_size", opts->in_capsule_data_size);
|
|
spdk_json_write_named_uint32(w, "max_io_size", opts->max_io_size);
|
|
spdk_json_write_named_uint32(w, "io_unit_size", opts->io_unit_size);
|
|
spdk_json_write_named_uint32(w, "max_aq_depth", opts->max_aq_depth);
|
|
spdk_json_write_named_uint32(w, "num_shared_buffers", opts->num_shared_buffers);
|
|
spdk_json_write_named_uint32(w, "buf_cache_size", opts->buf_cache_size);
|
|
spdk_json_write_named_bool(w, "dif_insert_or_strip", opts->dif_insert_or_strip);
|
|
spdk_json_write_named_bool(w, "zcopy", opts->zcopy);
|
|
|
|
if (transport->ops->dump_opts) {
|
|
transport->ops->dump_opts(transport, w);
|
|
}
|
|
|
|
spdk_json_write_named_uint32(w, "abort_timeout_sec", opts->abort_timeout_sec);
|
|
spdk_json_write_object_end(w);
|
|
}
|
|
|
|
void
|
|
nvmf_transport_listen_dump_opts(struct spdk_nvmf_transport *transport,
|
|
const struct spdk_nvme_transport_id *trid, struct spdk_json_write_ctx *w)
|
|
{
|
|
const char *adrfam = spdk_nvme_transport_id_adrfam_str(trid->adrfam);
|
|
|
|
spdk_json_write_named_object_begin(w, "listen_address");
|
|
|
|
spdk_json_write_named_string(w, "trtype", trid->trstring);
|
|
spdk_json_write_named_string(w, "adrfam", adrfam ? adrfam : "unknown");
|
|
spdk_json_write_named_string(w, "traddr", trid->traddr);
|
|
spdk_json_write_named_string(w, "trsvcid", trid->trsvcid);
|
|
|
|
if (transport->ops->listen_dump_opts) {
|
|
transport->ops->listen_dump_opts(transport, trid, w);
|
|
}
|
|
|
|
spdk_json_write_object_end(w);
|
|
}
|
|
|
|
spdk_nvme_transport_type_t
|
|
spdk_nvmf_get_transport_type(struct spdk_nvmf_transport *transport)
|
|
{
|
|
return transport->ops->type;
|
|
}
|
|
|
|
const char *
|
|
spdk_nvmf_get_transport_name(struct spdk_nvmf_transport *transport)
|
|
{
|
|
return transport->ops->name;
|
|
}
|
|
|
|
static void nvmf_transport_opts_copy(struct spdk_nvmf_transport_opts *opts,
|
|
struct spdk_nvmf_transport_opts *opts_src,
|
|
size_t opts_size)
|
|
{
|
|
assert(opts);
|
|
assert(opts_src);
|
|
|
|
opts->opts_size = opts_size;
|
|
|
|
#define SET_FIELD(field) \
|
|
if (offsetof(struct spdk_nvmf_transport_opts, field) + sizeof(opts->field) <= opts_size) { \
|
|
opts->field = opts_src->field; \
|
|
} \
|
|
|
|
SET_FIELD(max_queue_depth);
|
|
SET_FIELD(max_qpairs_per_ctrlr);
|
|
SET_FIELD(in_capsule_data_size);
|
|
SET_FIELD(max_io_size);
|
|
SET_FIELD(io_unit_size);
|
|
SET_FIELD(max_aq_depth);
|
|
SET_FIELD(buf_cache_size);
|
|
SET_FIELD(num_shared_buffers);
|
|
SET_FIELD(dif_insert_or_strip);
|
|
SET_FIELD(abort_timeout_sec);
|
|
SET_FIELD(association_timeout);
|
|
SET_FIELD(transport_specific);
|
|
SET_FIELD(acceptor_poll_rate);
|
|
SET_FIELD(zcopy);
|
|
|
|
/* Do not remove this statement, you should always update this statement when you adding a new field,
|
|
* and do not forget to add the SET_FIELD statement for your added field. */
|
|
SPDK_STATIC_ASSERT(sizeof(struct spdk_nvmf_transport_opts) == 64, "Incorrect size");
|
|
|
|
#undef SET_FIELD
|
|
#undef FILED_CHECK
|
|
}
|
|
|
|
struct spdk_nvmf_transport *
|
|
spdk_nvmf_transport_create(const char *transport_name, struct spdk_nvmf_transport_opts *opts)
|
|
{
|
|
const struct spdk_nvmf_transport_ops *ops = NULL;
|
|
struct spdk_nvmf_transport *transport;
|
|
char spdk_mempool_name[MAX_MEMPOOL_NAME_LENGTH];
|
|
int chars_written;
|
|
struct spdk_nvmf_transport_opts opts_local = {};
|
|
|
|
if (!opts) {
|
|
SPDK_ERRLOG("opts should not be NULL\n");
|
|
return NULL;
|
|
}
|
|
|
|
if (!opts->opts_size) {
|
|
SPDK_ERRLOG("The opts_size in opts structure should not be zero\n");
|
|
return NULL;
|
|
}
|
|
|
|
ops = nvmf_get_transport_ops(transport_name);
|
|
if (!ops) {
|
|
SPDK_ERRLOG("Transport type '%s' unavailable.\n", transport_name);
|
|
return NULL;
|
|
}
|
|
nvmf_transport_opts_copy(&opts_local, opts, opts->opts_size);
|
|
|
|
if (opts_local.max_io_size != 0 && (!spdk_u32_is_pow2(opts_local.max_io_size) ||
|
|
opts_local.max_io_size < 8192)) {
|
|
SPDK_ERRLOG("max_io_size %u must be a power of 2 and be greater than or equal 8KB\n",
|
|
opts_local.max_io_size);
|
|
return NULL;
|
|
}
|
|
|
|
if (opts_local.max_aq_depth < SPDK_NVMF_MIN_ADMIN_MAX_SQ_SIZE) {
|
|
SPDK_ERRLOG("max_aq_depth %u is less than minimum defined by NVMf spec, use min value\n",
|
|
opts_local.max_aq_depth);
|
|
opts_local.max_aq_depth = SPDK_NVMF_MIN_ADMIN_MAX_SQ_SIZE;
|
|
}
|
|
|
|
transport = ops->create(&opts_local);
|
|
if (!transport) {
|
|
SPDK_ERRLOG("Unable to create new transport of type %s\n", transport_name);
|
|
return NULL;
|
|
}
|
|
|
|
TAILQ_INIT(&transport->listeners);
|
|
|
|
transport->ops = ops;
|
|
transport->opts = opts_local;
|
|
|
|
chars_written = snprintf(spdk_mempool_name, MAX_MEMPOOL_NAME_LENGTH, "%s_%s_%s", "spdk_nvmf",
|
|
transport_name, "data");
|
|
if (chars_written < 0) {
|
|
SPDK_ERRLOG("Unable to generate transport data buffer pool name.\n");
|
|
ops->destroy(transport, NULL, NULL);
|
|
return NULL;
|
|
}
|
|
|
|
if (opts_local.num_shared_buffers) {
|
|
transport->data_buf_pool = spdk_mempool_create(spdk_mempool_name,
|
|
opts_local.num_shared_buffers,
|
|
opts_local.io_unit_size + NVMF_DATA_BUFFER_ALIGNMENT,
|
|
SPDK_MEMPOOL_DEFAULT_CACHE_SIZE,
|
|
SPDK_ENV_SOCKET_ID_ANY);
|
|
|
|
if (!transport->data_buf_pool) {
|
|
SPDK_ERRLOG("Unable to allocate buffer pool for poll group\n");
|
|
ops->destroy(transport, NULL, NULL);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
return transport;
|
|
}
|
|
|
|
struct spdk_nvmf_transport *
|
|
spdk_nvmf_transport_get_first(struct spdk_nvmf_tgt *tgt)
|
|
{
|
|
return TAILQ_FIRST(&tgt->transports);
|
|
}
|
|
|
|
struct spdk_nvmf_transport *
|
|
spdk_nvmf_transport_get_next(struct spdk_nvmf_transport *transport)
|
|
{
|
|
return TAILQ_NEXT(transport, link);
|
|
}
|
|
|
|
int
|
|
spdk_nvmf_transport_destroy(struct spdk_nvmf_transport *transport,
|
|
spdk_nvmf_transport_destroy_done_cb cb_fn, void *cb_arg)
|
|
{
|
|
struct spdk_nvmf_listener *listener, *listener_tmp;
|
|
|
|
if (transport->data_buf_pool != NULL) {
|
|
if (spdk_mempool_count(transport->data_buf_pool) !=
|
|
transport->opts.num_shared_buffers) {
|
|
SPDK_ERRLOG("transport buffer pool count is %zu but should be %u\n",
|
|
spdk_mempool_count(transport->data_buf_pool),
|
|
transport->opts.num_shared_buffers);
|
|
}
|
|
spdk_mempool_free(transport->data_buf_pool);
|
|
}
|
|
|
|
TAILQ_FOREACH_SAFE(listener, &transport->listeners, link, listener_tmp) {
|
|
TAILQ_REMOVE(&transport->listeners, listener, link);
|
|
transport->ops->stop_listen(transport, &listener->trid);
|
|
free(listener);
|
|
}
|
|
|
|
return transport->ops->destroy(transport, cb_fn, cb_arg);
|
|
}
|
|
|
|
struct spdk_nvmf_listener *
|
|
nvmf_transport_find_listener(struct spdk_nvmf_transport *transport,
|
|
const struct spdk_nvme_transport_id *trid)
|
|
{
|
|
struct spdk_nvmf_listener *listener;
|
|
|
|
TAILQ_FOREACH(listener, &transport->listeners, link) {
|
|
if (spdk_nvme_transport_id_compare(&listener->trid, trid) == 0) {
|
|
return listener;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
int
|
|
spdk_nvmf_transport_listen(struct spdk_nvmf_transport *transport,
|
|
const struct spdk_nvme_transport_id *trid, struct spdk_nvmf_listen_opts *opts)
|
|
{
|
|
struct spdk_nvmf_listener *listener;
|
|
int rc;
|
|
|
|
listener = nvmf_transport_find_listener(transport, trid);
|
|
if (!listener) {
|
|
listener = calloc(1, sizeof(*listener));
|
|
if (!listener) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
listener->ref = 1;
|
|
listener->trid = *trid;
|
|
TAILQ_INSERT_TAIL(&transport->listeners, listener, link);
|
|
rc = transport->ops->listen(transport, &listener->trid, opts);
|
|
if (rc != 0) {
|
|
TAILQ_REMOVE(&transport->listeners, listener, link);
|
|
free(listener);
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
++listener->ref;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
spdk_nvmf_transport_stop_listen(struct spdk_nvmf_transport *transport,
|
|
const struct spdk_nvme_transport_id *trid)
|
|
{
|
|
struct spdk_nvmf_listener *listener;
|
|
|
|
listener = nvmf_transport_find_listener(transport, trid);
|
|
if (!listener) {
|
|
return -ENOENT;
|
|
}
|
|
|
|
if (--listener->ref == 0) {
|
|
TAILQ_REMOVE(&transport->listeners, listener, link);
|
|
transport->ops->stop_listen(transport, trid);
|
|
free(listener);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
struct nvmf_stop_listen_ctx {
|
|
struct spdk_nvmf_transport *transport;
|
|
struct spdk_nvme_transport_id trid;
|
|
struct spdk_nvmf_subsystem *subsystem;
|
|
spdk_nvmf_tgt_subsystem_listen_done_fn cb_fn;
|
|
void *cb_arg;
|
|
};
|
|
|
|
static void
|
|
nvmf_stop_listen_fini(struct spdk_io_channel_iter *i, int status)
|
|
{
|
|
struct nvmf_stop_listen_ctx *ctx;
|
|
struct spdk_nvmf_transport *transport;
|
|
int rc = status;
|
|
|
|
ctx = spdk_io_channel_iter_get_ctx(i);
|
|
transport = ctx->transport;
|
|
assert(transport != NULL);
|
|
|
|
rc = spdk_nvmf_transport_stop_listen(transport, &ctx->trid);
|
|
if (rc) {
|
|
SPDK_ERRLOG("Failed to stop listening on address '%s'\n", ctx->trid.traddr);
|
|
}
|
|
|
|
if (ctx->cb_fn) {
|
|
ctx->cb_fn(ctx->cb_arg, rc);
|
|
}
|
|
free(ctx);
|
|
}
|
|
|
|
static void
|
|
nvmf_stop_listen_disconnect_qpairs(struct spdk_io_channel_iter *i)
|
|
{
|
|
struct nvmf_stop_listen_ctx *ctx;
|
|
struct spdk_nvmf_poll_group *group;
|
|
struct spdk_io_channel *ch;
|
|
struct spdk_nvmf_qpair *qpair, *tmp_qpair;
|
|
struct spdk_nvme_transport_id tmp_trid;
|
|
|
|
ctx = spdk_io_channel_iter_get_ctx(i);
|
|
ch = spdk_io_channel_iter_get_channel(i);
|
|
group = spdk_io_channel_get_ctx(ch);
|
|
|
|
TAILQ_FOREACH_SAFE(qpair, &group->qpairs, link, tmp_qpair) {
|
|
/* skip qpairs that don't match the TRID. */
|
|
if (spdk_nvmf_qpair_get_listen_trid(qpair, &tmp_trid)) {
|
|
continue;
|
|
}
|
|
|
|
if (!spdk_nvme_transport_id_compare(&ctx->trid, &tmp_trid)) {
|
|
if (ctx->subsystem == NULL || qpair->ctrlr == NULL ||
|
|
ctx->subsystem == qpair->ctrlr->subsys) {
|
|
spdk_nvmf_qpair_disconnect(qpair, NULL, NULL);
|
|
}
|
|
}
|
|
}
|
|
spdk_for_each_channel_continue(i, 0);
|
|
}
|
|
|
|
int
|
|
spdk_nvmf_transport_stop_listen_async(struct spdk_nvmf_transport *transport,
|
|
const struct spdk_nvme_transport_id *trid,
|
|
struct spdk_nvmf_subsystem *subsystem,
|
|
spdk_nvmf_tgt_subsystem_listen_done_fn cb_fn,
|
|
void *cb_arg)
|
|
{
|
|
struct nvmf_stop_listen_ctx *ctx;
|
|
|
|
ctx = calloc(1, sizeof(struct nvmf_stop_listen_ctx));
|
|
if (ctx == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
ctx->trid = *trid;
|
|
ctx->subsystem = subsystem;
|
|
ctx->transport = transport;
|
|
ctx->cb_fn = cb_fn;
|
|
ctx->cb_arg = cb_arg;
|
|
|
|
spdk_for_each_channel(transport->tgt, nvmf_stop_listen_disconnect_qpairs, ctx,
|
|
nvmf_stop_listen_fini);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
nvmf_transport_listener_discover(struct spdk_nvmf_transport *transport,
|
|
struct spdk_nvme_transport_id *trid,
|
|
struct spdk_nvmf_discovery_log_page_entry *entry)
|
|
{
|
|
transport->ops->listener_discover(transport, trid, entry);
|
|
}
|
|
|
|
struct spdk_nvmf_transport_poll_group *
|
|
nvmf_transport_poll_group_create(struct spdk_nvmf_transport *transport,
|
|
struct spdk_nvmf_poll_group *group)
|
|
{
|
|
struct spdk_nvmf_transport_poll_group *tgroup;
|
|
struct spdk_nvmf_transport_pg_cache_buf **bufs;
|
|
uint32_t i;
|
|
|
|
tgroup = transport->ops->poll_group_create(transport, group);
|
|
if (!tgroup) {
|
|
return NULL;
|
|
}
|
|
tgroup->transport = transport;
|
|
|
|
STAILQ_INIT(&tgroup->pending_buf_queue);
|
|
STAILQ_INIT(&tgroup->buf_cache);
|
|
|
|
if (transport->opts.buf_cache_size) {
|
|
tgroup->buf_cache_size = transport->opts.buf_cache_size;
|
|
bufs = calloc(tgroup->buf_cache_size, sizeof(struct spdk_nvmf_transport_pg_cache_buf *));
|
|
|
|
if (!bufs) {
|
|
SPDK_ERRLOG("Memory allocation failed, can't reserve buffers for the pg buffer cache\n");
|
|
return tgroup;
|
|
}
|
|
|
|
if (spdk_mempool_get_bulk(transport->data_buf_pool, (void **)bufs, tgroup->buf_cache_size)) {
|
|
tgroup->buf_cache_size = (uint32_t)spdk_mempool_count(transport->data_buf_pool);
|
|
SPDK_NOTICELOG("Unable to reserve the full number of buffers for the pg buffer cache. "
|
|
"Decrease the number of cached buffers from %u to %u\n",
|
|
transport->opts.buf_cache_size, tgroup->buf_cache_size);
|
|
/* Sanity check */
|
|
assert(tgroup->buf_cache_size <= transport->opts.buf_cache_size);
|
|
/* Try again with less number of buffers */
|
|
if (spdk_mempool_get_bulk(transport->data_buf_pool, (void **)bufs, tgroup->buf_cache_size)) {
|
|
SPDK_NOTICELOG("Failed to reserve %u buffers\n", tgroup->buf_cache_size);
|
|
tgroup->buf_cache_size = 0;
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < tgroup->buf_cache_size; i++) {
|
|
STAILQ_INSERT_HEAD(&tgroup->buf_cache, bufs[i], link);
|
|
}
|
|
tgroup->buf_cache_count = tgroup->buf_cache_size;
|
|
|
|
free(bufs);
|
|
}
|
|
|
|
return tgroup;
|
|
}
|
|
|
|
struct spdk_nvmf_transport_poll_group *
|
|
nvmf_transport_get_optimal_poll_group(struct spdk_nvmf_transport *transport,
|
|
struct spdk_nvmf_qpair *qpair)
|
|
{
|
|
if (transport->ops->get_optimal_poll_group) {
|
|
return transport->ops->get_optimal_poll_group(qpair);
|
|
} else {
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
void
|
|
nvmf_transport_poll_group_destroy(struct spdk_nvmf_transport_poll_group *group)
|
|
{
|
|
struct spdk_nvmf_transport_pg_cache_buf *buf, *tmp;
|
|
|
|
if (!STAILQ_EMPTY(&group->pending_buf_queue)) {
|
|
SPDK_ERRLOG("Pending I/O list wasn't empty on poll group destruction\n");
|
|
}
|
|
|
|
STAILQ_FOREACH_SAFE(buf, &group->buf_cache, link, tmp) {
|
|
STAILQ_REMOVE(&group->buf_cache, buf, spdk_nvmf_transport_pg_cache_buf, link);
|
|
spdk_mempool_put(group->transport->data_buf_pool, buf);
|
|
}
|
|
group->transport->ops->poll_group_destroy(group);
|
|
}
|
|
|
|
int
|
|
nvmf_transport_poll_group_add(struct spdk_nvmf_transport_poll_group *group,
|
|
struct spdk_nvmf_qpair *qpair)
|
|
{
|
|
if (qpair->transport) {
|
|
assert(qpair->transport == group->transport);
|
|
if (qpair->transport != group->transport) {
|
|
return -1;
|
|
}
|
|
} else {
|
|
qpair->transport = group->transport;
|
|
}
|
|
|
|
SPDK_DTRACE_PROBE3(nvmf_transport_poll_group_add, qpair, qpair->qid,
|
|
spdk_thread_get_id(group->group->thread));
|
|
|
|
return group->transport->ops->poll_group_add(group, qpair);
|
|
}
|
|
|
|
int
|
|
nvmf_transport_poll_group_remove(struct spdk_nvmf_transport_poll_group *group,
|
|
struct spdk_nvmf_qpair *qpair)
|
|
{
|
|
int rc = ENOTSUP;
|
|
|
|
SPDK_DTRACE_PROBE3(nvmf_transport_poll_group_remove, qpair, qpair->qid,
|
|
spdk_thread_get_id(group->group->thread));
|
|
|
|
assert(qpair->transport == group->transport);
|
|
if (group->transport->ops->poll_group_remove) {
|
|
rc = group->transport->ops->poll_group_remove(group, qpair);
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
int
|
|
nvmf_transport_poll_group_poll(struct spdk_nvmf_transport_poll_group *group)
|
|
{
|
|
return group->transport->ops->poll_group_poll(group);
|
|
}
|
|
|
|
int
|
|
nvmf_transport_req_free(struct spdk_nvmf_request *req)
|
|
{
|
|
return req->qpair->transport->ops->req_free(req);
|
|
}
|
|
|
|
int
|
|
nvmf_transport_req_complete(struct spdk_nvmf_request *req)
|
|
{
|
|
return req->qpair->transport->ops->req_complete(req);
|
|
}
|
|
|
|
void
|
|
nvmf_transport_qpair_fini(struct spdk_nvmf_qpair *qpair,
|
|
spdk_nvmf_transport_qpair_fini_cb cb_fn,
|
|
void *cb_arg)
|
|
{
|
|
SPDK_DTRACE_PROBE1(nvmf_transport_qpair_fini, qpair);
|
|
|
|
qpair->transport->ops->qpair_fini(qpair, cb_fn, cb_arg);
|
|
}
|
|
|
|
int
|
|
nvmf_transport_qpair_get_peer_trid(struct spdk_nvmf_qpair *qpair,
|
|
struct spdk_nvme_transport_id *trid)
|
|
{
|
|
return qpair->transport->ops->qpair_get_peer_trid(qpair, trid);
|
|
}
|
|
|
|
int
|
|
nvmf_transport_qpair_get_local_trid(struct spdk_nvmf_qpair *qpair,
|
|
struct spdk_nvme_transport_id *trid)
|
|
{
|
|
return qpair->transport->ops->qpair_get_local_trid(qpair, trid);
|
|
}
|
|
|
|
int
|
|
nvmf_transport_qpair_get_listen_trid(struct spdk_nvmf_qpair *qpair,
|
|
struct spdk_nvme_transport_id *trid)
|
|
{
|
|
return qpair->transport->ops->qpair_get_listen_trid(qpair, trid);
|
|
}
|
|
|
|
void
|
|
nvmf_transport_qpair_abort_request(struct spdk_nvmf_qpair *qpair,
|
|
struct spdk_nvmf_request *req)
|
|
{
|
|
if (qpair->transport->ops->qpair_abort_request) {
|
|
qpair->transport->ops->qpair_abort_request(qpair, req);
|
|
}
|
|
}
|
|
|
|
bool
|
|
spdk_nvmf_transport_opts_init(const char *transport_name,
|
|
struct spdk_nvmf_transport_opts *opts, size_t opts_size)
|
|
{
|
|
const struct spdk_nvmf_transport_ops *ops;
|
|
struct spdk_nvmf_transport_opts opts_local = {};
|
|
|
|
ops = nvmf_get_transport_ops(transport_name);
|
|
if (!ops) {
|
|
SPDK_ERRLOG("Transport type %s unavailable.\n", transport_name);
|
|
return false;
|
|
}
|
|
|
|
if (!opts) {
|
|
SPDK_ERRLOG("opts should not be NULL\n");
|
|
return false;
|
|
}
|
|
|
|
if (!opts_size) {
|
|
SPDK_ERRLOG("opts_size inside opts should not be zero value\n");
|
|
return false;
|
|
}
|
|
|
|
opts_local.association_timeout = NVMF_TRANSPORT_DEFAULT_ASSOCIATION_TIMEOUT_IN_MS;
|
|
opts_local.acceptor_poll_rate = SPDK_NVMF_DEFAULT_ACCEPT_POLL_RATE_US;
|
|
ops->opts_init(&opts_local);
|
|
|
|
nvmf_transport_opts_copy(opts, &opts_local, opts_size);
|
|
|
|
return true;
|
|
}
|
|
|
|
void
|
|
spdk_nvmf_request_free_buffers(struct spdk_nvmf_request *req,
|
|
struct spdk_nvmf_transport_poll_group *group,
|
|
struct spdk_nvmf_transport *transport)
|
|
{
|
|
uint32_t i;
|
|
|
|
for (i = 0; i < req->iovcnt; i++) {
|
|
if (group->buf_cache_count < group->buf_cache_size) {
|
|
STAILQ_INSERT_HEAD(&group->buf_cache,
|
|
(struct spdk_nvmf_transport_pg_cache_buf *)req->buffers[i],
|
|
link);
|
|
group->buf_cache_count++;
|
|
} else {
|
|
spdk_mempool_put(transport->data_buf_pool, req->buffers[i]);
|
|
}
|
|
req->iov[i].iov_base = NULL;
|
|
req->buffers[i] = NULL;
|
|
req->iov[i].iov_len = 0;
|
|
}
|
|
req->data_from_pool = false;
|
|
}
|
|
|
|
typedef int (*set_buffer_callback)(struct spdk_nvmf_request *req, void *buf,
|
|
uint32_t length, uint32_t io_unit_size);
|
|
static int
|
|
nvmf_request_set_buffer(struct spdk_nvmf_request *req, void *buf, uint32_t length,
|
|
uint32_t io_unit_size)
|
|
{
|
|
req->buffers[req->iovcnt] = buf;
|
|
req->iov[req->iovcnt].iov_base = (void *)((uintptr_t)(buf + NVMF_DATA_BUFFER_MASK) &
|
|
~NVMF_DATA_BUFFER_MASK);
|
|
req->iov[req->iovcnt].iov_len = spdk_min(length, io_unit_size);
|
|
length -= req->iov[req->iovcnt].iov_len;
|
|
req->iovcnt++;
|
|
|
|
return length;
|
|
}
|
|
|
|
static int
|
|
nvmf_request_get_buffers(struct spdk_nvmf_request *req,
|
|
struct spdk_nvmf_transport_poll_group *group,
|
|
struct spdk_nvmf_transport *transport,
|
|
uint32_t length, uint32_t io_unit_size,
|
|
set_buffer_callback cb_func)
|
|
{
|
|
uint32_t num_buffers;
|
|
uint32_t i = 0, j;
|
|
void *buffer, *buffers[NVMF_REQ_MAX_BUFFERS];
|
|
|
|
/* If the number of buffers is too large, then we know the I/O is larger than allowed.
|
|
* Fail it.
|
|
*/
|
|
num_buffers = SPDK_CEIL_DIV(length, io_unit_size);
|
|
if (num_buffers > NVMF_REQ_MAX_BUFFERS) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
while (i < num_buffers) {
|
|
if (!(STAILQ_EMPTY(&group->buf_cache))) {
|
|
group->buf_cache_count--;
|
|
buffer = STAILQ_FIRST(&group->buf_cache);
|
|
STAILQ_REMOVE_HEAD(&group->buf_cache, link);
|
|
assert(buffer != NULL);
|
|
|
|
length = cb_func(req, buffer, length, io_unit_size);
|
|
i++;
|
|
} else {
|
|
if (spdk_mempool_get_bulk(transport->data_buf_pool, buffers,
|
|
num_buffers - i)) {
|
|
return -ENOMEM;
|
|
}
|
|
for (j = 0; j < num_buffers - i; j++) {
|
|
length = cb_func(req, buffers[j], length, io_unit_size);
|
|
}
|
|
i += num_buffers - i;
|
|
}
|
|
}
|
|
|
|
assert(length == 0);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
spdk_nvmf_request_get_buffers(struct spdk_nvmf_request *req,
|
|
struct spdk_nvmf_transport_poll_group *group,
|
|
struct spdk_nvmf_transport *transport,
|
|
uint32_t length)
|
|
{
|
|
int rc;
|
|
|
|
req->iovcnt = 0;
|
|
rc = nvmf_request_get_buffers(req, group, transport, length,
|
|
transport->opts.io_unit_size,
|
|
nvmf_request_set_buffer);
|
|
if (!rc) {
|
|
req->data_from_pool = true;
|
|
} else if (rc == -ENOMEM) {
|
|
spdk_nvmf_request_free_buffers(req, group, transport);
|
|
return rc;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int
|
|
nvmf_request_set_stripped_buffer(struct spdk_nvmf_request *req, void *buf, uint32_t length,
|
|
uint32_t io_unit_size)
|
|
{
|
|
struct spdk_nvmf_stripped_data *data = req->stripped_data;
|
|
|
|
data->buffers[data->iovcnt] = buf;
|
|
data->iov[data->iovcnt].iov_base = (void *)((uintptr_t)(buf + NVMF_DATA_BUFFER_MASK) &
|
|
~NVMF_DATA_BUFFER_MASK);
|
|
data->iov[data->iovcnt].iov_len = spdk_min(length, io_unit_size);
|
|
length -= data->iov[data->iovcnt].iov_len;
|
|
data->iovcnt++;
|
|
|
|
return length;
|
|
}
|
|
|
|
void
|
|
nvmf_request_free_stripped_buffers(struct spdk_nvmf_request *req,
|
|
struct spdk_nvmf_transport_poll_group *group,
|
|
struct spdk_nvmf_transport *transport)
|
|
{
|
|
struct spdk_nvmf_stripped_data *data = req->stripped_data;
|
|
uint32_t i;
|
|
|
|
for (i = 0; i < data->iovcnt; i++) {
|
|
if (group->buf_cache_count < group->buf_cache_size) {
|
|
STAILQ_INSERT_HEAD(&group->buf_cache,
|
|
(struct spdk_nvmf_transport_pg_cache_buf *)data->buffers[i],
|
|
link);
|
|
group->buf_cache_count++;
|
|
} else {
|
|
spdk_mempool_put(transport->data_buf_pool, data->buffers[i]);
|
|
}
|
|
}
|
|
free(data);
|
|
req->stripped_data = NULL;
|
|
}
|
|
|
|
int
|
|
nvmf_request_get_stripped_buffers(struct spdk_nvmf_request *req,
|
|
struct spdk_nvmf_transport_poll_group *group,
|
|
struct spdk_nvmf_transport *transport,
|
|
uint32_t length)
|
|
{
|
|
uint32_t block_size = req->dif.dif_ctx.block_size;
|
|
uint32_t data_block_size = block_size - req->dif.dif_ctx.md_size;
|
|
uint32_t io_unit_size = transport->opts.io_unit_size / block_size * data_block_size;
|
|
struct spdk_nvmf_stripped_data *data;
|
|
uint32_t i;
|
|
int rc;
|
|
|
|
/* Data blocks must be block aligned */
|
|
for (i = 0; i < req->iovcnt; i++) {
|
|
if (req->iov[i].iov_len % block_size) {
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
data = calloc(1, sizeof(*data));
|
|
if (data == NULL) {
|
|
SPDK_ERRLOG("Unable to allocate memory for stripped_data.\n");
|
|
return -ENOMEM;
|
|
}
|
|
req->stripped_data = data;
|
|
req->stripped_data->iovcnt = 0;
|
|
|
|
rc = nvmf_request_get_buffers(req, group, transport, length, io_unit_size,
|
|
nvmf_request_set_stripped_buffer);
|
|
if (rc == -ENOMEM) {
|
|
nvmf_request_free_stripped_buffers(req, group, transport);
|
|
return rc;
|
|
}
|
|
return rc;
|
|
}
|