The revert asynchronous API doesn't run as the *real* asynchronous way, because the drive can only support synchronous module and only 1 session is supported. The reason why we added this API is that RPC call has the default timeout value here, while the revert may take over several minutes, the API itself doesn't short the revert action, so just remove it and use the synchronous API instead. The revert action will erase all the users data and bring the drive back to the factory state, it should run in the synchronous mode, so just remove the asynchronous API and we can increase the timeout value when using RPC to call this API. Change-Id: I08a082edea6385e378399423bbb229d05f8bc262 Signed-off-by: Changpeng Liu <changpeng.liu@intel.com> Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/1232 Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com> Reviewed-by: Jim Harris <james.r.harris@intel.com>
1200 lines
34 KiB
C
1200 lines
34 KiB
C
/*-
|
|
* BSD LICENSE
|
|
*
|
|
* Copyright (c) Intel Corporation.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include "spdk/stdinc.h"
|
|
#include "spdk_cunit.h"
|
|
#include "spdk/nvme_ocssd_spec.h"
|
|
#include "spdk/thread.h"
|
|
#include "spdk/bdev_module.h"
|
|
#include "spdk/util.h"
|
|
#include "spdk_internal/mock.h"
|
|
|
|
#include "bdev/nvme/bdev_ocssd.c"
|
|
#include "bdev/nvme/common.c"
|
|
#include "common/lib/test_env.c"
|
|
#include "unit/lib/json_mock.c"
|
|
|
|
DEFINE_STUB_V(spdk_bdev_module_list_add, (struct spdk_bdev_module *bdev_module));
|
|
DEFINE_STUB(spdk_nvme_ctrlr_is_ocssd_ns, bool, (struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid),
|
|
true);
|
|
DEFINE_STUB(spdk_nvme_ns_get_extended_sector_size, uint32_t, (struct spdk_nvme_ns *ns), 4096);
|
|
DEFINE_STUB(spdk_nvme_ns_is_active, bool, (struct spdk_nvme_ns *ns), true);
|
|
DEFINE_STUB_V(spdk_opal_dev_destruct, (struct spdk_opal_dev *dev));
|
|
DEFINE_STUB_V(spdk_bdev_io_complete_nvme_status, (struct spdk_bdev_io *bdev_io, uint32_t cdw0,
|
|
int sct, int sc));
|
|
DEFINE_STUB(spdk_bdev_io_get_io_channel, struct spdk_io_channel *, (struct spdk_bdev_io *bdev_io),
|
|
NULL);
|
|
DEFINE_STUB(spdk_bdev_push_media_events, int, (struct spdk_bdev *bdev,
|
|
const struct spdk_bdev_media_event *events,
|
|
size_t num_events), 0);
|
|
DEFINE_STUB_V(spdk_bdev_notify_media_management, (struct spdk_bdev *bdev));
|
|
DEFINE_STUB_V(nvme_ctrlr_depopulate_namespace_done, (struct nvme_bdev_ctrlr *ctrlr));
|
|
DEFINE_STUB_V(spdk_bdev_module_finish_done, (void));
|
|
|
|
struct nvme_request {
|
|
spdk_nvme_cmd_cb cb_fn;
|
|
void *cb_arg;
|
|
TAILQ_ENTRY(nvme_request) tailq;
|
|
};
|
|
|
|
struct spdk_nvme_qpair {
|
|
TAILQ_HEAD(, nvme_request) requests;
|
|
};
|
|
|
|
struct spdk_nvme_ns {
|
|
uint32_t nsid;
|
|
};
|
|
|
|
struct spdk_nvme_ctrlr {
|
|
struct spdk_nvme_transport_id trid;
|
|
struct spdk_ocssd_geometry_data geometry;
|
|
struct spdk_nvme_qpair *admin_qpair;
|
|
struct spdk_nvme_ns *ns;
|
|
uint32_t ns_count;
|
|
struct spdk_ocssd_chunk_information_entry *chunk_info;
|
|
uint64_t num_chunks;
|
|
|
|
LIST_ENTRY(spdk_nvme_ctrlr) list;
|
|
};
|
|
|
|
static LIST_HEAD(, spdk_nvme_ctrlr) g_ctrlr_list = LIST_HEAD_INITIALIZER(g_ctrlr_list);
|
|
static TAILQ_HEAD(, spdk_bdev) g_bdev_list = TAILQ_HEAD_INITIALIZER(g_bdev_list);
|
|
static struct spdk_thread *g_thread;
|
|
|
|
static struct spdk_nvme_ctrlr *
|
|
find_controller(const struct spdk_nvme_transport_id *trid)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
|
|
LIST_FOREACH(ctrlr, &g_ctrlr_list, list) {
|
|
if (!spdk_nvme_transport_id_compare(trid, &ctrlr->trid)) {
|
|
return ctrlr;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
free_controller(struct spdk_nvme_ctrlr *ctrlr)
|
|
{
|
|
CU_ASSERT(!nvme_bdev_ctrlr_get(&ctrlr->trid));
|
|
LIST_REMOVE(ctrlr, list);
|
|
spdk_nvme_ctrlr_free_io_qpair(ctrlr->admin_qpair);
|
|
free(ctrlr->chunk_info);
|
|
free(ctrlr->ns);
|
|
free(ctrlr);
|
|
}
|
|
|
|
static uint64_t
|
|
chunk_offset_to_lba(struct spdk_ocssd_geometry_data *geo, uint64_t offset)
|
|
{
|
|
uint64_t chk, pu, grp;
|
|
uint64_t chk_off, pu_off, grp_off;
|
|
|
|
chk_off = geo->lbaf.lbk_len;
|
|
pu_off = geo->lbaf.chk_len + chk_off;
|
|
grp_off = geo->lbaf.pu_len + pu_off;
|
|
|
|
chk = offset % geo->num_chk;
|
|
pu = (offset / geo->num_chk) % geo->num_pu;
|
|
grp = (offset / (geo->num_chk * geo->num_pu)) % geo->num_grp;
|
|
|
|
return chk << chk_off |
|
|
pu << pu_off |
|
|
grp << grp_off;
|
|
}
|
|
|
|
static struct spdk_nvme_ctrlr *
|
|
create_controller(const struct spdk_nvme_transport_id *trid, uint32_t ns_count,
|
|
const struct spdk_ocssd_geometry_data *geo)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
uint32_t nsid, offset;
|
|
|
|
SPDK_CU_ASSERT_FATAL(!find_controller(trid));
|
|
|
|
ctrlr = calloc(1, sizeof(*ctrlr));
|
|
SPDK_CU_ASSERT_FATAL(ctrlr != NULL);
|
|
|
|
ctrlr->ns = calloc(ns_count, sizeof(*ctrlr->ns));
|
|
SPDK_CU_ASSERT_FATAL(ctrlr->ns != NULL);
|
|
|
|
ctrlr->num_chunks = geo->num_grp * geo->num_pu * geo->num_chk;
|
|
ctrlr->chunk_info = calloc(ctrlr->num_chunks, sizeof(*ctrlr->chunk_info));
|
|
SPDK_CU_ASSERT_FATAL(ctrlr->chunk_info != NULL);
|
|
|
|
for (nsid = 0; nsid < ns_count; ++nsid) {
|
|
ctrlr->ns[nsid].nsid = nsid + 1;
|
|
}
|
|
|
|
ctrlr->geometry = *geo;
|
|
ctrlr->trid = *trid;
|
|
ctrlr->ns_count = ns_count;
|
|
ctrlr->admin_qpair = spdk_nvme_ctrlr_alloc_io_qpair(ctrlr, NULL, 0);
|
|
|
|
for (offset = 0; offset < ctrlr->num_chunks; ++offset) {
|
|
ctrlr->chunk_info[offset].cs.free = 1;
|
|
ctrlr->chunk_info[offset].slba = chunk_offset_to_lba(&ctrlr->geometry, offset);
|
|
ctrlr->chunk_info[offset].wp = ctrlr->chunk_info[offset].slba;
|
|
}
|
|
|
|
SPDK_CU_ASSERT_FATAL(ctrlr->admin_qpair != NULL);
|
|
|
|
LIST_INSERT_HEAD(&g_ctrlr_list, ctrlr, list);
|
|
|
|
return ctrlr;
|
|
}
|
|
|
|
static int
|
|
io_channel_create_cb(void *io_device, void *ctx_buf)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
io_channel_destroy_cb(void *io_device, void *ctx_buf)
|
|
{}
|
|
|
|
void
|
|
nvme_ctrlr_populate_namespace_done(struct nvme_async_probe_ctx *ctx,
|
|
struct nvme_bdev_ns *ns, int rc)
|
|
{
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
}
|
|
|
|
static struct nvme_bdev_ctrlr *
|
|
create_nvme_bdev_controller(const struct spdk_nvme_transport_id *trid, const char *name)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
struct nvme_bdev_ctrlr *nvme_bdev_ctrlr;
|
|
uint32_t nsid;
|
|
|
|
ctrlr = find_controller(trid);
|
|
|
|
SPDK_CU_ASSERT_FATAL(ctrlr != NULL);
|
|
SPDK_CU_ASSERT_FATAL(!nvme_bdev_ctrlr_get(trid));
|
|
|
|
nvme_bdev_ctrlr = calloc(1, sizeof(*nvme_bdev_ctrlr));
|
|
SPDK_CU_ASSERT_FATAL(nvme_bdev_ctrlr != NULL);
|
|
|
|
nvme_bdev_ctrlr->namespaces = calloc(ctrlr->ns_count, sizeof(struct nvme_bdev_ns *));
|
|
SPDK_CU_ASSERT_FATAL(nvme_bdev_ctrlr->namespaces != NULL);
|
|
|
|
nvme_bdev_ctrlr->ctrlr = ctrlr;
|
|
nvme_bdev_ctrlr->num_ns = ctrlr->ns_count;
|
|
nvme_bdev_ctrlr->ref = 0;
|
|
nvme_bdev_ctrlr->trid = *trid;
|
|
nvme_bdev_ctrlr->name = strdup(name);
|
|
|
|
for (nsid = 0; nsid < ctrlr->ns_count; ++nsid) {
|
|
nvme_bdev_ctrlr->namespaces[nsid] = calloc(1, sizeof(struct nvme_bdev_ns));
|
|
SPDK_CU_ASSERT_FATAL(nvme_bdev_ctrlr->namespaces[nsid] != NULL);
|
|
|
|
nvme_bdev_ctrlr->namespaces[nsid]->id = nsid + 1;
|
|
nvme_bdev_ctrlr->namespaces[nsid]->ctrlr = nvme_bdev_ctrlr;
|
|
nvme_bdev_ctrlr->namespaces[nsid]->type = NVME_BDEV_NS_OCSSD;
|
|
TAILQ_INIT(&nvme_bdev_ctrlr->namespaces[nsid]->bdevs);
|
|
|
|
bdev_ocssd_populate_namespace(nvme_bdev_ctrlr, nvme_bdev_ctrlr->namespaces[nsid], NULL);
|
|
}
|
|
|
|
while (spdk_thread_poll(g_thread, 0, 0) > 0) {}
|
|
|
|
spdk_io_device_register(nvme_bdev_ctrlr, io_channel_create_cb,
|
|
io_channel_destroy_cb, 0, name);
|
|
|
|
TAILQ_INSERT_TAIL(&g_nvme_bdev_ctrlrs, nvme_bdev_ctrlr, tailq);
|
|
|
|
return nvme_bdev_ctrlr;
|
|
}
|
|
|
|
static struct nvme_request *
|
|
alloc_request(spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
{
|
|
struct nvme_request *ctx;
|
|
|
|
ctx = calloc(1, sizeof(*ctx));
|
|
SPDK_CU_ASSERT_FATAL(ctx != NULL);
|
|
|
|
ctx->cb_fn = cb_fn;
|
|
ctx->cb_arg = cb_arg;
|
|
|
|
return ctx;
|
|
}
|
|
|
|
uint32_t
|
|
spdk_nvme_ctrlr_get_num_ns(struct spdk_nvme_ctrlr *ctrlr)
|
|
{
|
|
return ctrlr->ns_count;
|
|
}
|
|
|
|
uint32_t
|
|
spdk_nvme_ns_get_id(struct spdk_nvme_ns *ns)
|
|
{
|
|
return ns->nsid;
|
|
}
|
|
|
|
struct spdk_nvme_ns *
|
|
spdk_nvme_ctrlr_get_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid)
|
|
{
|
|
if (nsid == 0 || nsid > ctrlr->ns_count) {
|
|
return NULL;
|
|
}
|
|
|
|
return &ctrlr->ns[nsid - 1];
|
|
}
|
|
|
|
struct spdk_nvme_ctrlr *
|
|
spdk_nvme_connect(const struct spdk_nvme_transport_id *trid,
|
|
const struct spdk_nvme_ctrlr_opts *opts,
|
|
size_t opts_size)
|
|
{
|
|
return find_controller(trid);
|
|
}
|
|
|
|
int
|
|
spdk_nvme_detach(struct spdk_nvme_ctrlr *ctrlr)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
struct spdk_bdev *
|
|
spdk_bdev_get_by_name(const char *bdev_name)
|
|
{
|
|
struct spdk_bdev *bdev;
|
|
|
|
SPDK_CU_ASSERT_FATAL(bdev_name != NULL);
|
|
|
|
TAILQ_FOREACH(bdev, &g_bdev_list, internal.link) {
|
|
if (!strcmp(bdev->name, bdev_name)) {
|
|
return bdev;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
const char *
|
|
spdk_bdev_get_name(const struct spdk_bdev *bdev)
|
|
{
|
|
return bdev->name;
|
|
}
|
|
|
|
int
|
|
spdk_bdev_register(struct spdk_bdev *bdev)
|
|
{
|
|
CU_ASSERT_PTR_NULL(spdk_bdev_get_by_name(bdev->name));
|
|
TAILQ_INSERT_TAIL(&g_bdev_list, bdev, internal.link);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
spdk_bdev_unregister(struct spdk_bdev *bdev, spdk_bdev_unregister_cb cb_fn, void *cb_arg)
|
|
{
|
|
int rc;
|
|
|
|
CU_ASSERT_EQUAL(spdk_bdev_get_by_name(bdev->name), bdev);
|
|
TAILQ_REMOVE(&g_bdev_list, bdev, internal.link);
|
|
|
|
rc = bdev->fn_table->destruct(bdev->ctxt);
|
|
if (rc <= 0 && cb_fn != NULL) {
|
|
cb_fn(cb_arg, 0);
|
|
}
|
|
}
|
|
|
|
size_t
|
|
spdk_bdev_get_zone_size(const struct spdk_bdev *bdev)
|
|
{
|
|
return bdev->zone_size;
|
|
}
|
|
|
|
int
|
|
spdk_nvme_ocssd_ctrlr_cmd_geometry(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
|
|
void *payload, uint32_t payload_size,
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
{
|
|
struct spdk_nvme_cpl cpl = {};
|
|
|
|
CU_ASSERT_EQUAL(payload_size, sizeof(ctrlr->geometry));
|
|
memcpy(payload, &ctrlr->geometry, sizeof(ctrlr->geometry));
|
|
|
|
cb_fn(cb_arg, &cpl);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
spdk_nvme_transport_id_compare(const struct spdk_nvme_transport_id *trid1,
|
|
const struct spdk_nvme_transport_id *trid2)
|
|
{
|
|
return memcmp(trid1, trid2, sizeof(*trid1));
|
|
}
|
|
|
|
void
|
|
spdk_bdev_io_get_buf(struct spdk_bdev_io *bdev_io, spdk_bdev_io_get_buf_cb cb, uint64_t len)
|
|
{
|
|
}
|
|
|
|
void
|
|
spdk_bdev_io_complete(struct spdk_bdev_io *bdev_io, enum spdk_bdev_io_status status)
|
|
{
|
|
}
|
|
|
|
int32_t
|
|
spdk_nvme_ctrlr_process_admin_completions(struct spdk_nvme_ctrlr *ctrlr)
|
|
{
|
|
return spdk_nvme_qpair_process_completions(ctrlr->admin_qpair, 0);
|
|
}
|
|
|
|
struct spdk_nvme_qpair *
|
|
spdk_nvme_ctrlr_alloc_io_qpair(struct spdk_nvme_ctrlr *ctrlr,
|
|
const struct spdk_nvme_io_qpair_opts *opts,
|
|
size_t opts_size)
|
|
{
|
|
struct spdk_nvme_qpair *qpair;
|
|
|
|
qpair = calloc(1, sizeof(*qpair));
|
|
SPDK_CU_ASSERT_FATAL(qpair != NULL);
|
|
|
|
TAILQ_INIT(&qpair->requests);
|
|
return qpair;
|
|
}
|
|
|
|
int
|
|
spdk_nvme_ctrlr_free_io_qpair(struct spdk_nvme_qpair *qpair)
|
|
{
|
|
CU_ASSERT(TAILQ_EMPTY(&qpair->requests));
|
|
free(qpair);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int32_t
|
|
spdk_nvme_qpair_process_completions(struct spdk_nvme_qpair *qpair, uint32_t max_completions)
|
|
{
|
|
struct nvme_request *req;
|
|
struct spdk_nvme_cpl cpl = {};
|
|
int32_t num_requests = 0;
|
|
|
|
while ((req = TAILQ_FIRST(&qpair->requests))) {
|
|
TAILQ_REMOVE(&qpair->requests, req, tailq);
|
|
|
|
req->cb_fn(req->cb_arg, &cpl);
|
|
free(req);
|
|
|
|
num_requests++;
|
|
}
|
|
|
|
return num_requests;
|
|
}
|
|
|
|
int
|
|
spdk_nvme_ns_cmd_readv_with_md(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
uint64_t lba, uint32_t lba_count,
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg, uint32_t io_flags,
|
|
spdk_nvme_req_reset_sgl_cb reset_sgl_fn,
|
|
spdk_nvme_req_next_sge_cb next_sge_fn, void *metadata,
|
|
uint16_t apptag_mask, uint16_t apptag)
|
|
{
|
|
struct nvme_request *req;
|
|
|
|
req = alloc_request(cb_fn, cb_arg);
|
|
TAILQ_INSERT_TAIL(&qpair->requests, req, tailq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
spdk_nvme_ns_cmd_writev_with_md(struct spdk_nvme_ns *ns, struct spdk_nvme_qpair *qpair,
|
|
uint64_t lba, uint32_t lba_count,
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg, uint32_t io_flags,
|
|
spdk_nvme_req_reset_sgl_cb reset_sgl_fn,
|
|
spdk_nvme_req_next_sge_cb next_sge_fn, void *metadata,
|
|
uint16_t apptag_mask, uint16_t apptag)
|
|
{
|
|
struct nvme_request *req;
|
|
|
|
req = alloc_request(cb_fn, cb_arg);
|
|
TAILQ_INSERT_TAIL(&qpair->requests, req, tailq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
spdk_nvme_ocssd_ns_cmd_vector_reset(struct spdk_nvme_ns *ns,
|
|
struct spdk_nvme_qpair *qpair,
|
|
uint64_t *lba_list, uint32_t num_lbas,
|
|
struct spdk_ocssd_chunk_information_entry *chunk_info,
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
{
|
|
struct nvme_request *req;
|
|
|
|
req = alloc_request(cb_fn, cb_arg);
|
|
TAILQ_INSERT_TAIL(&qpair->requests, req, tailq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct spdk_nvme_cpl g_chunk_info_cpl;
|
|
static bool g_zone_info_status = true;
|
|
|
|
int
|
|
spdk_nvme_ctrlr_cmd_get_log_page(struct spdk_nvme_ctrlr *ctrlr,
|
|
uint8_t log_page, uint32_t nsid,
|
|
void *payload, uint32_t payload_size,
|
|
uint64_t offset,
|
|
spdk_nvme_cmd_cb cb_fn, void *cb_arg)
|
|
{
|
|
SPDK_CU_ASSERT_FATAL(offset + payload_size <= sizeof(*ctrlr->chunk_info) * ctrlr->num_chunks);
|
|
memcpy(payload, ((char *)ctrlr->chunk_info) + offset, payload_size);
|
|
|
|
cb_fn(cb_arg, &g_chunk_info_cpl);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
create_bdev_cb(const char *bdev_name, int status, void *ctx)
|
|
{
|
|
*(int *)ctx = status;
|
|
}
|
|
|
|
static int
|
|
create_bdev(const char *ctrlr_name, const char *bdev_name, uint32_t nsid,
|
|
const struct bdev_ocssd_range *range)
|
|
{
|
|
int status = EFAULT;
|
|
|
|
bdev_ocssd_create_bdev(ctrlr_name, bdev_name, nsid, range, create_bdev_cb, &status);
|
|
|
|
while (spdk_thread_poll(g_thread, 0, 0) > 0) {}
|
|
|
|
return status;
|
|
}
|
|
|
|
static void
|
|
delete_nvme_bdev_controller(struct nvme_bdev_ctrlr *nvme_bdev_ctrlr)
|
|
{
|
|
struct nvme_bdev *nvme_bdev, *tmp;
|
|
struct nvme_bdev_ns *nvme_ns;
|
|
bool empty = true;
|
|
uint32_t nsid;
|
|
|
|
nvme_bdev_ctrlr->destruct = true;
|
|
|
|
for (nsid = 0; nsid < nvme_bdev_ctrlr->num_ns; ++nsid) {
|
|
nvme_ns = nvme_bdev_ctrlr->namespaces[nsid];
|
|
|
|
if (!TAILQ_EMPTY(&nvme_ns->bdevs)) {
|
|
TAILQ_FOREACH_SAFE(nvme_bdev, &nvme_ns->bdevs, tailq, tmp) {
|
|
spdk_bdev_unregister(&nvme_bdev->disk, NULL, NULL);
|
|
}
|
|
|
|
empty = false;
|
|
}
|
|
|
|
bdev_ocssd_depopulate_namespace(nvme_bdev_ctrlr->namespaces[nsid]);
|
|
}
|
|
|
|
if (empty) {
|
|
nvme_bdev_ctrlr_destruct(nvme_bdev_ctrlr);
|
|
}
|
|
|
|
while (spdk_thread_poll(g_thread, 0, 0) > 0) {}
|
|
|
|
CU_ASSERT(TAILQ_EMPTY(&g_nvme_bdev_ctrlrs));
|
|
}
|
|
|
|
static void
|
|
test_create_controller(void)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
struct nvme_bdev_ctrlr *nvme_bdev_ctrlr;
|
|
struct spdk_nvme_transport_id trid = { .traddr = "00:00:00" };
|
|
struct spdk_ocssd_geometry_data geometry = {};
|
|
struct spdk_bdev *bdev;
|
|
struct bdev_ocssd_range range;
|
|
const char *controller_name = "nvme0";
|
|
const size_t ns_count = 16;
|
|
char namebuf[128];
|
|
uint32_t nsid;
|
|
int rc;
|
|
|
|
geometry = (struct spdk_ocssd_geometry_data) {
|
|
.clba = 512,
|
|
.num_chk = 64,
|
|
.num_pu = 8,
|
|
.num_grp = 4,
|
|
.maxoc = 69,
|
|
.maxocpu = 68,
|
|
.ws_opt = 86,
|
|
.lbaf = {
|
|
.lbk_len = 9,
|
|
.chk_len = 6,
|
|
.pu_len = 3,
|
|
.grp_len = 2,
|
|
}
|
|
};
|
|
|
|
ctrlr = create_controller(&trid, ns_count, &geometry);
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
for (nsid = 1; nsid <= ns_count; ++nsid) {
|
|
snprintf(namebuf, sizeof(namebuf), "%sn%"PRIu32, controller_name, nsid);
|
|
rc = create_bdev(controller_name, namebuf, nsid, NULL);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
bdev = spdk_bdev_get_by_name(namebuf);
|
|
SPDK_CU_ASSERT_FATAL(bdev != NULL);
|
|
CU_ASSERT_TRUE(bdev->zoned);
|
|
}
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
/* Verify that after deletion the bdevs can still be created */
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
for (nsid = 1; nsid <= ns_count; ++nsid) {
|
|
snprintf(namebuf, sizeof(namebuf), "%sn%"PRIu32, controller_name, nsid);
|
|
rc = create_bdev(controller_name, namebuf, nsid, NULL);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
bdev = spdk_bdev_get_by_name(namebuf);
|
|
SPDK_CU_ASSERT_FATAL(bdev != NULL);
|
|
CU_ASSERT_TRUE(bdev->zoned);
|
|
}
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
/* Verify it's not possible to create a bdev on non-existent namespace */
|
|
rc = create_bdev(controller_name, "invalid", ns_count + 1, NULL);
|
|
CU_ASSERT_EQUAL(rc, -ENODEV);
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
/* Verify the correctness of parallel unit range validation */
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
range.begin = 0;
|
|
range.end = geometry.num_grp * geometry.num_pu;
|
|
|
|
rc = create_bdev(controller_name, "invalid", 1, &range);
|
|
CU_ASSERT_EQUAL(rc, -EINVAL);
|
|
|
|
/* Verify it's not possible for the bdevs to overlap */
|
|
range.begin = 0;
|
|
range.end = 16;
|
|
rc = create_bdev(controller_name, "valid", 1, &range);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
bdev = spdk_bdev_get_by_name("valid");
|
|
CU_ASSERT_PTR_NOT_NULL(bdev);
|
|
|
|
range.begin = 16;
|
|
range.end = 31;
|
|
rc = create_bdev(controller_name, "invalid", 1, &range);
|
|
CU_ASSERT_EQUAL(rc, -EINVAL);
|
|
|
|
/* But it is possible to create them without overlap */
|
|
range.begin = 17;
|
|
range.end = 31;
|
|
rc = create_bdev(controller_name, "valid2", 1, &range);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
bdev = spdk_bdev_get_by_name("valid2");
|
|
CU_ASSERT_PTR_NOT_NULL(bdev);
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
free_controller(ctrlr);
|
|
}
|
|
|
|
static void
|
|
test_device_geometry(void)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
struct nvme_bdev_ctrlr *nvme_bdev_ctrlr;
|
|
struct spdk_nvme_transport_id trid = { .traddr = "00:00:00" };
|
|
const char *controller_name = "nvme0";
|
|
const char *bdev_name = "nvme0n1";
|
|
struct spdk_ocssd_geometry_data geometry;
|
|
struct spdk_bdev *bdev;
|
|
int rc;
|
|
|
|
geometry = (struct spdk_ocssd_geometry_data) {
|
|
.clba = 512,
|
|
.num_chk = 64,
|
|
.num_pu = 8,
|
|
.num_grp = 4,
|
|
.maxoc = 69,
|
|
.maxocpu = 68,
|
|
.ws_opt = 86,
|
|
.lbaf = {
|
|
.lbk_len = 9,
|
|
.chk_len = 6,
|
|
.pu_len = 3,
|
|
.grp_len = 2,
|
|
}
|
|
};
|
|
|
|
ctrlr = create_controller(&trid, 1, &geometry);
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
rc = create_bdev(controller_name, bdev_name, 1, NULL);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
bdev = spdk_bdev_get_by_name(bdev_name);
|
|
CU_ASSERT_EQUAL(bdev->blockcnt, geometry.clba *
|
|
geometry.num_chk *
|
|
geometry.num_pu *
|
|
geometry.num_grp);
|
|
CU_ASSERT_EQUAL(bdev->zone_size, geometry.clba);
|
|
CU_ASSERT_EQUAL(bdev->optimal_open_zones, geometry.num_pu * geometry.num_grp);
|
|
CU_ASSERT_EQUAL(bdev->max_open_zones, geometry.maxocpu);
|
|
CU_ASSERT_EQUAL(bdev->write_unit_size, geometry.ws_opt);
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
free_controller(ctrlr);
|
|
}
|
|
|
|
static uint64_t
|
|
generate_lba(const struct spdk_ocssd_geometry_data *geo, uint64_t lbk,
|
|
uint64_t chk, uint64_t pu, uint64_t grp)
|
|
{
|
|
uint64_t lba, len;
|
|
|
|
lba = lbk;
|
|
len = geo->lbaf.lbk_len;
|
|
CU_ASSERT(lbk < (1ull << geo->lbaf.lbk_len));
|
|
|
|
lba |= chk << len;
|
|
len += geo->lbaf.chk_len;
|
|
CU_ASSERT(chk < (1ull << geo->lbaf.chk_len));
|
|
|
|
lba |= pu << len;
|
|
len += geo->lbaf.pu_len;
|
|
CU_ASSERT(pu < (1ull << geo->lbaf.pu_len));
|
|
|
|
lba |= grp << len;
|
|
|
|
return lba;
|
|
}
|
|
|
|
static void
|
|
test_lba_translation(void)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
struct nvme_bdev_ctrlr *nvme_bdev_ctrlr;
|
|
struct spdk_nvme_transport_id trid = { .traddr = "00:00:00" };
|
|
const char *controller_name = "nvme0";
|
|
const char *bdev_name = "nvme0n1";
|
|
struct spdk_ocssd_geometry_data geometry = {};
|
|
struct ocssd_bdev *ocssd_bdev;
|
|
struct spdk_bdev *bdev;
|
|
uint64_t lba;
|
|
int rc;
|
|
|
|
geometry = (struct spdk_ocssd_geometry_data) {
|
|
.clba = 512,
|
|
.num_chk = 64,
|
|
.num_pu = 8,
|
|
.num_grp = 4,
|
|
.lbaf = {
|
|
.lbk_len = 9,
|
|
.chk_len = 6,
|
|
.pu_len = 3,
|
|
.grp_len = 2,
|
|
}
|
|
};
|
|
|
|
ctrlr = create_controller(&trid, 1, &geometry);
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
rc = create_bdev(controller_name, bdev_name, 1, NULL);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
bdev = spdk_bdev_get_by_name(bdev_name);
|
|
SPDK_CU_ASSERT_FATAL(bdev != NULL);
|
|
ocssd_bdev = SPDK_CONTAINEROF(bdev, struct ocssd_bdev, nvme_bdev.disk);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, 0);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 0, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), 0);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size - 1);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, bdev->zone_size - 1, 0, 0, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size - 1);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 1, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size * geometry.num_pu);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 0, 1));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size * geometry.num_pu);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size * geometry.num_pu + 68);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 68, 0, 0, 1));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size * geometry.num_pu + 68);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size + 68);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 68, 0, 1, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size + 68);
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
free_controller(ctrlr);
|
|
|
|
geometry = (struct spdk_ocssd_geometry_data) {
|
|
.clba = 5120,
|
|
.num_chk = 501,
|
|
.num_pu = 9,
|
|
.num_grp = 1,
|
|
.lbaf = {
|
|
.lbk_len = 13,
|
|
.chk_len = 9,
|
|
.pu_len = 4,
|
|
.grp_len = 1,
|
|
}
|
|
};
|
|
|
|
ctrlr = create_controller(&trid, 1, &geometry);
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
rc = create_bdev(controller_name, bdev_name, 1, NULL);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
bdev = spdk_bdev_get_by_name(bdev_name);
|
|
SPDK_CU_ASSERT_FATAL(bdev != NULL);
|
|
ocssd_bdev = SPDK_CONTAINEROF(bdev, struct ocssd_bdev, nvme_bdev.disk);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, 0);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 0, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), 0);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size - 1);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, bdev->zone_size - 1, 0, 0, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size - 1);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 1, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size * (geometry.num_pu - 1));
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, geometry.num_pu - 1, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba), bdev->zone_size * (geometry.num_pu - 1));
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size * geometry.num_pu * geometry.num_grp);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 1, 0, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba),
|
|
bdev->zone_size * geometry.num_pu * geometry.num_grp);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev, bdev->zone_size * geometry.num_pu * geometry.num_grp + 68);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 68, 1, 0, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev, lba),
|
|
bdev->zone_size * geometry.num_pu * geometry.num_grp + 68);
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
free_controller(ctrlr);
|
|
}
|
|
|
|
static void
|
|
punit_range_to_addr(const struct spdk_nvme_ctrlr *ctrlr, uint64_t punit,
|
|
uint64_t *grp, uint64_t *pu)
|
|
{
|
|
const struct spdk_ocssd_geometry_data *geo = &ctrlr->geometry;
|
|
|
|
*grp = punit / geo->num_pu;
|
|
*pu = punit % geo->num_pu;
|
|
|
|
CU_ASSERT(*grp < geo->num_grp);
|
|
}
|
|
|
|
static void
|
|
test_parallel_unit_range(void)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
struct nvme_bdev_ctrlr *nvme_bdev_ctrlr;
|
|
struct spdk_nvme_transport_id trid = { .traddr = "00:00:00" };
|
|
const char *controller_name = "nvme0";
|
|
const char *bdev_name[] = { "nvme0n1", "nvme0n2", "nvme0n3" };
|
|
const struct bdev_ocssd_range range[3] = { { 0, 5 }, { 6, 18 }, { 19, 23 } };
|
|
struct ocssd_bdev *ocssd_bdev[3];
|
|
struct spdk_ocssd_geometry_data geometry = {};
|
|
struct spdk_bdev *bdev[3];
|
|
uint64_t lba, i, offset, grp, pu, zone_size;
|
|
int rc;
|
|
|
|
geometry = (struct spdk_ocssd_geometry_data) {
|
|
.clba = 500,
|
|
.num_chk = 60,
|
|
.num_pu = 8,
|
|
.num_grp = 3,
|
|
.lbaf = {
|
|
.lbk_len = 9,
|
|
.chk_len = 6,
|
|
.pu_len = 3,
|
|
.grp_len = 2,
|
|
}
|
|
};
|
|
|
|
ctrlr = create_controller(&trid, 1, &geometry);
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
for (i = 0; i < SPDK_COUNTOF(range); ++i) {
|
|
rc = create_bdev(controller_name, bdev_name[i], 1, &range[i]);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
bdev[i] = spdk_bdev_get_by_name(bdev_name[i]);
|
|
SPDK_CU_ASSERT_FATAL(bdev[i] != NULL);
|
|
ocssd_bdev[i] = SPDK_CONTAINEROF(bdev[i], struct ocssd_bdev, nvme_bdev.disk);
|
|
}
|
|
|
|
zone_size = bdev[0]->zone_size;
|
|
CU_ASSERT_EQUAL(zone_size, bdev[1]->zone_size);
|
|
CU_ASSERT_EQUAL(zone_size, bdev[2]->zone_size);
|
|
|
|
/* Verify the first addresses are correct */
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[0], 0);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 0, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[0], lba), 0);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[1], 0);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 6, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[1], lba), 0);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[2], 0);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 0, 0, 3, 2));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[2], lba), 0);
|
|
|
|
/* Verify last address correctness */
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[0], bdev[0]->blockcnt - 1);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, geometry.clba - 1, geometry.num_chk - 1, 5, 0));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[0], lba), bdev[0]->blockcnt - 1);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[1], bdev[1]->blockcnt - 1);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, geometry.clba - 1, geometry.num_chk - 1, 2, 2));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[1], lba), bdev[1]->blockcnt - 1);
|
|
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[2], bdev[2]->blockcnt - 1);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, geometry.clba - 1, geometry.num_chk - 1, 7, 2));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[2], lba), bdev[2]->blockcnt - 1);
|
|
|
|
/* Verify correct jumps across parallel units / groups */
|
|
for (i = 0; i < SPDK_COUNTOF(range); ++i) {
|
|
for (offset = 0; offset < bdev_ocssd_num_parallel_units(ocssd_bdev[i]); ++offset) {
|
|
punit_range_to_addr(ctrlr, range[i].begin + offset, &grp, &pu);
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[i], offset * zone_size + 68);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 68, 0, pu, grp));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[i], lba),
|
|
offset * zone_size + 68);
|
|
}
|
|
}
|
|
|
|
/* Verify correct address wrapping */
|
|
for (i = 0; i < SPDK_COUNTOF(range); ++i) {
|
|
punit_range_to_addr(ctrlr, range[i].begin, &grp, &pu);
|
|
|
|
offset = bdev_ocssd_num_parallel_units(ocssd_bdev[i]) * zone_size + 68;
|
|
lba = bdev_ocssd_to_disk_lba(ocssd_bdev[i], offset);
|
|
CU_ASSERT_EQUAL(lba, generate_lba(&geometry, 68, 1, pu, grp));
|
|
assert(lba == generate_lba(&geometry, 68, 1, pu, grp));
|
|
CU_ASSERT_EQUAL(bdev_ocssd_from_disk_lba(ocssd_bdev[i], lba), offset);
|
|
}
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
free_controller(ctrlr);
|
|
}
|
|
|
|
static void
|
|
get_zone_info_cb(struct spdk_bdev_io *bdev_io, bool success, void *cb_arg)
|
|
{
|
|
CU_ASSERT_EQUAL(g_zone_info_status, success);
|
|
}
|
|
|
|
static uint64_t
|
|
generate_chunk_offset(const struct spdk_ocssd_geometry_data *geo, uint64_t chk,
|
|
uint64_t pu, uint64_t grp)
|
|
{
|
|
return grp * geo->num_pu * geo->num_chk +
|
|
pu * geo->num_chk + chk;
|
|
}
|
|
|
|
static struct spdk_bdev_io *
|
|
alloc_ocssd_io(void)
|
|
{
|
|
struct spdk_bdev_io *bdev_io;
|
|
|
|
bdev_io = calloc(1, sizeof(struct spdk_bdev_io) + sizeof(struct bdev_ocssd_io));
|
|
SPDK_CU_ASSERT_FATAL(bdev_io != NULL);
|
|
|
|
return bdev_io;
|
|
}
|
|
|
|
static struct spdk_ocssd_chunk_information_entry *
|
|
get_chunk_info(struct spdk_nvme_ctrlr *ctrlr, uint64_t offset)
|
|
{
|
|
assert(offset < ctrlr->num_chunks);
|
|
SPDK_CU_ASSERT_FATAL(offset < ctrlr->num_chunks);
|
|
return &ctrlr->chunk_info[offset];
|
|
}
|
|
|
|
enum chunk_state {
|
|
CHUNK_STATE_FREE,
|
|
CHUNK_STATE_CLOSED,
|
|
CHUNK_STATE_OPEN,
|
|
CHUNK_STATE_OFFLINE
|
|
};
|
|
|
|
static void
|
|
set_chunk_state(struct spdk_ocssd_chunk_information_entry *chunk, enum chunk_state state)
|
|
{
|
|
memset(&chunk->cs, 0, sizeof(chunk->cs));
|
|
switch (state) {
|
|
case CHUNK_STATE_FREE:
|
|
chunk->cs.free = 1;
|
|
break;
|
|
case CHUNK_STATE_CLOSED:
|
|
chunk->cs.closed = 1;
|
|
break;
|
|
case CHUNK_STATE_OPEN:
|
|
chunk->cs.open = 1;
|
|
break;
|
|
case CHUNK_STATE_OFFLINE:
|
|
chunk->cs.offline = 1;
|
|
break;
|
|
default:
|
|
SPDK_CU_ASSERT_FATAL(0 && "Invalid state");
|
|
}
|
|
}
|
|
|
|
static void
|
|
test_get_zone_info(void)
|
|
{
|
|
struct spdk_nvme_ctrlr *ctrlr;
|
|
struct nvme_bdev_ctrlr *nvme_bdev_ctrlr;
|
|
struct spdk_nvme_transport_id trid = { .traddr = "00:00:00" };
|
|
const char *controller_name = "nvme0";
|
|
const char *bdev_name = "nvme0n1";
|
|
struct spdk_bdev *bdev;
|
|
struct spdk_bdev_io *bdev_io;
|
|
#define MAX_ZONE_INFO_COUNT 64
|
|
struct spdk_bdev_zone_info zone_info[MAX_ZONE_INFO_COUNT];
|
|
struct spdk_ocssd_chunk_information_entry *chunk_info;
|
|
struct spdk_ocssd_geometry_data geometry;
|
|
uint64_t chunk_offset;
|
|
int rc, offset;
|
|
|
|
geometry = (struct spdk_ocssd_geometry_data) {
|
|
.clba = 512,
|
|
.num_chk = 64,
|
|
.num_pu = 8,
|
|
.num_grp = 4,
|
|
.lbaf = {
|
|
.lbk_len = 9,
|
|
.chk_len = 6,
|
|
.pu_len = 3,
|
|
.grp_len = 2,
|
|
}
|
|
};
|
|
|
|
ctrlr = create_controller(&trid, 1, &geometry);
|
|
nvme_bdev_ctrlr = create_nvme_bdev_controller(&trid, controller_name);
|
|
|
|
rc = create_bdev(controller_name, bdev_name, 1, NULL);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
bdev = spdk_bdev_get_by_name(bdev_name);
|
|
SPDK_CU_ASSERT_FATAL(bdev != NULL);
|
|
|
|
bdev_io = alloc_ocssd_io();
|
|
bdev_io->internal.cb = get_zone_info_cb;
|
|
bdev_io->bdev = bdev;
|
|
|
|
/* Verify empty zone */
|
|
bdev_io->u.zone_mgmt.zone_id = 0;
|
|
bdev_io->u.zone_mgmt.num_zones = 1;
|
|
bdev_io->u.zone_mgmt.buf = &zone_info;
|
|
chunk_info = get_chunk_info(ctrlr, 0);
|
|
set_chunk_state(chunk_info, CHUNK_STATE_FREE);
|
|
chunk_info->wp = 0;
|
|
|
|
rc = bdev_ocssd_get_zone_info(NULL, bdev_io);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
CU_ASSERT_EQUAL(zone_info[0].state, SPDK_BDEV_ZONE_STATE_EMPTY);
|
|
CU_ASSERT_EQUAL(zone_info[0].zone_id, 0);
|
|
CU_ASSERT_EQUAL(zone_info[0].write_pointer, 0);
|
|
CU_ASSERT_EQUAL(zone_info[0].capacity, geometry.clba);
|
|
|
|
/* Verify open zone */
|
|
bdev_io->u.zone_mgmt.zone_id = bdev->zone_size;
|
|
bdev_io->u.zone_mgmt.num_zones = 1;
|
|
bdev_io->u.zone_mgmt.buf = &zone_info;
|
|
chunk_info = get_chunk_info(ctrlr, generate_chunk_offset(&geometry, 0, 1, 0));
|
|
set_chunk_state(chunk_info, CHUNK_STATE_OPEN);
|
|
chunk_info->wp = chunk_info->slba + 68;
|
|
chunk_info->cnlb = 511;
|
|
chunk_info->ct.size_deviate = 1;
|
|
|
|
rc = bdev_ocssd_get_zone_info(NULL, bdev_io);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
CU_ASSERT_EQUAL(zone_info[0].state, SPDK_BDEV_ZONE_STATE_OPEN);
|
|
CU_ASSERT_EQUAL(zone_info[0].zone_id, bdev->zone_size);
|
|
CU_ASSERT_EQUAL(zone_info[0].write_pointer, bdev->zone_size + 68);
|
|
CU_ASSERT_EQUAL(zone_info[0].capacity, chunk_info->cnlb);
|
|
|
|
/* Verify offline zone at 2nd chunk */
|
|
bdev_io->u.zone_mgmt.zone_id = bdev->zone_size * geometry.num_pu * geometry.num_grp;
|
|
bdev_io->u.zone_mgmt.num_zones = 1;
|
|
bdev_io->u.zone_mgmt.buf = &zone_info;
|
|
chunk_info = get_chunk_info(ctrlr, generate_chunk_offset(&geometry, 1, 0, 0));
|
|
set_chunk_state(chunk_info, CHUNK_STATE_OFFLINE);
|
|
chunk_info->wp = chunk_info->slba;
|
|
|
|
rc = bdev_ocssd_get_zone_info(NULL, bdev_io);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
CU_ASSERT_EQUAL(zone_info[0].state, SPDK_BDEV_ZONE_STATE_OFFLINE);
|
|
CU_ASSERT_EQUAL(zone_info[0].zone_id, bdev_io->u.zone_mgmt.zone_id);
|
|
CU_ASSERT_EQUAL(zone_info[0].write_pointer, bdev_io->u.zone_mgmt.zone_id);
|
|
|
|
/* Verify multiple zones at a time */
|
|
bdev_io->u.zone_mgmt.zone_id = 0;
|
|
bdev_io->u.zone_mgmt.num_zones = MAX_ZONE_INFO_COUNT;
|
|
bdev_io->u.zone_mgmt.buf = &zone_info;
|
|
|
|
for (offset = 0; offset < MAX_ZONE_INFO_COUNT; ++offset) {
|
|
chunk_offset = generate_chunk_offset(&geometry,
|
|
(offset / (geometry.num_grp * geometry.num_pu)) % geometry.num_chk,
|
|
offset % geometry.num_pu,
|
|
(offset / geometry.num_pu) % geometry.num_grp);
|
|
|
|
|
|
chunk_info = get_chunk_info(ctrlr, chunk_offset);
|
|
set_chunk_state(chunk_info, CHUNK_STATE_OPEN);
|
|
chunk_info->wp = chunk_info->slba + 68;
|
|
chunk_info->ct.size_deviate = 0;
|
|
}
|
|
|
|
rc = bdev_ocssd_get_zone_info(NULL, bdev_io);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
|
|
for (offset = 0; offset < MAX_ZONE_INFO_COUNT; ++offset) {
|
|
CU_ASSERT_EQUAL(zone_info[offset].state, SPDK_BDEV_ZONE_STATE_OPEN);
|
|
CU_ASSERT_EQUAL(zone_info[offset].zone_id, bdev->zone_size * offset);
|
|
CU_ASSERT_EQUAL(zone_info[offset].write_pointer, bdev->zone_size * offset + 68);
|
|
CU_ASSERT_EQUAL(zone_info[offset].capacity, geometry.clba);
|
|
}
|
|
|
|
/* Verify misaligned start zone LBA */
|
|
bdev_io->u.zone_mgmt.zone_id = 1;
|
|
bdev_io->u.zone_mgmt.num_zones = MAX_ZONE_INFO_COUNT;
|
|
bdev_io->u.zone_mgmt.buf = &zone_info;
|
|
|
|
rc = bdev_ocssd_get_zone_info(NULL, bdev_io);
|
|
CU_ASSERT_EQUAL(rc, -EINVAL);
|
|
|
|
/* Verify correct NVMe error forwarding */
|
|
bdev_io->u.zone_mgmt.zone_id = 0;
|
|
bdev_io->u.zone_mgmt.num_zones = MAX_ZONE_INFO_COUNT;
|
|
bdev_io->u.zone_mgmt.buf = &zone_info;
|
|
chunk_info = get_chunk_info(ctrlr, 0);
|
|
set_chunk_state(chunk_info, CHUNK_STATE_FREE);
|
|
|
|
rc = bdev_ocssd_get_zone_info(NULL, bdev_io);
|
|
CU_ASSERT_EQUAL(rc, 0);
|
|
g_chunk_info_cpl = (struct spdk_nvme_cpl) {
|
|
.status = {
|
|
.sct = SPDK_NVME_SCT_GENERIC,
|
|
.sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR
|
|
}
|
|
};
|
|
g_zone_info_status = false;
|
|
|
|
g_chunk_info_cpl = (struct spdk_nvme_cpl) {};
|
|
g_zone_info_status = true;
|
|
|
|
delete_nvme_bdev_controller(nvme_bdev_ctrlr);
|
|
|
|
free(bdev_io);
|
|
free_controller(ctrlr);
|
|
}
|
|
|
|
int
|
|
main(int argc, const char **argv)
|
|
{
|
|
CU_pSuite suite = NULL;
|
|
unsigned int num_failures;
|
|
|
|
if (CU_initialize_registry() != CUE_SUCCESS) {
|
|
return CU_get_error();
|
|
}
|
|
|
|
suite = CU_add_suite("ocssd", NULL, NULL);
|
|
if (suite == NULL) {
|
|
CU_cleanup_registry();
|
|
return CU_get_error();
|
|
}
|
|
|
|
if (
|
|
CU_add_test(suite, "test_create_controller", test_create_controller) == NULL ||
|
|
CU_add_test(suite, "test_device_geometry", test_device_geometry) == NULL ||
|
|
CU_add_test(suite, "test_lba_translation", test_lba_translation) == NULL ||
|
|
CU_add_test(suite, "test_parallel_unit_range", test_parallel_unit_range) == NULL ||
|
|
CU_add_test(suite, "test_get_zone_info", test_get_zone_info) == NULL
|
|
) {
|
|
CU_cleanup_registry();
|
|
return CU_get_error();
|
|
}
|
|
|
|
g_thread = spdk_thread_create("test", NULL);
|
|
spdk_set_thread(g_thread);
|
|
|
|
CU_basic_set_mode(CU_BRM_VERBOSE);
|
|
CU_basic_run_tests();
|
|
num_failures = CU_get_number_of_failures();
|
|
|
|
spdk_thread_exit(g_thread);
|
|
spdk_thread_destroy(g_thread);
|
|
|
|
CU_cleanup_registry();
|
|
|
|
return num_failures;
|
|
}
|