NVMf target reports copy command support if all bdevs in the subsystem support copy IO type. Maximum copy size is reported for each namespace independently in namespace identify data. For now we support just one source range. Note, that command support in the controller is initialized once on controller create. If another namespace which doesn't support copy command is added to the subsystem later, it will not be reflected in the controller data structure and will not be communicated to the initiator. Attempt to execute copy command on such namespace will fail. This issue is not specific to copy command and applies also to write zeroes and unmap (dataset management) commands. Signed-off-by: Evgeniy Kochetov <evgeniik@nvidia.com> Change-Id: I5f06564eb43d66d2852bf7eeda8b17830c53c9bc Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/14350 Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com> Reviewed-by: Shuhei Matsumoto <smatsumoto@nvidia.com> Reviewed-by: Aleksey Marchuk <alexeymar@nvidia.com>
948 lines
30 KiB
C
948 lines
30 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright (C) 2017 Intel Corporation.
|
|
* All rights reserved.
|
|
* Copyright (c) 2021 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
*/
|
|
|
|
#include "spdk/stdinc.h"
|
|
|
|
#include "spdk_cunit.h"
|
|
|
|
#include "spdk_internal/mock.h"
|
|
#include "thread/thread_internal.h"
|
|
|
|
#include "nvmf/ctrlr_bdev.c"
|
|
|
|
#include "spdk/bdev_module.h"
|
|
|
|
SPDK_LOG_REGISTER_COMPONENT(nvmf)
|
|
|
|
DEFINE_STUB(spdk_nvmf_request_complete, int, (struct spdk_nvmf_request *req), -1);
|
|
|
|
DEFINE_STUB(spdk_bdev_get_name, const char *, (const struct spdk_bdev *bdev), "test");
|
|
|
|
DEFINE_STUB(spdk_bdev_get_physical_block_size, uint32_t,
|
|
(const struct spdk_bdev *bdev), 4096);
|
|
|
|
DEFINE_STUB(nvmf_ctrlr_process_admin_cmd, int, (struct spdk_nvmf_request *req), 0);
|
|
|
|
DEFINE_STUB(spdk_bdev_comparev_blocks, int, (struct spdk_bdev_desc *desc,
|
|
struct spdk_io_channel *ch, struct iovec *iov, int iovcnt,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg), 0);
|
|
|
|
DEFINE_STUB(spdk_bdev_nvme_admin_passthru, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
const struct spdk_nvme_cmd *cmd, void *buf, size_t nbytes,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg), 0);
|
|
|
|
DEFINE_STUB(spdk_bdev_abort, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
void *bio_cb_arg, spdk_bdev_io_completion_cb cb, void *cb_arg), 0);
|
|
|
|
DEFINE_STUB_V(spdk_bdev_io_get_iovec,
|
|
(struct spdk_bdev_io *bdev_io, struct iovec **iovp, int *iovcntp));
|
|
|
|
uint32_t
|
|
spdk_bdev_get_optimal_io_boundary(const struct spdk_bdev *bdev)
|
|
{
|
|
return bdev->optimal_io_boundary;
|
|
}
|
|
|
|
uint32_t
|
|
spdk_bdev_get_md_size(const struct spdk_bdev *bdev)
|
|
{
|
|
return bdev->md_len;
|
|
}
|
|
|
|
bool
|
|
spdk_bdev_is_md_interleaved(const struct spdk_bdev *bdev)
|
|
{
|
|
return (bdev->md_len != 0) && bdev->md_interleave;
|
|
}
|
|
|
|
/* We have to use the typedef in the function declaration to appease astyle. */
|
|
typedef enum spdk_dif_type spdk_dif_type_t;
|
|
|
|
spdk_dif_type_t
|
|
spdk_bdev_get_dif_type(const struct spdk_bdev *bdev)
|
|
{
|
|
if (bdev->md_len != 0) {
|
|
return bdev->dif_type;
|
|
} else {
|
|
return SPDK_DIF_DISABLE;
|
|
}
|
|
}
|
|
|
|
bool
|
|
spdk_bdev_is_dif_head_of_md(const struct spdk_bdev *bdev)
|
|
{
|
|
if (spdk_bdev_get_dif_type(bdev) != SPDK_DIF_DISABLE) {
|
|
return bdev->dif_is_head_of_md;
|
|
} else {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
uint32_t
|
|
spdk_bdev_get_data_block_size(const struct spdk_bdev *bdev)
|
|
{
|
|
if (spdk_bdev_is_md_interleaved(bdev)) {
|
|
return bdev->blocklen - bdev->md_len;
|
|
} else {
|
|
return bdev->blocklen;
|
|
}
|
|
}
|
|
|
|
uint16_t
|
|
spdk_bdev_get_acwu(const struct spdk_bdev *bdev)
|
|
{
|
|
return bdev->acwu;
|
|
}
|
|
|
|
uint32_t
|
|
spdk_bdev_get_block_size(const struct spdk_bdev *bdev)
|
|
{
|
|
return bdev->blocklen;
|
|
}
|
|
|
|
uint64_t
|
|
spdk_bdev_get_num_blocks(const struct spdk_bdev *bdev)
|
|
{
|
|
return bdev->blockcnt;
|
|
}
|
|
|
|
DEFINE_STUB(spdk_bdev_comparev_and_writev_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
struct iovec *compare_iov, int compare_iovcnt,
|
|
struct iovec *write_iov, int write_iovcnt,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(nvmf_ctrlr_process_io_cmd, int, (struct spdk_nvmf_request *req), 0);
|
|
|
|
DEFINE_STUB_V(spdk_bdev_io_get_nvme_fused_status, (const struct spdk_bdev_io *bdev_io,
|
|
uint32_t *cdw0, int *cmp_sct, int *cmp_sc, int *wr_sct, int *wr_sc));
|
|
|
|
DEFINE_STUB(spdk_bdev_is_dif_check_enabled, bool,
|
|
(const struct spdk_bdev *bdev, enum spdk_dif_check_type check_type), false);
|
|
|
|
DEFINE_STUB(spdk_bdev_get_io_channel, struct spdk_io_channel *,
|
|
(struct spdk_bdev_desc *desc), NULL);
|
|
|
|
DEFINE_STUB(spdk_bdev_flush_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_unmap_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_io_type_supported, bool,
|
|
(struct spdk_bdev *bdev, enum spdk_bdev_io_type io_type), false);
|
|
|
|
DEFINE_STUB(spdk_bdev_queue_io_wait, int,
|
|
(struct spdk_bdev *bdev, struct spdk_io_channel *ch,
|
|
struct spdk_bdev_io_wait_entry *entry),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_write_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, void *buf,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_writev_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
struct iovec *iov, int iovcnt, uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_read_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, void *buf,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_readv_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
struct iovec *iov, int iovcnt, uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_write_zeroes_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_nvme_io_passthru, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
const struct spdk_nvme_cmd *cmd, void *buf, size_t nbytes,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB_V(spdk_bdev_free_io, (struct spdk_bdev_io *bdev_io));
|
|
|
|
DEFINE_STUB(spdk_nvmf_subsystem_get_nqn, const char *,
|
|
(const struct spdk_nvmf_subsystem *subsystem), NULL);
|
|
|
|
DEFINE_STUB(spdk_bdev_zcopy_start, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
struct iovec *iov, int iovcnt,
|
|
uint64_t offset_blocks, uint64_t num_blocks,
|
|
bool populate,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_zcopy_end, int,
|
|
(struct spdk_bdev_io *bdev_io, bool commit,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_copy_blocks, int,
|
|
(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
|
|
uint64_t dst_offset_blocks, uint64_t src_offset_blocks, uint64_t num_blocks,
|
|
spdk_bdev_io_completion_cb cb, void *cb_arg),
|
|
0);
|
|
|
|
DEFINE_STUB(spdk_bdev_get_max_copy, uint32_t, (const struct spdk_bdev *bdev), 0);
|
|
|
|
struct spdk_nvmf_ns *
|
|
spdk_nvmf_subsystem_get_ns(struct spdk_nvmf_subsystem *subsystem, uint32_t nsid)
|
|
{
|
|
abort();
|
|
return NULL;
|
|
}
|
|
|
|
struct spdk_nvmf_ns *
|
|
spdk_nvmf_subsystem_get_first_ns(struct spdk_nvmf_subsystem *subsystem)
|
|
{
|
|
abort();
|
|
return NULL;
|
|
}
|
|
|
|
struct spdk_nvmf_ns *
|
|
spdk_nvmf_subsystem_get_next_ns(struct spdk_nvmf_subsystem *subsystem, struct spdk_nvmf_ns *prev_ns)
|
|
{
|
|
abort();
|
|
return NULL;
|
|
}
|
|
|
|
int
|
|
spdk_dif_ctx_init(struct spdk_dif_ctx *ctx, uint32_t block_size, uint32_t md_size,
|
|
bool md_interleave, bool dif_loc, enum spdk_dif_type dif_type, uint32_t dif_flags,
|
|
uint32_t init_ref_tag, uint16_t apptag_mask, uint16_t app_tag,
|
|
uint32_t data_offset, uint16_t guard_seed)
|
|
{
|
|
ctx->block_size = block_size;
|
|
ctx->md_size = md_size;
|
|
ctx->init_ref_tag = init_ref_tag;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static uint32_t g_bdev_nvme_status_cdw0;
|
|
static uint32_t g_bdev_nvme_status_sct = SPDK_NVME_SCT_GENERIC;
|
|
static uint32_t g_bdev_nvme_status_sc = SPDK_NVME_SC_SUCCESS;
|
|
|
|
static void
|
|
reset_bdev_nvme_status(void)
|
|
{
|
|
g_bdev_nvme_status_cdw0 = 0;
|
|
g_bdev_nvme_status_sct = SPDK_NVME_SCT_GENERIC;
|
|
g_bdev_nvme_status_sc = SPDK_NVME_SC_SUCCESS;
|
|
}
|
|
|
|
void
|
|
spdk_bdev_io_get_nvme_status(const struct spdk_bdev_io *bdev_io, uint32_t *cdw0, int *sct,
|
|
int *sc)
|
|
{
|
|
*cdw0 = g_bdev_nvme_status_cdw0;
|
|
*sct = g_bdev_nvme_status_sct;
|
|
*sc = g_bdev_nvme_status_sc;
|
|
}
|
|
|
|
static void
|
|
test_get_rw_params(void)
|
|
{
|
|
struct spdk_nvme_cmd cmd = {0};
|
|
uint64_t lba;
|
|
uint64_t count;
|
|
|
|
lba = 0;
|
|
count = 0;
|
|
to_le64(&cmd.cdw10, 0x1234567890ABCDEF);
|
|
to_le32(&cmd.cdw12, 0x9875 | SPDK_NVME_IO_FLAGS_FORCE_UNIT_ACCESS);
|
|
nvmf_bdev_ctrlr_get_rw_params(&cmd, &lba, &count);
|
|
CU_ASSERT(lba == 0x1234567890ABCDEF);
|
|
CU_ASSERT(count == 0x9875 + 1); /* NOTE: this field is 0's based, hence the +1 */
|
|
}
|
|
|
|
static void
|
|
test_lba_in_range(void)
|
|
{
|
|
/* Trivial cases (no overflow) */
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 0, 1) == true);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 0, 1000) == true);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 0, 1001) == false);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1, 999) == true);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1, 1000) == false);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 999, 1) == true);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1000, 1) == false);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1001, 1) == false);
|
|
|
|
/* Overflow edge cases */
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, 0, UINT64_MAX) == true);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, 1, UINT64_MAX) == false);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, UINT64_MAX - 1, 1) == true);
|
|
CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, UINT64_MAX, 1) == false);
|
|
}
|
|
|
|
static void
|
|
test_get_dif_ctx(void)
|
|
{
|
|
struct spdk_bdev bdev = {};
|
|
struct spdk_nvme_cmd cmd = {};
|
|
struct spdk_dif_ctx dif_ctx = {};
|
|
bool ret;
|
|
|
|
bdev.md_len = 0;
|
|
|
|
ret = nvmf_bdev_ctrlr_get_dif_ctx(&bdev, &cmd, &dif_ctx);
|
|
CU_ASSERT(ret == false);
|
|
|
|
to_le64(&cmd.cdw10, 0x1234567890ABCDEF);
|
|
bdev.blocklen = 520;
|
|
bdev.md_len = 8;
|
|
|
|
ret = nvmf_bdev_ctrlr_get_dif_ctx(&bdev, &cmd, &dif_ctx);
|
|
CU_ASSERT(ret == true);
|
|
CU_ASSERT(dif_ctx.block_size = 520);
|
|
CU_ASSERT(dif_ctx.md_size == 8);
|
|
CU_ASSERT(dif_ctx.init_ref_tag == 0x90ABCDEF);
|
|
}
|
|
|
|
static void
|
|
test_spdk_nvmf_bdev_ctrlr_compare_and_write_cmd(void)
|
|
{
|
|
int rc;
|
|
struct spdk_bdev bdev = {};
|
|
struct spdk_bdev_desc *desc = NULL;
|
|
struct spdk_io_channel ch = {};
|
|
|
|
struct spdk_nvmf_request cmp_req = {};
|
|
union nvmf_c2h_msg cmp_rsp = {};
|
|
|
|
struct spdk_nvmf_request write_req = {};
|
|
union nvmf_c2h_msg write_rsp = {};
|
|
|
|
struct spdk_nvmf_qpair qpair = {};
|
|
|
|
struct spdk_nvme_cmd cmp_cmd = {};
|
|
struct spdk_nvme_cmd write_cmd = {};
|
|
|
|
struct spdk_nvmf_ctrlr ctrlr = {};
|
|
struct spdk_nvmf_subsystem subsystem = {};
|
|
struct spdk_nvmf_ns ns = {};
|
|
struct spdk_nvmf_ns *subsys_ns[1] = {};
|
|
|
|
struct spdk_nvmf_poll_group group = {};
|
|
struct spdk_nvmf_subsystem_poll_group sgroups = {};
|
|
struct spdk_nvmf_subsystem_pg_ns_info ns_info = {};
|
|
|
|
bdev.blocklen = 512;
|
|
bdev.blockcnt = 10;
|
|
ns.bdev = &bdev;
|
|
|
|
subsystem.id = 0;
|
|
subsystem.max_nsid = 1;
|
|
subsys_ns[0] = &ns;
|
|
subsystem.ns = (struct spdk_nvmf_ns **)&subsys_ns;
|
|
|
|
/* Enable controller */
|
|
ctrlr.vcprop.cc.bits.en = 1;
|
|
ctrlr.subsys = &subsystem;
|
|
|
|
group.num_sgroups = 1;
|
|
sgroups.num_ns = 1;
|
|
sgroups.ns_info = &ns_info;
|
|
group.sgroups = &sgroups;
|
|
|
|
qpair.ctrlr = &ctrlr;
|
|
qpair.group = &group;
|
|
|
|
cmp_req.qpair = &qpair;
|
|
cmp_req.cmd = (union nvmf_h2c_msg *)&cmp_cmd;
|
|
cmp_req.rsp = &cmp_rsp;
|
|
|
|
cmp_cmd.nsid = 1;
|
|
cmp_cmd.fuse = SPDK_NVME_CMD_FUSE_FIRST;
|
|
cmp_cmd.opc = SPDK_NVME_OPC_COMPARE;
|
|
|
|
write_req.qpair = &qpair;
|
|
write_req.cmd = (union nvmf_h2c_msg *)&write_cmd;
|
|
write_req.rsp = &write_rsp;
|
|
|
|
write_cmd.nsid = 1;
|
|
write_cmd.fuse = SPDK_NVME_CMD_FUSE_SECOND;
|
|
write_cmd.opc = SPDK_NVME_OPC_WRITE;
|
|
|
|
/* 1. SUCCESS */
|
|
cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
cmp_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
|
|
write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen;
|
|
|
|
rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req);
|
|
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sct == 0);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sc == 0);
|
|
|
|
/* 2. Fused command start lba / num blocks mismatch */
|
|
cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
cmp_cmd.cdw12 = 2; /* NLB: CDW12 bits 15:00, 0's based */
|
|
|
|
write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen;
|
|
|
|
rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req);
|
|
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_FIELD);
|
|
|
|
/* 3. SPDK_NVME_SC_LBA_OUT_OF_RANGE */
|
|
cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
cmp_cmd.cdw12 = 100; /* NLB: CDW12 bits 15:00, 0's based */
|
|
|
|
write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
write_cmd.cdw12 = 100; /* NLB: CDW12 bits 15:00, 0's based */
|
|
write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen;
|
|
|
|
rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req);
|
|
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sc == SPDK_NVME_SC_LBA_OUT_OF_RANGE);
|
|
|
|
/* 4. SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID */
|
|
cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
cmp_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
|
|
write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen - 1;
|
|
|
|
rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req);
|
|
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0);
|
|
CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(write_rsp.nvme_cpl.status.sc == SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID);
|
|
}
|
|
|
|
static void
|
|
test_nvmf_bdev_ctrlr_identify_ns(void)
|
|
{
|
|
struct spdk_nvmf_ns ns = {};
|
|
struct spdk_nvme_ns_data nsdata = {};
|
|
struct spdk_bdev bdev = {};
|
|
uint8_t ns_g_id[16] = "abcdefgh";
|
|
uint8_t eui64[8] = "12345678";
|
|
|
|
ns.bdev = &bdev;
|
|
ns.ptpl_file = (void *)0xDEADBEEF;
|
|
memcpy(ns.opts.nguid, ns_g_id, 16);
|
|
memcpy(ns.opts.eui64, eui64, 8);
|
|
|
|
bdev.blockcnt = 10;
|
|
bdev.acwu = 1;
|
|
bdev.md_len = 512;
|
|
bdev.dif_type = SPDK_DIF_TYPE1;
|
|
bdev.blocklen = 4096;
|
|
bdev.md_interleave = 0;
|
|
bdev.optimal_io_boundary = SPDK_BDEV_IO_NUM_CHILD_IOV;
|
|
bdev.dif_is_head_of_md = true;
|
|
|
|
nvmf_bdev_ctrlr_identify_ns(&ns, &nsdata, false);
|
|
CU_ASSERT(nsdata.nsze == 10);
|
|
CU_ASSERT(nsdata.ncap == 10);
|
|
CU_ASSERT(nsdata.nuse == 10);
|
|
CU_ASSERT(nsdata.nlbaf == 0);
|
|
CU_ASSERT(nsdata.flbas.format == 0);
|
|
CU_ASSERT(nsdata.nacwu == 0);
|
|
CU_ASSERT(nsdata.lbaf[0].lbads == spdk_u32log2(4096));
|
|
CU_ASSERT(nsdata.lbaf[0].ms == 512);
|
|
CU_ASSERT(nsdata.dps.pit == SPDK_NVME_FMT_NVM_PROTECTION_DISABLE);
|
|
CU_ASSERT(nsdata.noiob == SPDK_BDEV_IO_NUM_CHILD_IOV);
|
|
CU_ASSERT(nsdata.nmic.can_share == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.persist == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_reg_only == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_reg_only == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_all_reg == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_all_reg == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.ignore_existing_key == 1);
|
|
CU_ASSERT(nsdata.flbas.extended == 1);
|
|
CU_ASSERT(nsdata.mc.extended == 1);
|
|
CU_ASSERT(nsdata.mc.pointer == 0);
|
|
CU_ASSERT(nsdata.dps.md_start == true);
|
|
CU_ASSERT(!strncmp(nsdata.nguid, ns_g_id, 16));
|
|
CU_ASSERT(!strncmp((uint8_t *)&nsdata.eui64, eui64, 8));
|
|
|
|
memset(&nsdata, 0, sizeof(nsdata));
|
|
nvmf_bdev_ctrlr_identify_ns(&ns, &nsdata, true);
|
|
CU_ASSERT(nsdata.nsze == 10);
|
|
CU_ASSERT(nsdata.ncap == 10);
|
|
CU_ASSERT(nsdata.nuse == 10);
|
|
CU_ASSERT(nsdata.nlbaf == 0);
|
|
CU_ASSERT(nsdata.flbas.format == 0);
|
|
CU_ASSERT(nsdata.nacwu == 0);
|
|
CU_ASSERT(nsdata.lbaf[0].lbads == spdk_u32log2(4096));
|
|
CU_ASSERT(nsdata.noiob == SPDK_BDEV_IO_NUM_CHILD_IOV);
|
|
CU_ASSERT(nsdata.nmic.can_share == 1);
|
|
CU_ASSERT(nsdata.lbaf[0].ms == 0);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.persist == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_reg_only == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_reg_only == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_all_reg == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_all_reg == 1);
|
|
CU_ASSERT(nsdata.nsrescap.rescap.ignore_existing_key == 1);
|
|
CU_ASSERT(!strncmp(nsdata.nguid, ns_g_id, 16));
|
|
CU_ASSERT(!strncmp((uint8_t *)&nsdata.eui64, eui64, 8));
|
|
}
|
|
|
|
static void
|
|
test_nvmf_bdev_ctrlr_zcopy_start(void)
|
|
{
|
|
int rc;
|
|
struct spdk_bdev bdev = {};
|
|
struct spdk_bdev_desc *desc = NULL;
|
|
struct spdk_io_channel ch = {};
|
|
|
|
struct spdk_nvmf_request write_req = {};
|
|
union nvmf_c2h_msg write_rsp = {};
|
|
|
|
struct spdk_nvmf_qpair qpair = {};
|
|
|
|
struct spdk_nvme_cmd write_cmd = {};
|
|
|
|
struct spdk_nvmf_ctrlr ctrlr = {};
|
|
struct spdk_nvmf_subsystem subsystem = {};
|
|
struct spdk_nvmf_ns ns = {};
|
|
struct spdk_nvmf_ns *subsys_ns[1] = {};
|
|
|
|
struct spdk_nvmf_poll_group group = {};
|
|
struct spdk_nvmf_subsystem_poll_group sgroups = {};
|
|
struct spdk_nvmf_subsystem_pg_ns_info ns_info = {};
|
|
|
|
bdev.blocklen = 512;
|
|
bdev.blockcnt = 10;
|
|
ns.bdev = &bdev;
|
|
|
|
subsystem.id = 0;
|
|
subsystem.max_nsid = 1;
|
|
subsys_ns[0] = &ns;
|
|
subsystem.ns = (struct spdk_nvmf_ns **)&subsys_ns;
|
|
|
|
/* Enable controller */
|
|
ctrlr.vcprop.cc.bits.en = 1;
|
|
ctrlr.subsys = &subsystem;
|
|
|
|
group.num_sgroups = 1;
|
|
sgroups.num_ns = 1;
|
|
sgroups.ns_info = &ns_info;
|
|
group.sgroups = &sgroups;
|
|
|
|
qpair.ctrlr = &ctrlr;
|
|
qpair.group = &group;
|
|
|
|
write_req.qpair = &qpair;
|
|
write_req.cmd = (union nvmf_h2c_msg *)&write_cmd;
|
|
write_req.rsp = &write_rsp;
|
|
|
|
write_cmd.nsid = 1;
|
|
write_cmd.opc = SPDK_NVME_OPC_WRITE;
|
|
|
|
/* 1. SUCCESS */
|
|
write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen;
|
|
|
|
rc = nvmf_bdev_ctrlr_zcopy_start(&bdev, desc, &ch, &write_req);
|
|
|
|
CU_ASSERT_EQUAL(rc, SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sct, SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sc, SPDK_NVME_SC_SUCCESS);
|
|
|
|
/* 2. SPDK_NVME_SC_LBA_OUT_OF_RANGE */
|
|
write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
write_cmd.cdw12 = 100; /* NLB: CDW12 bits 15:00, 0's based */
|
|
write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen;
|
|
|
|
rc = nvmf_bdev_ctrlr_zcopy_start(&bdev, desc, &ch, &write_req);
|
|
|
|
CU_ASSERT_EQUAL(rc, SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sct, SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sc, SPDK_NVME_SC_LBA_OUT_OF_RANGE);
|
|
|
|
/* 3. SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID */
|
|
write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen - 1;
|
|
|
|
rc = nvmf_bdev_ctrlr_zcopy_start(&bdev, desc, &ch, &write_req);
|
|
|
|
CU_ASSERT_EQUAL(rc, SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sct, SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sc, SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID);
|
|
}
|
|
|
|
static void
|
|
test_nvmf_bdev_ctrlr_cmd(void)
|
|
{
|
|
int rc;
|
|
struct spdk_bdev bdev = {};
|
|
struct spdk_io_channel ch = {};
|
|
struct spdk_nvmf_request req = {};
|
|
struct spdk_nvmf_qpair qpair = {};
|
|
union nvmf_h2c_msg cmd = {};
|
|
union nvmf_c2h_msg rsp = {};
|
|
struct spdk_nvme_scc_source_range range = {};
|
|
|
|
req.cmd = &cmd;
|
|
req.rsp = &rsp;
|
|
req.qpair = &qpair;
|
|
req.length = 4096;
|
|
bdev.blocklen = 512;
|
|
bdev.blockcnt = 3;
|
|
cmd.nvme_cmd.cdw10 = 0;
|
|
cmd.nvme_cmd.cdw12 = 2;
|
|
|
|
/* Compare status asynchronous */
|
|
rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
|
|
/* SLBA out of range */
|
|
cmd.nvme_cmd.cdw10 = 3;
|
|
|
|
rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_LBA_OUT_OF_RANGE);
|
|
|
|
/* SGL length invalid */
|
|
cmd.nvme_cmd.cdw10 = 0;
|
|
req.length = 512;
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID);
|
|
|
|
/* Device error */
|
|
req.length = 4096;
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
MOCK_SET(spdk_bdev_comparev_blocks, -1);
|
|
|
|
rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR);
|
|
|
|
/* bdev not support flush */
|
|
MOCK_SET(spdk_bdev_io_type_supported, false);
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_flush_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_SUCCESS);
|
|
|
|
/* Flush error */
|
|
MOCK_SET(spdk_bdev_io_type_supported, true);
|
|
MOCK_SET(spdk_bdev_flush_blocks, -1);
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_flush_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR);
|
|
|
|
/* Flush blocks status asynchronous */
|
|
MOCK_SET(spdk_bdev_flush_blocks, 0);
|
|
|
|
rc = nvmf_bdev_ctrlr_flush_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
MOCK_CLEAR(spdk_bdev_io_type_supported);
|
|
MOCK_CLEAR(spdk_bdev_flush_blocks);
|
|
|
|
/* Write zeroes blocks status asynchronous */
|
|
rc = nvmf_bdev_ctrlr_write_zeroes_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
|
|
/* SLBA out of range */
|
|
cmd.nvme_cmd.cdw10 = 3;
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_write_zeroes_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_LBA_OUT_OF_RANGE);
|
|
|
|
/* Write block error */
|
|
MOCK_SET(spdk_bdev_write_zeroes_blocks, -1);
|
|
cmd.nvme_cmd.cdw10 = 0;
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_write_zeroes_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR);
|
|
|
|
/* Copy blocks status asynchronous */
|
|
MOCK_SET(spdk_bdev_io_type_supported, true);
|
|
cmd.nvme_cmd.cdw10 = 1024;
|
|
cmd.nvme_cmd.cdw11 = 0;
|
|
cmd.nvme_cmd.cdw12 = 0;
|
|
cmd.nvme_cmd.cdw12_bits.copy.nr = 0;
|
|
req.length = 32;
|
|
range.slba = 512;
|
|
range.nlb = 511;
|
|
req.data = ⦥
|
|
rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
|
|
/* Copy command not supported */
|
|
MOCK_SET(spdk_bdev_io_type_supported, false);
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_OPCODE);
|
|
|
|
MOCK_SET(spdk_bdev_io_type_supported, true);
|
|
|
|
/* Unsupported number of source ranges */
|
|
cmd.nvme_cmd.cdw12_bits.copy.nr = 1;
|
|
req.length = 64;
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_COMMAND_SPECIFIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_CMD_SIZE_LIMIT_SIZE_EXCEEDED);
|
|
|
|
cmd.nvme_cmd.cdw12_bits.copy.nr = 0;
|
|
req.length = 32;
|
|
|
|
/* Unsupported source range descriptor format */
|
|
cmd.nvme_cmd.cdw12_bits.copy.df = 1;
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_FIELD);
|
|
|
|
cmd.nvme_cmd.cdw12_bits.copy.df = 0;
|
|
|
|
/* Bdev copy command failed */
|
|
MOCK_SET(spdk_bdev_copy_blocks, -1);
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
|
|
rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR);
|
|
|
|
MOCK_CLEAR(spdk_bdev_copy_blocks);
|
|
MOCK_CLEAR(spdk_bdev_io_type_supported);
|
|
}
|
|
|
|
static void
|
|
test_nvmf_bdev_ctrlr_read_write_cmd(void)
|
|
{
|
|
struct spdk_bdev bdev = {};
|
|
struct spdk_nvmf_request req = {};
|
|
union nvmf_c2h_msg rsp = {};
|
|
union nvmf_h2c_msg cmd = {};
|
|
int rc;
|
|
|
|
req.cmd = &cmd;
|
|
req.rsp = &rsp;
|
|
|
|
/* Read two blocks, block size 4096 */
|
|
cmd.nvme_cmd.cdw12 = 1;
|
|
bdev.blockcnt = 100;
|
|
bdev.blocklen = 4096;
|
|
req.length = 8192;
|
|
req.zcopy_phase = NVMF_ZCOPY_PHASE_NONE;
|
|
|
|
rc = nvmf_bdev_ctrlr_read_cmd(&bdev, NULL, NULL, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
|
|
/* Write two blocks, block size 4096 */
|
|
cmd.nvme_cmd.cdw12 = 1;
|
|
bdev.blockcnt = 100;
|
|
bdev.blocklen = 4096;
|
|
req.length = 8192;
|
|
req.zcopy_phase = NVMF_ZCOPY_PHASE_NONE;
|
|
|
|
rc = nvmf_bdev_ctrlr_write_cmd(&bdev, NULL, NULL, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
}
|
|
|
|
static void
|
|
test_nvmf_bdev_ctrlr_nvme_passthru(void)
|
|
{
|
|
int rc;
|
|
struct spdk_bdev bdev = {};
|
|
struct spdk_bdev_desc *desc = NULL;
|
|
struct spdk_io_channel ch = {};
|
|
struct spdk_nvmf_qpair qpair = {};
|
|
struct spdk_nvmf_poll_group group = {};
|
|
|
|
struct spdk_nvmf_request req = {};
|
|
union nvmf_c2h_msg rsp = {};
|
|
struct spdk_nvme_cmd cmd = {};
|
|
struct spdk_bdev_io bdev_io;
|
|
|
|
bdev.blocklen = 512;
|
|
bdev.blockcnt = 10;
|
|
|
|
qpair.group = &group;
|
|
|
|
req.qpair = &qpair;
|
|
req.cmd = (union nvmf_h2c_msg *)&cmd;
|
|
req.rsp = &rsp;
|
|
|
|
cmd.nsid = 1;
|
|
cmd.opc = 0xFF;
|
|
|
|
cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */
|
|
cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */
|
|
|
|
/* NVME_IO success */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
nvmf_bdev_ctrlr_complete_cmd(&bdev_io, true, &req);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_SUCCESS);
|
|
|
|
/* NVME_IO fail */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
g_bdev_nvme_status_sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR;
|
|
nvmf_bdev_ctrlr_complete_cmd(&bdev_io, false, &req);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR);
|
|
reset_bdev_nvme_status();
|
|
|
|
/* NVME_IO not supported */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
MOCK_SET(spdk_bdev_nvme_io_passthru, -ENOTSUP);
|
|
rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_OPCODE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.dnr == 1);
|
|
|
|
/* NVME_IO no channel - queue IO */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
MOCK_SET(spdk_bdev_nvme_io_passthru, -ENOMEM);
|
|
rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
CU_ASSERT(group.stat.pending_bdev_io == 1);
|
|
|
|
MOCK_SET(spdk_bdev_nvme_io_passthru, 0);
|
|
|
|
/* NVME_ADMIN success */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
nvmf_bdev_ctrlr_complete_admin_cmd(&bdev_io, true, &req);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_SUCCESS);
|
|
|
|
/* NVME_ADMIN fail */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
g_bdev_nvme_status_sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR;
|
|
nvmf_bdev_ctrlr_complete_admin_cmd(&bdev_io, true, &req);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR);
|
|
reset_bdev_nvme_status();
|
|
|
|
/* NVME_ADMIN not supported */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
MOCK_SET(spdk_bdev_nvme_admin_passthru, -ENOTSUP);
|
|
rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC);
|
|
CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_OPCODE);
|
|
CU_ASSERT(rsp.nvme_cpl.status.dnr == 1);
|
|
|
|
/* NVME_ADMIN no channel - queue IO */
|
|
memset(&rsp, 0, sizeof(rsp));
|
|
MOCK_SET(spdk_bdev_nvme_admin_passthru, -ENOMEM);
|
|
rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL);
|
|
CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS);
|
|
CU_ASSERT(group.stat.pending_bdev_io == 2);
|
|
|
|
MOCK_SET(spdk_bdev_nvme_admin_passthru, 0);
|
|
}
|
|
|
|
int
|
|
main(int argc, char **argv)
|
|
{
|
|
CU_pSuite suite = NULL;
|
|
unsigned int num_failures;
|
|
|
|
CU_set_error_action(CUEA_ABORT);
|
|
CU_initialize_registry();
|
|
|
|
suite = CU_add_suite("nvmf", NULL, NULL);
|
|
|
|
CU_ADD_TEST(suite, test_get_rw_params);
|
|
CU_ADD_TEST(suite, test_lba_in_range);
|
|
CU_ADD_TEST(suite, test_get_dif_ctx);
|
|
CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_identify_ns);
|
|
CU_ADD_TEST(suite, test_spdk_nvmf_bdev_ctrlr_compare_and_write_cmd);
|
|
CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_zcopy_start);
|
|
CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_cmd);
|
|
CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_read_write_cmd);
|
|
CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_nvme_passthru);
|
|
|
|
CU_basic_set_mode(CU_BRM_VERBOSE);
|
|
CU_basic_run_tests();
|
|
num_failures = CU_get_number_of_failures();
|
|
CU_cleanup_registry();
|
|
return num_failures;
|
|
}
|