/* SPDX-License-Identifier: BSD-3-Clause * Copyright (C) 2017 Intel Corporation. * All rights reserved. * Copyright (c) 2021 NVIDIA CORPORATION & AFFILIATES. All rights reserved. */ #include "spdk/stdinc.h" #include "spdk_cunit.h" #include "spdk_internal/mock.h" #include "thread/thread_internal.h" #include "nvmf/ctrlr_bdev.c" #include "spdk/bdev_module.h" SPDK_LOG_REGISTER_COMPONENT(nvmf) DEFINE_STUB(spdk_nvmf_request_complete, int, (struct spdk_nvmf_request *req), -1); DEFINE_STUB(spdk_bdev_get_name, const char *, (const struct spdk_bdev *bdev), "test"); DEFINE_STUB(spdk_bdev_get_physical_block_size, uint32_t, (const struct spdk_bdev *bdev), 4096); DEFINE_STUB(nvmf_ctrlr_process_admin_cmd, int, (struct spdk_nvmf_request *req), 0); DEFINE_STUB(spdk_bdev_comparev_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, struct iovec *iov, int iovcnt, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_nvme_admin_passthru, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, const struct spdk_nvme_cmd *cmd, void *buf, size_t nbytes, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_abort, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, void *bio_cb_arg, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB_V(spdk_bdev_io_get_iovec, (struct spdk_bdev_io *bdev_io, struct iovec **iovp, int *iovcntp)); DEFINE_STUB(spdk_bdev_get_write_unit_size, uint32_t, (const struct spdk_bdev *bdev), 1); uint32_t spdk_bdev_get_optimal_io_boundary(const struct spdk_bdev *bdev) { return bdev->optimal_io_boundary; } uint32_t spdk_bdev_get_md_size(const struct spdk_bdev *bdev) { return bdev->md_len; } bool spdk_bdev_is_md_interleaved(const struct spdk_bdev *bdev) { return (bdev->md_len != 0) && bdev->md_interleave; } /* We have to use the typedef in the function declaration to appease astyle. */ typedef enum spdk_dif_type spdk_dif_type_t; spdk_dif_type_t spdk_bdev_get_dif_type(const struct spdk_bdev *bdev) { if (bdev->md_len != 0) { return bdev->dif_type; } else { return SPDK_DIF_DISABLE; } } bool spdk_bdev_is_dif_head_of_md(const struct spdk_bdev *bdev) { if (spdk_bdev_get_dif_type(bdev) != SPDK_DIF_DISABLE) { return bdev->dif_is_head_of_md; } else { return false; } } uint32_t spdk_bdev_get_data_block_size(const struct spdk_bdev *bdev) { if (spdk_bdev_is_md_interleaved(bdev)) { return bdev->blocklen - bdev->md_len; } else { return bdev->blocklen; } } uint16_t spdk_bdev_get_acwu(const struct spdk_bdev *bdev) { return bdev->acwu; } uint32_t spdk_bdev_get_block_size(const struct spdk_bdev *bdev) { return bdev->blocklen; } uint64_t spdk_bdev_get_num_blocks(const struct spdk_bdev *bdev) { return bdev->blockcnt; } DEFINE_STUB(spdk_bdev_comparev_and_writev_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, struct iovec *compare_iov, int compare_iovcnt, struct iovec *write_iov, int write_iovcnt, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(nvmf_ctrlr_process_io_cmd, int, (struct spdk_nvmf_request *req), 0); DEFINE_STUB_V(spdk_bdev_io_get_nvme_fused_status, (const struct spdk_bdev_io *bdev_io, uint32_t *cdw0, int *cmp_sct, int *cmp_sc, int *wr_sct, int *wr_sc)); DEFINE_STUB(spdk_bdev_is_dif_check_enabled, bool, (const struct spdk_bdev *bdev, enum spdk_dif_check_type check_type), false); DEFINE_STUB(spdk_bdev_get_io_channel, struct spdk_io_channel *, (struct spdk_bdev_desc *desc), NULL); DEFINE_STUB(spdk_bdev_flush_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_unmap_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_io_type_supported, bool, (struct spdk_bdev *bdev, enum spdk_bdev_io_type io_type), false); DEFINE_STUB(spdk_bdev_queue_io_wait, int, (struct spdk_bdev *bdev, struct spdk_io_channel *ch, struct spdk_bdev_io_wait_entry *entry), 0); DEFINE_STUB(spdk_bdev_write_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, void *buf, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_writev_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, struct iovec *iov, int iovcnt, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_read_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, void *buf, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_readv_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, struct iovec *iov, int iovcnt, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_write_zeroes_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_nvme_io_passthru, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, const struct spdk_nvme_cmd *cmd, void *buf, size_t nbytes, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB_V(spdk_bdev_free_io, (struct spdk_bdev_io *bdev_io)); DEFINE_STUB(spdk_nvmf_subsystem_get_nqn, const char *, (const struct spdk_nvmf_subsystem *subsystem), NULL); DEFINE_STUB(spdk_bdev_zcopy_start, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, struct iovec *iov, int iovcnt, uint64_t offset_blocks, uint64_t num_blocks, bool populate, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_zcopy_end, int, (struct spdk_bdev_io *bdev_io, bool commit, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_copy_blocks, int, (struct spdk_bdev_desc *desc, struct spdk_io_channel *ch, uint64_t dst_offset_blocks, uint64_t src_offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb, void *cb_arg), 0); DEFINE_STUB(spdk_bdev_get_max_copy, uint32_t, (const struct spdk_bdev *bdev), 0); struct spdk_nvmf_ns * spdk_nvmf_subsystem_get_ns(struct spdk_nvmf_subsystem *subsystem, uint32_t nsid) { abort(); return NULL; } struct spdk_nvmf_ns * spdk_nvmf_subsystem_get_first_ns(struct spdk_nvmf_subsystem *subsystem) { abort(); return NULL; } struct spdk_nvmf_ns * spdk_nvmf_subsystem_get_next_ns(struct spdk_nvmf_subsystem *subsystem, struct spdk_nvmf_ns *prev_ns) { abort(); return NULL; } int spdk_dif_ctx_init(struct spdk_dif_ctx *ctx, uint32_t block_size, uint32_t md_size, bool md_interleave, bool dif_loc, enum spdk_dif_type dif_type, uint32_t dif_flags, uint32_t init_ref_tag, uint16_t apptag_mask, uint16_t app_tag, uint32_t data_offset, uint16_t guard_seed) { ctx->block_size = block_size; ctx->md_size = md_size; ctx->init_ref_tag = init_ref_tag; return 0; } static uint32_t g_bdev_nvme_status_cdw0; static uint32_t g_bdev_nvme_status_sct = SPDK_NVME_SCT_GENERIC; static uint32_t g_bdev_nvme_status_sc = SPDK_NVME_SC_SUCCESS; static void reset_bdev_nvme_status(void) { g_bdev_nvme_status_cdw0 = 0; g_bdev_nvme_status_sct = SPDK_NVME_SCT_GENERIC; g_bdev_nvme_status_sc = SPDK_NVME_SC_SUCCESS; } void spdk_bdev_io_get_nvme_status(const struct spdk_bdev_io *bdev_io, uint32_t *cdw0, int *sct, int *sc) { *cdw0 = g_bdev_nvme_status_cdw0; *sct = g_bdev_nvme_status_sct; *sc = g_bdev_nvme_status_sc; } static void test_get_rw_params(void) { struct spdk_nvme_cmd cmd = {0}; uint64_t lba; uint64_t count; lba = 0; count = 0; to_le64(&cmd.cdw10, 0x1234567890ABCDEF); to_le32(&cmd.cdw12, 0x9875 | SPDK_NVME_IO_FLAGS_FORCE_UNIT_ACCESS); nvmf_bdev_ctrlr_get_rw_params(&cmd, &lba, &count); CU_ASSERT(lba == 0x1234567890ABCDEF); CU_ASSERT(count == 0x9875 + 1); /* NOTE: this field is 0's based, hence the +1 */ } static void test_lba_in_range(void) { /* Trivial cases (no overflow) */ CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 0, 1) == true); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 0, 1000) == true); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 0, 1001) == false); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1, 999) == true); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1, 1000) == false); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 999, 1) == true); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1000, 1) == false); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(1000, 1001, 1) == false); /* Overflow edge cases */ CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, 0, UINT64_MAX) == true); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, 1, UINT64_MAX) == false); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, UINT64_MAX - 1, 1) == true); CU_ASSERT(nvmf_bdev_ctrlr_lba_in_range(UINT64_MAX, UINT64_MAX, 1) == false); } static void test_get_dif_ctx(void) { struct spdk_bdev bdev = {}; struct spdk_nvme_cmd cmd = {}; struct spdk_dif_ctx dif_ctx = {}; bool ret; bdev.md_len = 0; ret = nvmf_bdev_ctrlr_get_dif_ctx(&bdev, &cmd, &dif_ctx); CU_ASSERT(ret == false); to_le64(&cmd.cdw10, 0x1234567890ABCDEF); bdev.blocklen = 520; bdev.md_len = 8; ret = nvmf_bdev_ctrlr_get_dif_ctx(&bdev, &cmd, &dif_ctx); CU_ASSERT(ret == true); CU_ASSERT(dif_ctx.block_size = 520); CU_ASSERT(dif_ctx.md_size == 8); CU_ASSERT(dif_ctx.init_ref_tag == 0x90ABCDEF); } static void test_spdk_nvmf_bdev_ctrlr_compare_and_write_cmd(void) { int rc; struct spdk_bdev bdev = {}; struct spdk_bdev_desc *desc = NULL; struct spdk_io_channel ch = {}; struct spdk_nvmf_request cmp_req = {}; union nvmf_c2h_msg cmp_rsp = {}; struct spdk_nvmf_request write_req = {}; union nvmf_c2h_msg write_rsp = {}; struct spdk_nvmf_qpair qpair = {}; struct spdk_nvme_cmd cmp_cmd = {}; struct spdk_nvme_cmd write_cmd = {}; struct spdk_nvmf_ctrlr ctrlr = {}; struct spdk_nvmf_subsystem subsystem = {}; struct spdk_nvmf_ns ns = {}; struct spdk_nvmf_ns *subsys_ns[1] = {}; struct spdk_nvmf_poll_group group = {}; struct spdk_nvmf_subsystem_poll_group sgroups = {}; struct spdk_nvmf_subsystem_pg_ns_info ns_info = {}; bdev.blocklen = 512; bdev.blockcnt = 10; ns.bdev = &bdev; subsystem.id = 0; subsystem.max_nsid = 1; subsys_ns[0] = &ns; subsystem.ns = (struct spdk_nvmf_ns **)&subsys_ns; /* Enable controller */ ctrlr.vcprop.cc.bits.en = 1; ctrlr.subsys = &subsystem; group.num_sgroups = 1; sgroups.num_ns = 1; sgroups.ns_info = &ns_info; group.sgroups = &sgroups; qpair.ctrlr = &ctrlr; qpair.group = &group; cmp_req.qpair = &qpair; cmp_req.cmd = (union nvmf_h2c_msg *)&cmp_cmd; cmp_req.rsp = &cmp_rsp; cmp_cmd.nsid = 1; cmp_cmd.fuse = SPDK_NVME_CMD_FUSE_FIRST; cmp_cmd.opc = SPDK_NVME_OPC_COMPARE; write_req.qpair = &qpair; write_req.cmd = (union nvmf_h2c_msg *)&write_cmd; write_req.rsp = &write_rsp; write_cmd.nsid = 1; write_cmd.fuse = SPDK_NVME_CMD_FUSE_SECOND; write_cmd.opc = SPDK_NVME_OPC_WRITE; /* 1. SUCCESS */ cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ cmp_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen; rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0); CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0); CU_ASSERT(write_rsp.nvme_cpl.status.sct == 0); CU_ASSERT(write_rsp.nvme_cpl.status.sc == 0); /* 2. Fused command start lba / num blocks mismatch */ cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ cmp_cmd.cdw12 = 2; /* NLB: CDW12 bits 15:00, 0's based */ write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen; rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0); CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0); CU_ASSERT(write_rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(write_rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_FIELD); /* 3. SPDK_NVME_SC_LBA_OUT_OF_RANGE */ cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ cmp_cmd.cdw12 = 100; /* NLB: CDW12 bits 15:00, 0's based */ write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ write_cmd.cdw12 = 100; /* NLB: CDW12 bits 15:00, 0's based */ write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen; rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0); CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0); CU_ASSERT(write_rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(write_rsp.nvme_cpl.status.sc == SPDK_NVME_SC_LBA_OUT_OF_RANGE); /* 4. SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID */ cmp_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ cmp_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen - 1; rc = nvmf_bdev_ctrlr_compare_and_write_cmd(&bdev, desc, &ch, &cmp_req, &write_req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(cmp_rsp.nvme_cpl.status.sct == 0); CU_ASSERT(cmp_rsp.nvme_cpl.status.sc == 0); CU_ASSERT(write_rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(write_rsp.nvme_cpl.status.sc == SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID); } static void test_nvmf_bdev_ctrlr_identify_ns(void) { struct spdk_nvmf_ns ns = {}; struct spdk_nvme_ns_data nsdata = {}; struct spdk_bdev bdev = {}; uint8_t ns_g_id[16] = "abcdefgh"; uint8_t eui64[8] = "12345678"; ns.bdev = &bdev; ns.ptpl_file = (void *)0xDEADBEEF; memcpy(ns.opts.nguid, ns_g_id, 16); memcpy(ns.opts.eui64, eui64, 8); bdev.blockcnt = 10; bdev.acwu = 1; bdev.md_len = 512; bdev.dif_type = SPDK_DIF_TYPE1; bdev.blocklen = 4096; bdev.md_interleave = 0; bdev.optimal_io_boundary = SPDK_BDEV_IO_NUM_CHILD_IOV; bdev.dif_is_head_of_md = true; nvmf_bdev_ctrlr_identify_ns(&ns, &nsdata, false); CU_ASSERT(nsdata.nsze == 10); CU_ASSERT(nsdata.ncap == 10); CU_ASSERT(nsdata.nuse == 10); CU_ASSERT(nsdata.nlbaf == 0); CU_ASSERT(nsdata.flbas.format == 0); CU_ASSERT(nsdata.flbas.msb_format == 0); CU_ASSERT(nsdata.nacwu == 0); CU_ASSERT(nsdata.lbaf[0].lbads == spdk_u32log2(4096)); CU_ASSERT(nsdata.lbaf[0].ms == 512); CU_ASSERT(nsdata.dps.pit == SPDK_NVME_FMT_NVM_PROTECTION_DISABLE); CU_ASSERT(nsdata.noiob == SPDK_BDEV_IO_NUM_CHILD_IOV); CU_ASSERT(nsdata.nmic.can_share == 1); CU_ASSERT(nsdata.nsrescap.rescap.persist == 1); CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive == 1); CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access == 1); CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_reg_only == 1); CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_reg_only == 1); CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_all_reg == 1); CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_all_reg == 1); CU_ASSERT(nsdata.nsrescap.rescap.ignore_existing_key == 1); CU_ASSERT(nsdata.flbas.extended == 1); CU_ASSERT(nsdata.mc.extended == 1); CU_ASSERT(nsdata.mc.pointer == 0); CU_ASSERT(nsdata.dps.md_start == true); CU_ASSERT(!strncmp(nsdata.nguid, ns_g_id, 16)); CU_ASSERT(!strncmp((uint8_t *)&nsdata.eui64, eui64, 8)); memset(&nsdata, 0, sizeof(nsdata)); nvmf_bdev_ctrlr_identify_ns(&ns, &nsdata, true); CU_ASSERT(nsdata.nsze == 10); CU_ASSERT(nsdata.ncap == 10); CU_ASSERT(nsdata.nuse == 10); CU_ASSERT(nsdata.nlbaf == 0); CU_ASSERT(nsdata.flbas.format == 0); CU_ASSERT(nsdata.flbas.msb_format == 0); CU_ASSERT(nsdata.nacwu == 0); CU_ASSERT(nsdata.lbaf[0].lbads == spdk_u32log2(4096)); CU_ASSERT(nsdata.noiob == SPDK_BDEV_IO_NUM_CHILD_IOV); CU_ASSERT(nsdata.nmic.can_share == 1); CU_ASSERT(nsdata.lbaf[0].ms == 0); CU_ASSERT(nsdata.nsrescap.rescap.persist == 1); CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive == 1); CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access == 1); CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_reg_only == 1); CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_reg_only == 1); CU_ASSERT(nsdata.nsrescap.rescap.write_exclusive_all_reg == 1); CU_ASSERT(nsdata.nsrescap.rescap.exclusive_access_all_reg == 1); CU_ASSERT(nsdata.nsrescap.rescap.ignore_existing_key == 1); CU_ASSERT(!strncmp(nsdata.nguid, ns_g_id, 16)); CU_ASSERT(!strncmp((uint8_t *)&nsdata.eui64, eui64, 8)); } static void test_nvmf_bdev_ctrlr_zcopy_start(void) { int rc; struct spdk_bdev bdev = {}; struct spdk_bdev_desc *desc = NULL; struct spdk_io_channel ch = {}; struct spdk_nvmf_request write_req = {}; union nvmf_c2h_msg write_rsp = {}; struct spdk_nvmf_qpair qpair = {}; struct spdk_nvme_cmd write_cmd = {}; struct spdk_nvmf_ctrlr ctrlr = {}; struct spdk_nvmf_subsystem subsystem = {}; struct spdk_nvmf_ns ns = {}; struct spdk_nvmf_ns *subsys_ns[1] = {}; struct spdk_nvmf_poll_group group = {}; struct spdk_nvmf_subsystem_poll_group sgroups = {}; struct spdk_nvmf_subsystem_pg_ns_info ns_info = {}; bdev.blocklen = 512; bdev.blockcnt = 10; ns.bdev = &bdev; subsystem.id = 0; subsystem.max_nsid = 1; subsys_ns[0] = &ns; subsystem.ns = (struct spdk_nvmf_ns **)&subsys_ns; /* Enable controller */ ctrlr.vcprop.cc.bits.en = 1; ctrlr.subsys = &subsystem; group.num_sgroups = 1; sgroups.num_ns = 1; sgroups.ns_info = &ns_info; group.sgroups = &sgroups; qpair.ctrlr = &ctrlr; qpair.group = &group; write_req.qpair = &qpair; write_req.cmd = (union nvmf_h2c_msg *)&write_cmd; write_req.rsp = &write_rsp; write_cmd.nsid = 1; write_cmd.opc = SPDK_NVME_OPC_WRITE; /* 1. SUCCESS */ write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen; rc = nvmf_bdev_ctrlr_zcopy_start(&bdev, desc, &ch, &write_req); CU_ASSERT_EQUAL(rc, SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sct, SPDK_NVME_SCT_GENERIC); CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sc, SPDK_NVME_SC_SUCCESS); /* 2. SPDK_NVME_SC_LBA_OUT_OF_RANGE */ write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ write_cmd.cdw12 = 100; /* NLB: CDW12 bits 15:00, 0's based */ write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen; rc = nvmf_bdev_ctrlr_zcopy_start(&bdev, desc, &ch, &write_req); CU_ASSERT_EQUAL(rc, SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sct, SPDK_NVME_SCT_GENERIC); CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sc, SPDK_NVME_SC_LBA_OUT_OF_RANGE); /* 3. SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID */ write_cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ write_cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ write_req.length = (write_cmd.cdw12 + 1) * bdev.blocklen - 1; rc = nvmf_bdev_ctrlr_zcopy_start(&bdev, desc, &ch, &write_req); CU_ASSERT_EQUAL(rc, SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sct, SPDK_NVME_SCT_GENERIC); CU_ASSERT_EQUAL(write_rsp.nvme_cpl.status.sc, SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID); } static void test_nvmf_bdev_ctrlr_cmd(void) { int rc; struct spdk_bdev bdev = {}; struct spdk_io_channel ch = {}; struct spdk_nvmf_request req = {}; struct spdk_nvmf_qpair qpair = {}; union nvmf_h2c_msg cmd = {}; union nvmf_c2h_msg rsp = {}; struct spdk_nvme_scc_source_range range = {}; req.cmd = &cmd; req.rsp = &rsp; req.qpair = &qpair; req.length = 4096; bdev.blocklen = 512; bdev.blockcnt = 3; cmd.nvme_cmd.cdw10 = 0; cmd.nvme_cmd.cdw12 = 2; /* Compare status asynchronous */ rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); /* SLBA out of range */ cmd.nvme_cmd.cdw10 = 3; rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_LBA_OUT_OF_RANGE); /* SGL length invalid */ cmd.nvme_cmd.cdw10 = 0; req.length = 512; memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_DATA_SGL_LENGTH_INVALID); /* Device error */ req.length = 4096; memset(&rsp, 0, sizeof(rsp)); MOCK_SET(spdk_bdev_comparev_blocks, -1); rc = nvmf_bdev_ctrlr_compare_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR); /* bdev not support flush */ MOCK_SET(spdk_bdev_io_type_supported, false); memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_flush_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_SUCCESS); /* Flush error */ MOCK_SET(spdk_bdev_io_type_supported, true); MOCK_SET(spdk_bdev_flush_blocks, -1); memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_flush_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR); /* Flush blocks status asynchronous */ MOCK_SET(spdk_bdev_flush_blocks, 0); rc = nvmf_bdev_ctrlr_flush_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); MOCK_CLEAR(spdk_bdev_io_type_supported); MOCK_CLEAR(spdk_bdev_flush_blocks); /* Write zeroes blocks status asynchronous */ rc = nvmf_bdev_ctrlr_write_zeroes_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); /* SLBA out of range */ cmd.nvme_cmd.cdw10 = 3; memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_write_zeroes_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_LBA_OUT_OF_RANGE); /* Write block error */ MOCK_SET(spdk_bdev_write_zeroes_blocks, -1); cmd.nvme_cmd.cdw10 = 0; memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_write_zeroes_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR); /* Copy blocks status asynchronous */ MOCK_SET(spdk_bdev_io_type_supported, true); cmd.nvme_cmd.cdw10 = 1024; cmd.nvme_cmd.cdw11 = 0; cmd.nvme_cmd.cdw12 = 0; cmd.nvme_cmd.cdw12_bits.copy.nr = 0; range.slba = 512; range.nlb = 511; req.data = ⦥ req.length = 32; spdk_iov_one(req.iov, &req.iovcnt, &range, req.length); rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); /* Copy command not supported */ MOCK_SET(spdk_bdev_io_type_supported, false); memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); MOCK_SET(spdk_bdev_io_type_supported, true); /* Unsupported number of source ranges */ cmd.nvme_cmd.cdw12_bits.copy.nr = 1; req.length = 64; memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_COMMAND_SPECIFIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_CMD_SIZE_LIMIT_SIZE_EXCEEDED); cmd.nvme_cmd.cdw12_bits.copy.nr = 0; req.length = 32; /* Unsupported source range descriptor format */ cmd.nvme_cmd.cdw12_bits.copy.df = 1; memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_FIELD); cmd.nvme_cmd.cdw12_bits.copy.df = 0; /* Bdev copy command failed */ MOCK_SET(spdk_bdev_copy_blocks, -1); memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_copy_cmd(&bdev, NULL, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR); MOCK_CLEAR(spdk_bdev_copy_blocks); MOCK_CLEAR(spdk_bdev_io_type_supported); } static void test_nvmf_bdev_ctrlr_read_write_cmd(void) { struct spdk_bdev bdev = {}; struct spdk_nvmf_request req = {}; union nvmf_c2h_msg rsp = {}; union nvmf_h2c_msg cmd = {}; int rc; req.cmd = &cmd; req.rsp = &rsp; /* Read two blocks, block size 4096 */ cmd.nvme_cmd.cdw12 = 1; bdev.blockcnt = 100; bdev.blocklen = 4096; req.length = 8192; req.zcopy_phase = NVMF_ZCOPY_PHASE_NONE; rc = nvmf_bdev_ctrlr_read_cmd(&bdev, NULL, NULL, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); /* Write two blocks, block size 4096 */ cmd.nvme_cmd.cdw12 = 1; bdev.blockcnt = 100; bdev.blocklen = 4096; req.length = 8192; req.zcopy_phase = NVMF_ZCOPY_PHASE_NONE; rc = nvmf_bdev_ctrlr_write_cmd(&bdev, NULL, NULL, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); } static void test_nvmf_bdev_ctrlr_nvme_passthru(void) { int rc; struct spdk_bdev bdev = {}; struct spdk_bdev_desc *desc = NULL; struct spdk_io_channel ch = {}; struct spdk_nvmf_qpair qpair = {}; struct spdk_nvmf_poll_group group = {}; struct spdk_nvmf_request req = {}; union nvmf_c2h_msg rsp = {}; struct spdk_nvme_cmd cmd = {}; struct spdk_bdev_io bdev_io; bdev.blocklen = 512; bdev.blockcnt = 10; qpair.group = &group; req.qpair = &qpair; req.cmd = (union nvmf_h2c_msg *)&cmd; req.rsp = &rsp; spdk_iov_one(req.iov, &req.iovcnt, NULL, 0); cmd.nsid = 1; cmd.opc = 0xFF; cmd.cdw10 = 1; /* SLBA: CDW10 and CDW11 */ cmd.cdw12 = 1; /* NLB: CDW12 bits 15:00, 0's based */ /* NVME_IO success */ memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); nvmf_bdev_ctrlr_complete_cmd(&bdev_io, true, &req); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_SUCCESS); /* NVME_IO fail */ memset(&rsp, 0, sizeof(rsp)); rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); g_bdev_nvme_status_sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR; nvmf_bdev_ctrlr_complete_cmd(&bdev_io, false, &req); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR); reset_bdev_nvme_status(); /* NVME_IO not supported */ memset(&rsp, 0, sizeof(rsp)); MOCK_SET(spdk_bdev_nvme_io_passthru, -ENOTSUP); rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_OPCODE); CU_ASSERT(rsp.nvme_cpl.status.dnr == 1); /* NVME_IO no channel - queue IO */ memset(&rsp, 0, sizeof(rsp)); MOCK_SET(spdk_bdev_nvme_io_passthru, -ENOMEM); rc = nvmf_bdev_ctrlr_nvme_passthru_io(&bdev, desc, &ch, &req); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); CU_ASSERT(group.stat.pending_bdev_io == 1); MOCK_SET(spdk_bdev_nvme_io_passthru, 0); /* NVME_ADMIN success */ memset(&rsp, 0, sizeof(rsp)); rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); nvmf_bdev_ctrlr_complete_admin_cmd(&bdev_io, true, &req); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_SUCCESS); /* NVME_ADMIN fail */ memset(&rsp, 0, sizeof(rsp)); rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); g_bdev_nvme_status_sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR; nvmf_bdev_ctrlr_complete_admin_cmd(&bdev_io, true, &req); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INTERNAL_DEVICE_ERROR); reset_bdev_nvme_status(); /* NVME_ADMIN not supported */ memset(&rsp, 0, sizeof(rsp)); MOCK_SET(spdk_bdev_nvme_admin_passthru, -ENOTSUP); rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE); CU_ASSERT(rsp.nvme_cpl.status.sct == SPDK_NVME_SCT_GENERIC); CU_ASSERT(rsp.nvme_cpl.status.sc == SPDK_NVME_SC_INVALID_OPCODE); CU_ASSERT(rsp.nvme_cpl.status.dnr == 1); /* NVME_ADMIN no channel - queue IO */ memset(&rsp, 0, sizeof(rsp)); MOCK_SET(spdk_bdev_nvme_admin_passthru, -ENOMEM); rc = spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(&bdev, desc, &ch, &req, NULL); CU_ASSERT(rc == SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS); CU_ASSERT(group.stat.pending_bdev_io == 2); MOCK_SET(spdk_bdev_nvme_admin_passthru, 0); } int main(int argc, char **argv) { CU_pSuite suite = NULL; unsigned int num_failures; CU_set_error_action(CUEA_ABORT); CU_initialize_registry(); suite = CU_add_suite("nvmf", NULL, NULL); CU_ADD_TEST(suite, test_get_rw_params); CU_ADD_TEST(suite, test_lba_in_range); CU_ADD_TEST(suite, test_get_dif_ctx); CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_identify_ns); CU_ADD_TEST(suite, test_spdk_nvmf_bdev_ctrlr_compare_and_write_cmd); CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_zcopy_start); CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_cmd); CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_read_write_cmd); CU_ADD_TEST(suite, test_nvmf_bdev_ctrlr_nvme_passthru); CU_basic_set_mode(CU_BRM_VERBOSE); CU_basic_run_tests(); num_failures = CU_get_number_of_failures(); CU_cleanup_registry(); return num_failures; }