bdev/raid: Simplify variable names in struct raid_base_bdev_info
raid_base_bdev_info is sufficiently descriptive as name and so names of member variables in raid_base_bdev_info can be simplified. Change-Id: I759a758ec0f3ac21de767ad379a902f1b04cc66d Signed-off-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com> Reviewed-on: https://review.gerrithub.io/422919 Chandler-Test-Pool: SPDK Automated Test System <sys_sgsw@intel.com> Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Ben Walker <benjamin.walker@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com> Reviewed-by: Kunal Sablok <kunal.sablok@intel.com>
This commit is contained in:
parent
4bd9762165
commit
c7756fe71d
@ -105,7 +105,7 @@ raid_bdev_create_cb(void *io_device, void *ctx_buf)
|
|||||||
* bdev io channel.
|
* bdev io channel.
|
||||||
*/
|
*/
|
||||||
ch->base_bdevs_io_channel[i] = spdk_bdev_get_io_channel(
|
ch->base_bdevs_io_channel[i] = spdk_bdev_get_io_channel(
|
||||||
raid_bdev->base_bdev_info[i].base_bdev_desc);
|
raid_bdev->base_bdev_info[i].desc);
|
||||||
if (!ch->base_bdevs_io_channel[i]) {
|
if (!ch->base_bdevs_io_channel[i]) {
|
||||||
for (uint32_t j = 0; j < i; j++) {
|
for (uint32_t j = 0; j < i; j++) {
|
||||||
spdk_put_io_channel(ch->base_bdevs_io_channel[j]);
|
spdk_put_io_channel(ch->base_bdevs_io_channel[j]);
|
||||||
@ -201,10 +201,10 @@ raid_bdev_free_base_bdev_resource(struct raid_bdev *raid_bdev, uint32_t base_bde
|
|||||||
|
|
||||||
info = &raid_bdev->base_bdev_info[base_bdev_slot];
|
info = &raid_bdev->base_bdev_info[base_bdev_slot];
|
||||||
|
|
||||||
spdk_bdev_module_release_bdev(info->base_bdev);
|
spdk_bdev_module_release_bdev(info->bdev);
|
||||||
spdk_bdev_close(info->base_bdev_desc);
|
spdk_bdev_close(info->desc);
|
||||||
info->base_bdev_desc = NULL;
|
info->desc = NULL;
|
||||||
info->base_bdev = NULL;
|
info->bdev = NULL;
|
||||||
|
|
||||||
assert(raid_bdev->num_base_bdevs_discovered);
|
assert(raid_bdev->num_base_bdevs_discovered);
|
||||||
raid_bdev->num_base_bdevs_discovered--;
|
raid_bdev->num_base_bdevs_discovered--;
|
||||||
@ -233,8 +233,8 @@ raid_bdev_destruct(void *ctxt)
|
|||||||
* layers. Also close the descriptors if we have started shutdown.
|
* layers. Also close the descriptors if we have started shutdown.
|
||||||
*/
|
*/
|
||||||
if (g_shutdown_started ||
|
if (g_shutdown_started ||
|
||||||
((raid_bdev->base_bdev_info[i].base_bdev_remove_scheduled == true) &&
|
((raid_bdev->base_bdev_info[i].remove_scheduled == true) &&
|
||||||
(raid_bdev->base_bdev_info[i].base_bdev != NULL))) {
|
(raid_bdev->base_bdev_info[i].bdev != NULL))) {
|
||||||
raid_bdev_free_base_bdev_resource(raid_bdev, i);
|
raid_bdev_free_base_bdev_resource(raid_bdev, i);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -325,21 +325,21 @@ raid_bdev_send_passthru(struct spdk_io_channel *ch, struct spdk_bdev_io *bdev_io
|
|||||||
raid_bdev_io = (struct raid_bdev_io *)bdev_io->driver_ctx;
|
raid_bdev_io = (struct raid_bdev_io *)bdev_io->driver_ctx;
|
||||||
raid_bdev_io->status = SPDK_BDEV_IO_STATUS_SUCCESS;
|
raid_bdev_io->status = SPDK_BDEV_IO_STATUS_SUCCESS;
|
||||||
|
|
||||||
if (raid_bdev->base_bdev_info[0].base_bdev_desc == NULL) {
|
if (raid_bdev->base_bdev_info[0].desc == NULL) {
|
||||||
SPDK_ERRLOG("base bdev desc null for pd_idx %u\n", 0);
|
SPDK_ERRLOG("base bdev desc null for pd_idx %u\n", 0);
|
||||||
assert(0);
|
assert(0);
|
||||||
}
|
}
|
||||||
raid_bdev_io->splits_pending = 0;
|
raid_bdev_io->splits_pending = 0;
|
||||||
raid_bdev_io->splits_comp_outstanding = 1;
|
raid_bdev_io->splits_comp_outstanding = 1;
|
||||||
if (bdev_io->type == SPDK_BDEV_IO_TYPE_READ) {
|
if (bdev_io->type == SPDK_BDEV_IO_TYPE_READ) {
|
||||||
ret = spdk_bdev_read_blocks(raid_bdev->base_bdev_info[0].base_bdev_desc,
|
ret = spdk_bdev_read_blocks(raid_bdev->base_bdev_info[0].desc,
|
||||||
raid_bdev_io_channel->base_bdevs_io_channel[0],
|
raid_bdev_io_channel->base_bdevs_io_channel[0],
|
||||||
bdev_io->u.bdev.iovs->iov_base,
|
bdev_io->u.bdev.iovs->iov_base,
|
||||||
bdev_io->u.bdev.offset_blocks,
|
bdev_io->u.bdev.offset_blocks,
|
||||||
bdev_io->u.bdev.num_blocks, raid_bdev_io_completion,
|
bdev_io->u.bdev.num_blocks, raid_bdev_io_completion,
|
||||||
bdev_io);
|
bdev_io);
|
||||||
} else if (bdev_io->type == SPDK_BDEV_IO_TYPE_WRITE) {
|
} else if (bdev_io->type == SPDK_BDEV_IO_TYPE_WRITE) {
|
||||||
ret = spdk_bdev_write_blocks(raid_bdev->base_bdev_info[0].base_bdev_desc,
|
ret = spdk_bdev_write_blocks(raid_bdev->base_bdev_info[0].desc,
|
||||||
raid_bdev_io_channel->base_bdevs_io_channel[0],
|
raid_bdev_io_channel->base_bdevs_io_channel[0],
|
||||||
bdev_io->u.bdev.iovs->iov_base,
|
bdev_io->u.bdev.iovs->iov_base,
|
||||||
bdev_io->u.bdev.offset_blocks,
|
bdev_io->u.bdev.offset_blocks,
|
||||||
@ -423,7 +423,7 @@ raid_bdev_submit_children(struct spdk_io_channel *ch, struct spdk_bdev_io *bdev_
|
|||||||
raid_bdev_io->splits_comp_outstanding++;
|
raid_bdev_io->splits_comp_outstanding++;
|
||||||
assert(raid_bdev_io->splits_pending);
|
assert(raid_bdev_io->splits_pending);
|
||||||
raid_bdev_io->splits_pending--;
|
raid_bdev_io->splits_pending--;
|
||||||
if (raid_bdev->base_bdev_info[pd_idx].base_bdev_desc == NULL) {
|
if (raid_bdev->base_bdev_info[pd_idx].desc == NULL) {
|
||||||
SPDK_ERRLOG("base bdev desc null for pd_idx %u\n", pd_idx);
|
SPDK_ERRLOG("base bdev desc null for pd_idx %u\n", pd_idx);
|
||||||
assert(0);
|
assert(0);
|
||||||
}
|
}
|
||||||
@ -434,13 +434,13 @@ raid_bdev_submit_children(struct spdk_io_channel *ch, struct spdk_bdev_io *bdev_
|
|||||||
* function and function callback context
|
* function and function callback context
|
||||||
*/
|
*/
|
||||||
if (bdev_io->type == SPDK_BDEV_IO_TYPE_READ) {
|
if (bdev_io->type == SPDK_BDEV_IO_TYPE_READ) {
|
||||||
ret = spdk_bdev_read_blocks(raid_bdev->base_bdev_info[pd_idx].base_bdev_desc,
|
ret = spdk_bdev_read_blocks(raid_bdev->base_bdev_info[pd_idx].desc,
|
||||||
raid_bdev_io_channel->base_bdevs_io_channel[pd_idx],
|
raid_bdev_io_channel->base_bdevs_io_channel[pd_idx],
|
||||||
buf, pd_lba, pd_blocks, raid_bdev_io_completion,
|
buf, pd_lba, pd_blocks, raid_bdev_io_completion,
|
||||||
bdev_io);
|
bdev_io);
|
||||||
|
|
||||||
} else if (bdev_io->type == SPDK_BDEV_IO_TYPE_WRITE) {
|
} else if (bdev_io->type == SPDK_BDEV_IO_TYPE_WRITE) {
|
||||||
ret = spdk_bdev_write_blocks(raid_bdev->base_bdev_info[pd_idx].base_bdev_desc,
|
ret = spdk_bdev_write_blocks(raid_bdev->base_bdev_info[pd_idx].desc,
|
||||||
raid_bdev_io_channel->base_bdevs_io_channel[pd_idx],
|
raid_bdev_io_channel->base_bdevs_io_channel[pd_idx],
|
||||||
buf, pd_lba, pd_blocks, raid_bdev_io_completion,
|
buf, pd_lba, pd_blocks, raid_bdev_io_completion,
|
||||||
bdev_io);
|
bdev_io);
|
||||||
@ -544,11 +544,11 @@ raid_bdev_io_submit_fail_process(struct raid_bdev *raid_bdev, struct spdk_bdev_i
|
|||||||
} else {
|
} else {
|
||||||
/* Queue the IO to bdev layer wait queue */
|
/* Queue the IO to bdev layer wait queue */
|
||||||
pd_idx = get_curr_base_bdev_index(raid_bdev, raid_bdev_io);
|
pd_idx = get_curr_base_bdev_index(raid_bdev, raid_bdev_io);
|
||||||
raid_bdev_io->waitq_entry.bdev = raid_bdev->base_bdev_info[pd_idx].base_bdev;
|
raid_bdev_io->waitq_entry.bdev = raid_bdev->base_bdev_info[pd_idx].bdev;
|
||||||
raid_bdev_io->waitq_entry.cb_fn = raid_bdev_waitq_io_process;
|
raid_bdev_io->waitq_entry.cb_fn = raid_bdev_waitq_io_process;
|
||||||
raid_bdev_io->waitq_entry.cb_arg = raid_bdev_io;
|
raid_bdev_io->waitq_entry.cb_arg = raid_bdev_io;
|
||||||
raid_bdev_io_channel = spdk_io_channel_get_ctx(raid_bdev_io->ch);
|
raid_bdev_io_channel = spdk_io_channel_get_ctx(raid_bdev_io->ch);
|
||||||
if (spdk_bdev_queue_io_wait(raid_bdev->base_bdev_info[pd_idx].base_bdev,
|
if (spdk_bdev_queue_io_wait(raid_bdev->base_bdev_info[pd_idx].bdev,
|
||||||
raid_bdev_io_channel->base_bdevs_io_channel[pd_idx],
|
raid_bdev_io_channel->base_bdevs_io_channel[pd_idx],
|
||||||
&raid_bdev_io->waitq_entry) != 0) {
|
&raid_bdev_io->waitq_entry) != 0) {
|
||||||
SPDK_ERRLOG("bdev io waitq error, it should not happen\n");
|
SPDK_ERRLOG("bdev io waitq error, it should not happen\n");
|
||||||
@ -764,8 +764,8 @@ raid_bdev_dump_info_json(void *ctx, struct spdk_json_write_ctx *w)
|
|||||||
spdk_json_write_name(w, "base_bdevs_list");
|
spdk_json_write_name(w, "base_bdevs_list");
|
||||||
spdk_json_write_array_begin(w);
|
spdk_json_write_array_begin(w);
|
||||||
for (uint16_t i = 0; i < raid_bdev->num_base_bdevs; i++) {
|
for (uint16_t i = 0; i < raid_bdev->num_base_bdevs; i++) {
|
||||||
if (raid_bdev->base_bdev_info[i].base_bdev) {
|
if (raid_bdev->base_bdev_info[i].bdev) {
|
||||||
spdk_json_write_string(w, raid_bdev->base_bdev_info[i].base_bdev->name);
|
spdk_json_write_string(w, raid_bdev->base_bdev_info[i].bdev->name);
|
||||||
} else {
|
} else {
|
||||||
spdk_json_write_null(w);
|
spdk_json_write_null(w);
|
||||||
}
|
}
|
||||||
@ -1260,8 +1260,8 @@ raid_bdev_alloc_base_bdev_resource(struct raid_bdev *raid_bdev, struct spdk_bdev
|
|||||||
assert(raid_bdev->state != RAID_BDEV_STATE_ONLINE);
|
assert(raid_bdev->state != RAID_BDEV_STATE_ONLINE);
|
||||||
assert(base_bdev_slot < raid_bdev->num_base_bdevs);
|
assert(base_bdev_slot < raid_bdev->num_base_bdevs);
|
||||||
|
|
||||||
raid_bdev->base_bdev_info[base_bdev_slot].base_bdev = bdev;
|
raid_bdev->base_bdev_info[base_bdev_slot].bdev = bdev;
|
||||||
raid_bdev->base_bdev_info[base_bdev_slot].base_bdev_desc = desc;
|
raid_bdev->base_bdev_info[base_bdev_slot].desc = desc;
|
||||||
raid_bdev->num_base_bdevs_discovered++;
|
raid_bdev->num_base_bdevs_discovered++;
|
||||||
assert(raid_bdev->num_base_bdevs_discovered <= raid_bdev->num_base_bdevs);
|
assert(raid_bdev->num_base_bdevs_discovered <= raid_bdev->num_base_bdevs);
|
||||||
|
|
||||||
@ -1286,16 +1286,16 @@ raid_bdev_configure(struct raid_bdev *raid_bdev)
|
|||||||
uint64_t min_blockcnt;
|
uint64_t min_blockcnt;
|
||||||
struct spdk_bdev *raid_bdev_gen;
|
struct spdk_bdev *raid_bdev_gen;
|
||||||
|
|
||||||
blocklen = raid_bdev->base_bdev_info[0].base_bdev->blocklen;
|
blocklen = raid_bdev->base_bdev_info[0].bdev->blocklen;
|
||||||
min_blockcnt = raid_bdev->base_bdev_info[0].base_bdev->blockcnt;
|
min_blockcnt = raid_bdev->base_bdev_info[0].bdev->blockcnt;
|
||||||
for (uint32_t i = 1; i < raid_bdev->num_base_bdevs; i++) {
|
for (uint32_t i = 1; i < raid_bdev->num_base_bdevs; i++) {
|
||||||
/* Calculate minimum block count from all base bdevs */
|
/* Calculate minimum block count from all base bdevs */
|
||||||
if (raid_bdev->base_bdev_info[i].base_bdev->blockcnt < min_blockcnt) {
|
if (raid_bdev->base_bdev_info[i].bdev->blockcnt < min_blockcnt) {
|
||||||
min_blockcnt = raid_bdev->base_bdev_info[i].base_bdev->blockcnt;
|
min_blockcnt = raid_bdev->base_bdev_info[i].bdev->blockcnt;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Check blocklen for all base bdevs that it should be same */
|
/* Check blocklen for all base bdevs that it should be same */
|
||||||
if (blocklen != raid_bdev->base_bdev_info[i].base_bdev->blocklen) {
|
if (blocklen != raid_bdev->base_bdev_info[i].bdev->blocklen) {
|
||||||
/*
|
/*
|
||||||
* Assumption is that all the base bdevs for any raid bdev should
|
* Assumption is that all the base bdevs for any raid bdev should
|
||||||
* have same blocklen
|
* have same blocklen
|
||||||
@ -1416,7 +1416,7 @@ raid_bdev_remove_base_bdev(void *ctx)
|
|||||||
/* Find the raid_bdev which has claimed this base_bdev */
|
/* Find the raid_bdev which has claimed this base_bdev */
|
||||||
TAILQ_FOREACH_SAFE(raid_bdev, &g_spdk_raid_bdev_list, link_global_list, next_raid_bdev) {
|
TAILQ_FOREACH_SAFE(raid_bdev, &g_spdk_raid_bdev_list, link_global_list, next_raid_bdev) {
|
||||||
for (i = 0; i < raid_bdev->num_base_bdevs; i++) {
|
for (i = 0; i < raid_bdev->num_base_bdevs; i++) {
|
||||||
if (raid_bdev->base_bdev_info[i].base_bdev == base_bdev) {
|
if (raid_bdev->base_bdev_info[i].bdev == base_bdev) {
|
||||||
found = true;
|
found = true;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
@ -1432,11 +1432,11 @@ raid_bdev_remove_base_bdev(void *ctx)
|
|||||||
}
|
}
|
||||||
|
|
||||||
assert(raid_bdev != NULL);
|
assert(raid_bdev != NULL);
|
||||||
assert(raid_bdev->base_bdev_info[i].base_bdev);
|
assert(raid_bdev->base_bdev_info[i].bdev);
|
||||||
assert(raid_bdev->base_bdev_info[i].base_bdev_desc);
|
assert(raid_bdev->base_bdev_info[i].desc);
|
||||||
raid_bdev->base_bdev_info[i].base_bdev_remove_scheduled = true;
|
raid_bdev->base_bdev_info[i].remove_scheduled = true;
|
||||||
|
|
||||||
if (raid_bdev->destruct_called == true && raid_bdev->base_bdev_info[i].base_bdev != NULL) {
|
if (raid_bdev->destruct_called == true && raid_bdev->base_bdev_info[i].bdev != NULL) {
|
||||||
/* As raid bdev is already unregistered, so cleanup should be done here itself */
|
/* As raid bdev is already unregistered, so cleanup should be done here itself */
|
||||||
raid_bdev_free_base_bdev_resource(raid_bdev, i);
|
raid_bdev_free_base_bdev_resource(raid_bdev, i);
|
||||||
if (raid_bdev->num_base_bdevs_discovered == 0) {
|
if (raid_bdev->num_base_bdevs_discovered == 0) {
|
||||||
|
@ -64,17 +64,17 @@ enum raid_bdev_state {
|
|||||||
*/
|
*/
|
||||||
struct raid_base_bdev_info {
|
struct raid_base_bdev_info {
|
||||||
/* pointer to base spdk bdev */
|
/* pointer to base spdk bdev */
|
||||||
struct spdk_bdev *base_bdev;
|
struct spdk_bdev *bdev;
|
||||||
|
|
||||||
/* pointer to base bdev descriptor opened by raid bdev */
|
/* pointer to base bdev descriptor opened by raid bdev */
|
||||||
struct spdk_bdev_desc *base_bdev_desc;
|
struct spdk_bdev_desc *desc;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* When underlying base device calls the hot plug function on drive removal,
|
* When underlying base device calls the hot plug function on drive removal,
|
||||||
* this flag will be set and later after doing some processing, base device
|
* this flag will be set and later after doing some processing, base device
|
||||||
* descriptor will be closed
|
* descriptor will be closed
|
||||||
*/
|
*/
|
||||||
bool base_bdev_remove_scheduled;
|
bool remove_scheduled;
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -266,12 +266,12 @@ check_and_remove_raid_bdev(struct raid_bdev_config *raid_bdev_config)
|
|||||||
assert(raid_bdev->state == RAID_BDEV_STATE_CONFIGURING);
|
assert(raid_bdev->state == RAID_BDEV_STATE_CONFIGURING);
|
||||||
for (uint32_t i = 0; i < raid_bdev->num_base_bdevs; i++) {
|
for (uint32_t i = 0; i < raid_bdev->num_base_bdevs; i++) {
|
||||||
assert(raid_bdev->base_bdev_info != NULL);
|
assert(raid_bdev->base_bdev_info != NULL);
|
||||||
if (raid_bdev->base_bdev_info[i].base_bdev) {
|
if (raid_bdev->base_bdev_info[i].bdev) {
|
||||||
/* Release base bdev related resources */
|
/* Release base bdev related resources */
|
||||||
spdk_bdev_module_release_bdev(raid_bdev->base_bdev_info[i].base_bdev);
|
spdk_bdev_module_release_bdev(raid_bdev->base_bdev_info[i].bdev);
|
||||||
spdk_bdev_close(raid_bdev->base_bdev_info[i].base_bdev_desc);
|
spdk_bdev_close(raid_bdev->base_bdev_info[i].desc);
|
||||||
raid_bdev->base_bdev_info[i].base_bdev_desc = NULL;
|
raid_bdev->base_bdev_info[i].desc = NULL;
|
||||||
raid_bdev->base_bdev_info[i].base_bdev = NULL;
|
raid_bdev->base_bdev_info[i].bdev = NULL;
|
||||||
assert(raid_bdev->num_base_bdevs_discovered);
|
assert(raid_bdev->num_base_bdevs_discovered);
|
||||||
raid_bdev->num_base_bdevs_discovered--;
|
raid_bdev->num_base_bdevs_discovered--;
|
||||||
}
|
}
|
||||||
|
@ -720,7 +720,7 @@ verify_io(struct spdk_bdev_io *bdev_io, uint8_t num_base_drives,
|
|||||||
CU_ASSERT(pd_lba == g_io_output[index].offset_blocks);
|
CU_ASSERT(pd_lba == g_io_output[index].offset_blocks);
|
||||||
CU_ASSERT(pd_blocks == g_io_output[index].num_blocks);
|
CU_ASSERT(pd_blocks == g_io_output[index].num_blocks);
|
||||||
CU_ASSERT(ch_ctx->base_bdevs_io_channel[pd_idx] == g_io_output[index].ch);
|
CU_ASSERT(ch_ctx->base_bdevs_io_channel[pd_idx] == g_io_output[index].ch);
|
||||||
CU_ASSERT(raid_bdev->base_bdev_info[pd_idx].base_bdev_desc == g_io_output[index].desc);
|
CU_ASSERT(raid_bdev->base_bdev_info[pd_idx].desc == g_io_output[index].desc);
|
||||||
CU_ASSERT(buf == g_io_output[index].buf);
|
CU_ASSERT(buf == g_io_output[index].buf);
|
||||||
CU_ASSERT(bdev_io->type == g_io_output[index].iotype);
|
CU_ASSERT(bdev_io->type == g_io_output[index].iotype);
|
||||||
buf += (pd_blocks << spdk_u32log2(g_block_len));
|
buf += (pd_blocks << spdk_u32log2(g_block_len));
|
||||||
@ -730,7 +730,7 @@ verify_io(struct spdk_bdev_io *bdev_io, uint8_t num_base_drives,
|
|||||||
CU_ASSERT(bdev_io->u.bdev.offset_blocks == g_io_output[0].offset_blocks);
|
CU_ASSERT(bdev_io->u.bdev.offset_blocks == g_io_output[0].offset_blocks);
|
||||||
CU_ASSERT(bdev_io->u.bdev.num_blocks == g_io_output[0].num_blocks);
|
CU_ASSERT(bdev_io->u.bdev.num_blocks == g_io_output[0].num_blocks);
|
||||||
CU_ASSERT(ch_ctx->base_bdevs_io_channel[0] == g_io_output[0].ch);
|
CU_ASSERT(ch_ctx->base_bdevs_io_channel[0] == g_io_output[0].ch);
|
||||||
CU_ASSERT(raid_bdev->base_bdev_info[0].base_bdev_desc == g_io_output[0].desc);
|
CU_ASSERT(raid_bdev->base_bdev_info[0].desc == g_io_output[0].desc);
|
||||||
CU_ASSERT(buf == g_io_output[index].buf);
|
CU_ASSERT(buf == g_io_output[index].buf);
|
||||||
}
|
}
|
||||||
CU_ASSERT(g_io_comp_status == io_status);
|
CU_ASSERT(g_io_comp_status == io_status);
|
||||||
@ -839,10 +839,10 @@ verify_raid_bdev(struct rpc_construct_raid_bdev *r, bool presence, uint32_t raid
|
|||||||
CU_ASSERT(pbdev->raid_level == r->raid_level);
|
CU_ASSERT(pbdev->raid_level == r->raid_level);
|
||||||
CU_ASSERT(pbdev->destruct_called == false);
|
CU_ASSERT(pbdev->destruct_called == false);
|
||||||
for (i = 0; i < pbdev->num_base_bdevs; i++) {
|
for (i = 0; i < pbdev->num_base_bdevs; i++) {
|
||||||
if (pbdev->base_bdev_info && pbdev->base_bdev_info[i].base_bdev) {
|
if (pbdev->base_bdev_info && pbdev->base_bdev_info[i].bdev) {
|
||||||
bdev = spdk_bdev_get_by_name(pbdev->base_bdev_info[i].base_bdev->name);
|
bdev = spdk_bdev_get_by_name(pbdev->base_bdev_info[i].bdev->name);
|
||||||
CU_ASSERT(bdev != NULL);
|
CU_ASSERT(bdev != NULL);
|
||||||
CU_ASSERT(pbdev->base_bdev_info[i].base_bdev_remove_scheduled == false);
|
CU_ASSERT(pbdev->base_bdev_info[i].remove_scheduled == false);
|
||||||
} else {
|
} else {
|
||||||
CU_ASSERT(0);
|
CU_ASSERT(0);
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user