nvme: use nvme_robust_mutex related operations

Change-Id: I35416506dbafe5e9d21861e207e295e114bdb3db
Signed-off-by: GangCao <gang.cao@intel.com>
This commit is contained in:
GangCao 2016-12-05 15:54:16 -05:00 committed by Daniel Verkamp
parent 47341b89b7
commit 6bdcf5abe8
5 changed files with 88 additions and 88 deletions

View File

@ -59,7 +59,7 @@ nvme_attach(enum spdk_nvme_transport_type trtype,
int int
spdk_nvme_detach(struct spdk_nvme_ctrlr *ctrlr) spdk_nvme_detach(struct spdk_nvme_ctrlr *ctrlr)
{ {
pthread_mutex_lock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_lock(&g_spdk_nvme_driver->lock);
nvme_ctrlr_proc_put_ref(ctrlr); nvme_ctrlr_proc_put_ref(ctrlr);
@ -68,7 +68,7 @@ spdk_nvme_detach(struct spdk_nvme_ctrlr *ctrlr)
nvme_ctrlr_destruct(ctrlr); nvme_ctrlr_destruct(ctrlr);
} }
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
return 0; return 0;
} }
@ -289,7 +289,7 @@ nvme_driver_init(void)
return ret; return ret;
} }
pthread_mutex_lock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_lock(&g_spdk_nvme_driver->lock);
g_spdk_nvme_driver->initialized = false; g_spdk_nvme_driver->initialized = false;
@ -301,7 +301,7 @@ nvme_driver_init(void)
if (g_spdk_nvme_driver->request_mempool == NULL) { if (g_spdk_nvme_driver->request_mempool == NULL) {
SPDK_ERRLOG("unable to allocate pool of requests\n"); SPDK_ERRLOG("unable to allocate pool of requests\n");
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
pthread_mutex_destroy(&g_spdk_nvme_driver->lock); pthread_mutex_destroy(&g_spdk_nvme_driver->lock);
spdk_memzone_free(SPDK_NVME_DRIVER_NAME); spdk_memzone_free(SPDK_NVME_DRIVER_NAME);
@ -309,7 +309,7 @@ nvme_driver_init(void)
return -1; return -1;
} }
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
return ret; return ret;
} }
@ -343,7 +343,7 @@ nvme_init_controllers(void *cb_ctx, spdk_nvme_attach_cb attach_cb)
int start_rc; int start_rc;
struct spdk_nvme_ctrlr *ctrlr, *ctrlr_tmp; struct spdk_nvme_ctrlr *ctrlr, *ctrlr_tmp;
pthread_mutex_lock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_lock(&g_spdk_nvme_driver->lock);
/* Initialize all new controllers in the init_ctrlrs list in parallel. */ /* Initialize all new controllers in the init_ctrlrs list in parallel. */
while (!TAILQ_EMPTY(&g_spdk_nvme_driver->init_ctrlrs)) { while (!TAILQ_EMPTY(&g_spdk_nvme_driver->init_ctrlrs)) {
@ -356,9 +356,9 @@ nvme_init_controllers(void *cb_ctx, spdk_nvme_attach_cb attach_cb)
* the functions it calls (in particular nvme_ctrlr_set_num_qpairs()) * the functions it calls (in particular nvme_ctrlr_set_num_qpairs())
* can assume it is held. * can assume it is held.
*/ */
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
start_rc = nvme_ctrlr_process_init(ctrlr); start_rc = nvme_ctrlr_process_init(ctrlr);
pthread_mutex_lock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_lock(&g_spdk_nvme_driver->lock);
if (start_rc) { if (start_rc) {
/* Controller failed to initialize. */ /* Controller failed to initialize. */
@ -386,9 +386,9 @@ nvme_init_controllers(void *cb_ctx, spdk_nvme_attach_cb attach_cb)
* Unlock while calling attach_cb() so the user can call other functions * Unlock while calling attach_cb() so the user can call other functions
* that may take the driver lock, like nvme_detach(). * that may take the driver lock, like nvme_detach().
*/ */
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
attach_cb(cb_ctx, &ctrlr->probe_info, ctrlr, &ctrlr->opts); attach_cb(cb_ctx, &ctrlr->probe_info, ctrlr, &ctrlr->opts);
pthread_mutex_lock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_lock(&g_spdk_nvme_driver->lock);
break; break;
} }
@ -397,7 +397,7 @@ nvme_init_controllers(void *cb_ctx, spdk_nvme_attach_cb attach_cb)
g_spdk_nvme_driver->initialized = true; g_spdk_nvme_driver->initialized = true;
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
return rc; return rc;
} }
@ -423,7 +423,7 @@ _spdk_nvme_probe(const struct spdk_nvme_discover_info *info, void *cb_ctx,
return rc; return rc;
} }
pthread_mutex_lock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_lock(&g_spdk_nvme_driver->lock);
if (hotplug_fd < 0) { if (hotplug_fd < 0) {
hotplug_fd = spdk_uevent_connect(); hotplug_fd = spdk_uevent_connect();
@ -437,7 +437,7 @@ _spdk_nvme_probe(const struct spdk_nvme_discover_info *info, void *cb_ctx,
} else { } else {
if (!spdk_nvme_transport_available(info->trtype)) { if (!spdk_nvme_transport_available(info->trtype)) {
SPDK_ERRLOG("NVMe over Fabrics trtype %u not available\n", info->trtype); SPDK_ERRLOG("NVMe over Fabrics trtype %u not available\n", info->trtype);
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
return -1; return -1;
} }
@ -454,16 +454,16 @@ _spdk_nvme_probe(const struct spdk_nvme_discover_info *info, void *cb_ctx,
* Unlock while calling attach_cb() so the user can call other functions * Unlock while calling attach_cb() so the user can call other functions
* that may take the driver lock, like nvme_detach(). * that may take the driver lock, like nvme_detach().
*/ */
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
attach_cb(cb_ctx, &ctrlr->probe_info, ctrlr, &ctrlr->opts); attach_cb(cb_ctx, &ctrlr->probe_info, ctrlr, &ctrlr->opts);
pthread_mutex_lock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_lock(&g_spdk_nvme_driver->lock);
} }
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
return 0; return 0;
} }
pthread_mutex_unlock(&g_spdk_nvme_driver->lock); nvme_robust_mutex_unlock(&g_spdk_nvme_driver->lock);
/* /*
* Keep going even if one or more nvme_attach() calls failed, * Keep going even if one or more nvme_attach() calls failed,
* but maintain the value of rc to signal errors when we return. * but maintain the value of rc to signal errors when we return.

View File

@ -151,7 +151,7 @@ spdk_nvme_ctrlr_alloc_io_qpair(struct spdk_nvme_ctrlr *ctrlr,
return NULL; return NULL;
} }
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
/* /*
* Get the first available I/O queue ID. * Get the first available I/O queue ID.
@ -159,14 +159,14 @@ spdk_nvme_ctrlr_alloc_io_qpair(struct spdk_nvme_ctrlr *ctrlr,
qid = spdk_bit_array_find_first_set(ctrlr->free_io_qids, 1); qid = spdk_bit_array_find_first_set(ctrlr->free_io_qids, 1);
if (qid > ctrlr->opts.num_io_queues) { if (qid > ctrlr->opts.num_io_queues) {
SPDK_ERRLOG("No free I/O queue IDs\n"); SPDK_ERRLOG("No free I/O queue IDs\n");
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return NULL; return NULL;
} }
qpair = nvme_transport_ctrlr_create_io_qpair(ctrlr, qid, qprio); qpair = nvme_transport_ctrlr_create_io_qpair(ctrlr, qid, qprio);
if (qpair == NULL) { if (qpair == NULL) {
SPDK_ERRLOG("transport->ctrlr_create_io_qpair() failed\n"); SPDK_ERRLOG("transport->ctrlr_create_io_qpair() failed\n");
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return NULL; return NULL;
} }
spdk_bit_array_clear(ctrlr->free_io_qids, qid); spdk_bit_array_clear(ctrlr->free_io_qids, qid);
@ -174,7 +174,7 @@ spdk_nvme_ctrlr_alloc_io_qpair(struct spdk_nvme_ctrlr *ctrlr,
nvme_ctrlr_proc_add_io_qpair(qpair); nvme_ctrlr_proc_add_io_qpair(qpair);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return qpair; return qpair;
} }
@ -190,7 +190,7 @@ spdk_nvme_ctrlr_free_io_qpair(struct spdk_nvme_qpair *qpair)
ctrlr = qpair->ctrlr; ctrlr = qpair->ctrlr;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
nvme_ctrlr_proc_remove_io_qpair(qpair); nvme_ctrlr_proc_remove_io_qpair(qpair);
@ -198,11 +198,11 @@ spdk_nvme_ctrlr_free_io_qpair(struct spdk_nvme_qpair *qpair)
spdk_bit_array_set(ctrlr->free_io_qids, qpair->id); spdk_bit_array_set(ctrlr->free_io_qids, qpair->id);
if (nvme_transport_ctrlr_delete_io_qpair(ctrlr, qpair)) { if (nvme_transport_ctrlr_delete_io_qpair(ctrlr, qpair)) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -1; return -1;
} }
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return 0; return 0;
} }
@ -495,7 +495,7 @@ spdk_nvme_ctrlr_reset(struct spdk_nvme_ctrlr *ctrlr)
int rc = 0; int rc = 0;
struct spdk_nvme_qpair *qpair; struct spdk_nvme_qpair *qpair;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
if (ctrlr->is_resetting || ctrlr->is_failed) { if (ctrlr->is_resetting || ctrlr->is_failed) {
/* /*
@ -503,7 +503,7 @@ spdk_nvme_ctrlr_reset(struct spdk_nvme_ctrlr *ctrlr)
* immediately since there is no need to kick off another * immediately since there is no need to kick off another
* reset in these cases. * reset in these cases.
*/ */
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return 0; return 0;
} }
@ -541,7 +541,7 @@ spdk_nvme_ctrlr_reset(struct spdk_nvme_ctrlr *ctrlr)
ctrlr->is_resetting = false; ctrlr->is_resetting = false;
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -990,7 +990,7 @@ nvme_ctrlr_proc_get_ref(struct spdk_nvme_ctrlr *ctrlr)
struct spdk_nvme_ctrlr_process *active_proc; struct spdk_nvme_ctrlr_process *active_proc;
pid_t pid = getpid(); pid_t pid = getpid();
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
nvme_ctrlr_remove_inactive_proc(ctrlr); nvme_ctrlr_remove_inactive_proc(ctrlr);
@ -1001,7 +1001,7 @@ nvme_ctrlr_proc_get_ref(struct spdk_nvme_ctrlr *ctrlr)
} }
} }
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
void void
@ -1011,7 +1011,7 @@ nvme_ctrlr_proc_put_ref(struct spdk_nvme_ctrlr *ctrlr)
pid_t pid = getpid(); pid_t pid = getpid();
int proc_count; int proc_count;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
proc_count = nvme_ctrlr_remove_inactive_proc(ctrlr); proc_count = nvme_ctrlr_remove_inactive_proc(ctrlr);
@ -1032,7 +1032,7 @@ nvme_ctrlr_proc_put_ref(struct spdk_nvme_ctrlr *ctrlr)
} }
} }
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
int int
@ -1041,7 +1041,7 @@ nvme_ctrlr_get_ref_count(struct spdk_nvme_ctrlr *ctrlr)
struct spdk_nvme_ctrlr_process *active_proc; struct spdk_nvme_ctrlr_process *active_proc;
int ref = 0; int ref = 0;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
nvme_ctrlr_remove_inactive_proc(ctrlr); nvme_ctrlr_remove_inactive_proc(ctrlr);
@ -1049,7 +1049,7 @@ nvme_ctrlr_get_ref_count(struct spdk_nvme_ctrlr *ctrlr)
ref += active_proc->ref; ref += active_proc->ref;
} }
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return ref; return ref;
} }
@ -1362,12 +1362,12 @@ spdk_nvme_ctrlr_process_admin_completions(struct spdk_nvme_ctrlr *ctrlr)
{ {
int32_t num_completions; int32_t num_completions;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
if (ctrlr->keep_alive_interval_ticks) { if (ctrlr->keep_alive_interval_ticks) {
nvme_ctrlr_keep_alive(ctrlr); nvme_ctrlr_keep_alive(ctrlr);
} }
num_completions = spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); num_completions = spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return num_completions; return num_completions;
} }
@ -1447,9 +1447,9 @@ spdk_nvme_ctrlr_attach_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
if (res) if (res)
return res; return res;
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
SPDK_ERRLOG("spdk_nvme_ctrlr_attach_ns failed!\n"); SPDK_ERRLOG("spdk_nvme_ctrlr_attach_ns failed!\n");
@ -1472,9 +1472,9 @@ spdk_nvme_ctrlr_detach_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
if (res) if (res)
return res; return res;
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
SPDK_ERRLOG("spdk_nvme_ctrlr_detach_ns failed!\n"); SPDK_ERRLOG("spdk_nvme_ctrlr_detach_ns failed!\n");
@ -1495,9 +1495,9 @@ spdk_nvme_ctrlr_create_ns(struct spdk_nvme_ctrlr *ctrlr, struct spdk_nvme_ns_dat
if (res) if (res)
return 0; return 0;
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
SPDK_ERRLOG("spdk_nvme_ctrlr_create_ns failed!\n"); SPDK_ERRLOG("spdk_nvme_ctrlr_create_ns failed!\n");
@ -1524,9 +1524,9 @@ spdk_nvme_ctrlr_delete_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid)
if (res) if (res)
return res; return res;
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
SPDK_ERRLOG("spdk_nvme_ctrlr_delete_ns failed!\n"); SPDK_ERRLOG("spdk_nvme_ctrlr_delete_ns failed!\n");
@ -1549,9 +1549,9 @@ spdk_nvme_ctrlr_format(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
if (res) if (res)
return res; return res;
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
SPDK_ERRLOG("spdk_nvme_ctrlr_format failed!\n"); SPDK_ERRLOG("spdk_nvme_ctrlr_format failed!\n");
@ -1594,9 +1594,9 @@ spdk_nvme_ctrlr_update_firmware(struct spdk_nvme_ctrlr *ctrlr, void *payload, ui
return res; return res;
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
SPDK_ERRLOG("spdk_nvme_ctrlr_fw_image_download failed!\n"); SPDK_ERRLOG("spdk_nvme_ctrlr_fw_image_download failed!\n");
@ -1620,9 +1620,9 @@ spdk_nvme_ctrlr_update_firmware(struct spdk_nvme_ctrlr *ctrlr, void *payload, ui
return res; return res;
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ctrlr->adminq, 0);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
SPDK_ERRLOG("nvme_ctrlr_cmd_fw_commit failed!\n"); SPDK_ERRLOG("nvme_ctrlr_cmd_fw_commit failed!\n");

View File

@ -62,10 +62,10 @@ spdk_nvme_ctrlr_cmd_admin_raw(struct spdk_nvme_ctrlr *ctrlr,
struct nvme_request *req; struct nvme_request *req;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_contig(buf, len, cb_fn, cb_arg); req = nvme_allocate_request_contig(buf, len, cb_fn, cb_arg);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -73,7 +73,7 @@ spdk_nvme_ctrlr_cmd_admin_raw(struct spdk_nvme_ctrlr *ctrlr,
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -135,11 +135,11 @@ nvme_ctrlr_cmd_attach_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_user_copy(payload, sizeof(struct spdk_nvme_ctrlr_list), req = nvme_allocate_request_user_copy(payload, sizeof(struct spdk_nvme_ctrlr_list),
cb_fn, cb_arg, true); cb_fn, cb_arg, true);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -150,7 +150,7 @@ nvme_ctrlr_cmd_attach_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -162,11 +162,11 @@ nvme_ctrlr_cmd_detach_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_user_copy(payload, sizeof(struct spdk_nvme_ctrlr_list), req = nvme_allocate_request_user_copy(payload, sizeof(struct spdk_nvme_ctrlr_list),
cb_fn, cb_arg, true); cb_fn, cb_arg, true);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -177,7 +177,7 @@ nvme_ctrlr_cmd_detach_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid,
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -189,11 +189,11 @@ nvme_ctrlr_cmd_create_ns(struct spdk_nvme_ctrlr *ctrlr, struct spdk_nvme_ns_data
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_user_copy(payload, sizeof(struct spdk_nvme_ns_data), req = nvme_allocate_request_user_copy(payload, sizeof(struct spdk_nvme_ns_data),
cb_fn, cb_arg, true); cb_fn, cb_arg, true);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -203,7 +203,7 @@ nvme_ctrlr_cmd_create_ns(struct spdk_nvme_ctrlr *ctrlr, struct spdk_nvme_ns_data
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -215,10 +215,10 @@ nvme_ctrlr_cmd_delete_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid, spdk_nvme
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_null(cb_fn, cb_arg); req = nvme_allocate_request_null(cb_fn, cb_arg);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -229,7 +229,7 @@ nvme_ctrlr_cmd_delete_ns(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid, spdk_nvme
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -241,10 +241,10 @@ nvme_ctrlr_cmd_format(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid, struct spdk_
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_null(cb_fn, cb_arg); req = nvme_allocate_request_null(cb_fn, cb_arg);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -254,7 +254,7 @@ nvme_ctrlr_cmd_format(struct spdk_nvme_ctrlr *ctrlr, uint32_t nsid, struct spdk_
memcpy(&cmd->cdw10, format, sizeof(uint32_t)); memcpy(&cmd->cdw10, format, sizeof(uint32_t));
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -268,10 +268,10 @@ spdk_nvme_ctrlr_cmd_set_feature(struct spdk_nvme_ctrlr *ctrlr, uint8_t feature,
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_null(cb_fn, cb_arg); req = nvme_allocate_request_null(cb_fn, cb_arg);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -282,7 +282,7 @@ spdk_nvme_ctrlr_cmd_set_feature(struct spdk_nvme_ctrlr *ctrlr, uint8_t feature,
cmd->cdw12 = cdw12; cmd->cdw12 = cdw12;
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -296,10 +296,10 @@ spdk_nvme_ctrlr_cmd_get_feature(struct spdk_nvme_ctrlr *ctrlr, uint8_t feature,
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_null(cb_fn, cb_arg); req = nvme_allocate_request_null(cb_fn, cb_arg);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -309,7 +309,7 @@ spdk_nvme_ctrlr_cmd_get_feature(struct spdk_nvme_ctrlr *ctrlr, uint8_t feature,
cmd->cdw11 = cdw11; cmd->cdw11 = cdw11;
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -356,10 +356,10 @@ spdk_nvme_ctrlr_cmd_get_log_page(struct spdk_nvme_ctrlr *ctrlr, uint8_t log_page
numdl = numd & 0xFFFFu; numdl = numd & 0xFFFFu;
numdu = (numd >> 16) & 0xFFFFu; numdu = (numd >> 16) & 0xFFFFu;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_user_copy(payload, payload_size, cb_fn, cb_arg, false); req = nvme_allocate_request_user_copy(payload, payload_size, cb_fn, cb_arg, false);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -371,7 +371,7 @@ spdk_nvme_ctrlr_cmd_get_log_page(struct spdk_nvme_ctrlr *ctrlr, uint8_t log_page
cmd->cdw11 = numdu; cmd->cdw11 = numdu;
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }
@ -404,10 +404,10 @@ nvme_ctrlr_cmd_fw_commit(struct spdk_nvme_ctrlr *ctrlr,
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_null(cb_fn, cb_arg); req = nvme_allocate_request_null(cb_fn, cb_arg);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -416,7 +416,7 @@ nvme_ctrlr_cmd_fw_commit(struct spdk_nvme_ctrlr *ctrlr,
memcpy(&cmd->cdw10, fw_commit, sizeof(uint32_t)); memcpy(&cmd->cdw10, fw_commit, sizeof(uint32_t));
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
@ -431,10 +431,10 @@ nvme_ctrlr_cmd_fw_image_download(struct spdk_nvme_ctrlr *ctrlr,
struct spdk_nvme_cmd *cmd; struct spdk_nvme_cmd *cmd;
int rc; int rc;
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
req = nvme_allocate_request_user_copy(payload, size, cb_fn, cb_arg, true); req = nvme_allocate_request_user_copy(payload, size, cb_fn, cb_arg, true);
if (req == NULL) { if (req == NULL) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return -ENOMEM; return -ENOMEM;
} }
@ -444,7 +444,7 @@ nvme_ctrlr_cmd_fw_image_download(struct spdk_nvme_ctrlr *ctrlr,
cmd->cdw11 = offset >> 2; cmd->cdw11 = offset >> 2;
rc = nvme_ctrlr_submit_admin_request(ctrlr, req); rc = nvme_ctrlr_submit_admin_request(ctrlr, req);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
return rc; return rc;
} }

View File

@ -55,9 +55,9 @@ int nvme_ns_identify_update(struct spdk_nvme_ns *ns)
} }
while (status.done == false) { while (status.done == false) {
pthread_mutex_lock(&ns->ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ns->ctrlr->ctrlr_lock);
spdk_nvme_qpair_process_completions(ns->ctrlr->adminq, 0); spdk_nvme_qpair_process_completions(ns->ctrlr->adminq, 0);
pthread_mutex_unlock(&ns->ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ns->ctrlr->ctrlr_lock);
} }
if (spdk_nvme_cpl_is_error(&status.cpl)) { if (spdk_nvme_cpl_is_error(&status.cpl)) {
/* This can occur if the namespace is not active. Simply zero the /* This can occur if the namespace is not active. Simply zero the

View File

@ -1657,7 +1657,7 @@ nvme_pcie_qpair_submit_request(struct spdk_nvme_qpair *qpair, struct nvme_reques
nvme_pcie_qpair_check_enabled(qpair); nvme_pcie_qpair_check_enabled(qpair);
if (nvme_qpair_is_admin_queue(qpair)) { if (nvme_qpair_is_admin_queue(qpair)) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
} }
tr = LIST_FIRST(&pqpair->free_tr); tr = LIST_FIRST(&pqpair->free_tr);
@ -1706,7 +1706,7 @@ nvme_pcie_qpair_submit_request(struct spdk_nvme_qpair *qpair, struct nvme_reques
exit: exit:
if (nvme_qpair_is_admin_queue(qpair)) { if (nvme_qpair_is_admin_queue(qpair)) {
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
return rc; return rc;
@ -1733,7 +1733,7 @@ nvme_pcie_qpair_process_completions(struct spdk_nvme_qpair *qpair, uint32_t max_
} }
if (nvme_qpair_is_admin_queue(qpair)) { if (nvme_qpair_is_admin_queue(qpair)) {
pthread_mutex_lock(&ctrlr->ctrlr_lock); nvme_robust_mutex_lock(&ctrlr->ctrlr_lock);
} }
if (max_completions == 0 || (max_completions > (qpair->num_entries - 1U))) { if (max_completions == 0 || (max_completions > (qpair->num_entries - 1U))) {
@ -1782,7 +1782,7 @@ nvme_pcie_qpair_process_completions(struct spdk_nvme_qpair *qpair, uint32_t max_
if (nvme_qpair_is_admin_queue(qpair)) { if (nvme_qpair_is_admin_queue(qpair)) {
nvme_pcie_qpair_complete_pending_admin_request(qpair); nvme_pcie_qpair_complete_pending_admin_request(qpair);
pthread_mutex_unlock(&ctrlr->ctrlr_lock); nvme_robust_mutex_unlock(&ctrlr->ctrlr_lock);
} }
return num_completions; return num_completions;