Many open source projects have moved to using SPDX identifiers to specify license information, reducing the amount of boilerplate code in every source file. This patch replaces the bulk of SPDK .c, .cpp and Makefiles with the BSD-3-Clause identifier. Almost all of these files share the exact same license text, and this patch only modifies the files that contain the most common license text. There can be slight variations because the third clause contains company names - most say "Intel Corporation", but there are instances for Nvidia, Samsung, Eideticom and even "the copyright holder". Used a bash script to automate replacement of the license text with SPDX identifier which is checked into scripts/spdx.sh. Signed-off-by: Jim Harris <james.r.harris@intel.com> Change-Id: Iaa88ab5e92ea471691dc298cfe41ebfb5d169780 Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/12904 Community-CI: Broadcom CI <spdk-ci.pdl@broadcom.com> Community-CI: Mellanox Build Bot Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Aleksey Marchuk <alexeymar@nvidia.com> Reviewed-by: Changpeng Liu <changpeng.liu@intel.com> Reviewed-by: Dong Yi <dongx.yi@intel.com> Reviewed-by: Konrad Sztyber <konrad.sztyber@intel.com> Reviewed-by: Paul Luse <paul.e.luse@intel.com> Reviewed-by: <qun.wan@intel.com>
1870 lines
50 KiB
C
1870 lines
50 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright (c) Intel Corporation.
|
|
* All rights reserved.
|
|
* Copyright (c) 2021 Mellanox Technologies LTD. All rights reserved.
|
|
*/
|
|
|
|
#include "spdk/stdinc.h"
|
|
|
|
#include "spdk/env.h"
|
|
#include "spdk/likely.h"
|
|
#include "spdk/string.h"
|
|
#include "spdk/util.h"
|
|
#include "spdk/memory.h"
|
|
#include "spdk/barrier.h"
|
|
#include "spdk/vhost.h"
|
|
#include "vhost_internal.h"
|
|
#include <rte_version.h>
|
|
|
|
#include "spdk_internal/vhost_user.h"
|
|
|
|
/* Path to folder where character device will be created. Can be set by user. */
|
|
static char g_vhost_user_dev_dirname[PATH_MAX] = "";
|
|
|
|
static struct spdk_thread *g_vhost_user_init_thread;
|
|
|
|
/**
|
|
* DPDK calls our callbacks synchronously but the work those callbacks
|
|
* perform needs to be async. Luckily, all DPDK callbacks are called on
|
|
* a DPDK-internal pthread, so we'll just wait on a semaphore in there.
|
|
*/
|
|
static sem_t g_dpdk_sem;
|
|
|
|
/** Return code for the current DPDK callback */
|
|
static int g_dpdk_response;
|
|
|
|
struct vhost_session_fn_ctx {
|
|
/** Device pointer obtained before enqueueing the event */
|
|
struct spdk_vhost_dev *vdev;
|
|
|
|
/** ID of the session to send event to. */
|
|
uint32_t vsession_id;
|
|
|
|
/** User provided function to be executed on session's thread. */
|
|
spdk_vhost_session_fn cb_fn;
|
|
|
|
/**
|
|
* User provided function to be called on the init thread
|
|
* after iterating through all sessions.
|
|
*/
|
|
spdk_vhost_dev_fn cpl_fn;
|
|
|
|
/** Custom user context */
|
|
void *user_ctx;
|
|
};
|
|
|
|
static struct spdk_vhost_user_dev *
|
|
to_user_dev(struct spdk_vhost_dev *vdev)
|
|
{
|
|
assert(vdev != NULL);
|
|
return vdev->ctxt;
|
|
}
|
|
|
|
static void __attribute__((constructor))
|
|
_vhost_user_sem_init(void)
|
|
{
|
|
if (sem_init(&g_dpdk_sem, 0, 0) != 0) {
|
|
SPDK_ERRLOG("Failed to initialize semaphore for rte_vhost pthread.\n");
|
|
abort();
|
|
}
|
|
}
|
|
|
|
static void __attribute__((destructor))
|
|
_vhost_user_sem_destroy(void)
|
|
{
|
|
sem_destroy(&g_dpdk_sem);
|
|
}
|
|
|
|
void *vhost_gpa_to_vva(struct spdk_vhost_session *vsession, uint64_t addr, uint64_t len)
|
|
{
|
|
void *vva;
|
|
uint64_t newlen;
|
|
|
|
newlen = len;
|
|
vva = (void *)rte_vhost_va_from_guest_pa(vsession->mem, addr, &newlen);
|
|
if (newlen != len) {
|
|
return NULL;
|
|
}
|
|
|
|
return vva;
|
|
|
|
}
|
|
|
|
static void
|
|
vhost_log_req_desc(struct spdk_vhost_session *vsession, struct spdk_vhost_virtqueue *virtqueue,
|
|
uint16_t req_id)
|
|
{
|
|
struct vring_desc *desc, *desc_table;
|
|
uint32_t desc_table_size;
|
|
int rc;
|
|
|
|
if (spdk_likely(!vhost_dev_has_feature(vsession, VHOST_F_LOG_ALL))) {
|
|
return;
|
|
}
|
|
|
|
rc = vhost_vq_get_desc(vsession, virtqueue, req_id, &desc, &desc_table, &desc_table_size);
|
|
if (spdk_unlikely(rc != 0)) {
|
|
SPDK_ERRLOG("Can't log used ring descriptors!\n");
|
|
return;
|
|
}
|
|
|
|
do {
|
|
if (vhost_vring_desc_is_wr(desc)) {
|
|
/* To be honest, only pages realy touched should be logged, but
|
|
* doing so would require tracking those changes in each backed.
|
|
* Also backend most likely will touch all/most of those pages so
|
|
* for lets assume we touched all pages passed to as writeable buffers. */
|
|
rte_vhost_log_write(vsession->vid, desc->addr, desc->len);
|
|
}
|
|
vhost_vring_desc_get_next(&desc, desc_table, desc_table_size);
|
|
} while (desc);
|
|
}
|
|
|
|
static void
|
|
vhost_log_used_vring_elem(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue,
|
|
uint16_t idx)
|
|
{
|
|
uint64_t offset, len;
|
|
|
|
if (spdk_likely(!vhost_dev_has_feature(vsession, VHOST_F_LOG_ALL))) {
|
|
return;
|
|
}
|
|
|
|
if (spdk_unlikely(virtqueue->packed.packed_ring)) {
|
|
offset = idx * sizeof(struct vring_packed_desc);
|
|
len = sizeof(struct vring_packed_desc);
|
|
} else {
|
|
offset = offsetof(struct vring_used, ring[idx]);
|
|
len = sizeof(virtqueue->vring.used->ring[idx]);
|
|
}
|
|
|
|
rte_vhost_log_used_vring(vsession->vid, virtqueue->vring_idx, offset, len);
|
|
}
|
|
|
|
static void
|
|
vhost_log_used_vring_idx(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue)
|
|
{
|
|
uint64_t offset, len;
|
|
uint16_t vq_idx;
|
|
|
|
if (spdk_likely(!vhost_dev_has_feature(vsession, VHOST_F_LOG_ALL))) {
|
|
return;
|
|
}
|
|
|
|
offset = offsetof(struct vring_used, idx);
|
|
len = sizeof(virtqueue->vring.used->idx);
|
|
vq_idx = virtqueue - vsession->virtqueue;
|
|
|
|
rte_vhost_log_used_vring(vsession->vid, vq_idx, offset, len);
|
|
}
|
|
|
|
/*
|
|
* Get available requests from avail ring.
|
|
*/
|
|
uint16_t
|
|
vhost_vq_avail_ring_get(struct spdk_vhost_virtqueue *virtqueue, uint16_t *reqs,
|
|
uint16_t reqs_len)
|
|
{
|
|
struct rte_vhost_vring *vring = &virtqueue->vring;
|
|
struct vring_avail *avail = vring->avail;
|
|
uint16_t size_mask = vring->size - 1;
|
|
uint16_t last_idx = virtqueue->last_avail_idx, avail_idx = avail->idx;
|
|
uint16_t count, i;
|
|
int rc;
|
|
uint64_t u64_value;
|
|
|
|
spdk_smp_rmb();
|
|
|
|
if (virtqueue->vsession && spdk_unlikely(virtqueue->vsession->interrupt_mode)) {
|
|
/* Read to clear vring's kickfd */
|
|
rc = read(vring->kickfd, &u64_value, sizeof(u64_value));
|
|
if (rc < 0) {
|
|
SPDK_ERRLOG("failed to acknowledge kickfd: %s.\n", spdk_strerror(errno));
|
|
return -errno;
|
|
}
|
|
}
|
|
|
|
count = avail_idx - last_idx;
|
|
if (spdk_likely(count == 0)) {
|
|
return 0;
|
|
}
|
|
|
|
if (spdk_unlikely(count > vring->size)) {
|
|
/* TODO: the queue is unrecoverably broken and should be marked so.
|
|
* For now we will fail silently and report there are no new avail entries.
|
|
*/
|
|
return 0;
|
|
}
|
|
|
|
count = spdk_min(count, reqs_len);
|
|
|
|
virtqueue->last_avail_idx += count;
|
|
/* Check whether there are unprocessed reqs in vq, then kick vq manually */
|
|
if (virtqueue->vsession && spdk_unlikely(virtqueue->vsession->interrupt_mode)) {
|
|
/* If avail_idx is larger than virtqueue's last_avail_idx, then there is unprocessed reqs.
|
|
* avail_idx should get updated here from memory, in case of race condition with guest.
|
|
*/
|
|
avail_idx = * (volatile uint16_t *) &avail->idx;
|
|
if (avail_idx > virtqueue->last_avail_idx) {
|
|
/* Write to notify vring's kickfd */
|
|
rc = write(vring->kickfd, &u64_value, sizeof(u64_value));
|
|
if (rc < 0) {
|
|
SPDK_ERRLOG("failed to kick vring: %s.\n", spdk_strerror(errno));
|
|
return -errno;
|
|
}
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < count; i++) {
|
|
reqs[i] = vring->avail->ring[(last_idx + i) & size_mask];
|
|
}
|
|
|
|
SPDK_DEBUGLOG(vhost_ring,
|
|
"AVAIL: last_idx=%"PRIu16" avail_idx=%"PRIu16" count=%"PRIu16"\n",
|
|
last_idx, avail_idx, count);
|
|
|
|
return count;
|
|
}
|
|
|
|
static bool
|
|
vhost_vring_desc_is_indirect(struct vring_desc *cur_desc)
|
|
{
|
|
return !!(cur_desc->flags & VRING_DESC_F_INDIRECT);
|
|
}
|
|
|
|
static bool
|
|
vhost_vring_packed_desc_is_indirect(struct vring_packed_desc *cur_desc)
|
|
{
|
|
return (cur_desc->flags & VRING_DESC_F_INDIRECT) != 0;
|
|
}
|
|
|
|
static bool
|
|
vhost_inflight_packed_desc_is_indirect(spdk_vhost_inflight_desc *cur_desc)
|
|
{
|
|
return (cur_desc->flags & VRING_DESC_F_INDIRECT) != 0;
|
|
}
|
|
|
|
int
|
|
vhost_vq_get_desc(struct spdk_vhost_session *vsession, struct spdk_vhost_virtqueue *virtqueue,
|
|
uint16_t req_idx, struct vring_desc **desc, struct vring_desc **desc_table,
|
|
uint32_t *desc_table_size)
|
|
{
|
|
if (spdk_unlikely(req_idx >= virtqueue->vring.size)) {
|
|
return -1;
|
|
}
|
|
|
|
*desc = &virtqueue->vring.desc[req_idx];
|
|
|
|
if (vhost_vring_desc_is_indirect(*desc)) {
|
|
*desc_table_size = (*desc)->len / sizeof(**desc);
|
|
*desc_table = vhost_gpa_to_vva(vsession, (*desc)->addr,
|
|
sizeof(**desc) * *desc_table_size);
|
|
*desc = *desc_table;
|
|
if (*desc == NULL) {
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
*desc_table = virtqueue->vring.desc;
|
|
*desc_table_size = virtqueue->vring.size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool
|
|
vhost_packed_desc_indirect_to_desc_table(struct spdk_vhost_session *vsession,
|
|
uint64_t addr, uint32_t len,
|
|
struct vring_packed_desc **desc_table,
|
|
uint32_t *desc_table_size)
|
|
{
|
|
*desc_table_size = len / sizeof(struct vring_packed_desc);
|
|
|
|
*desc_table = vhost_gpa_to_vva(vsession, addr, len);
|
|
if (spdk_unlikely(*desc_table == NULL)) {
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
int
|
|
vhost_vq_get_desc_packed(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue,
|
|
uint16_t req_idx, struct vring_packed_desc **desc,
|
|
struct vring_packed_desc **desc_table, uint32_t *desc_table_size)
|
|
{
|
|
*desc = &virtqueue->vring.desc_packed[req_idx];
|
|
|
|
/* In packed ring when the desc is non-indirect we get next desc
|
|
* by judging (desc->flag & VRING_DESC_F_NEXT) != 0. When the desc
|
|
* is indirect we get next desc by idx and desc_table_size. It's
|
|
* different from split ring.
|
|
*/
|
|
if (vhost_vring_packed_desc_is_indirect(*desc)) {
|
|
if (!vhost_packed_desc_indirect_to_desc_table(vsession, (*desc)->addr, (*desc)->len,
|
|
desc_table, desc_table_size)) {
|
|
return -1;
|
|
}
|
|
|
|
*desc = *desc_table;
|
|
} else {
|
|
*desc_table = NULL;
|
|
*desc_table_size = 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
vhost_inflight_queue_get_desc(struct spdk_vhost_session *vsession,
|
|
spdk_vhost_inflight_desc *desc_array,
|
|
uint16_t req_idx, spdk_vhost_inflight_desc **desc,
|
|
struct vring_packed_desc **desc_table, uint32_t *desc_table_size)
|
|
{
|
|
*desc = &desc_array[req_idx];
|
|
|
|
if (vhost_inflight_packed_desc_is_indirect(*desc)) {
|
|
if (!vhost_packed_desc_indirect_to_desc_table(vsession, (*desc)->addr, (*desc)->len,
|
|
desc_table, desc_table_size)) {
|
|
return -1;
|
|
}
|
|
|
|
/* This desc is the inflight desc not the packed desc.
|
|
* When set the F_INDIRECT the table entry should be the packed desc
|
|
* so set the inflight desc NULL.
|
|
*/
|
|
*desc = NULL;
|
|
} else {
|
|
/* When not set the F_INDIRECT means there is no packed desc table */
|
|
*desc_table = NULL;
|
|
*desc_table_size = 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
vhost_vq_used_signal(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue)
|
|
{
|
|
if (virtqueue->used_req_cnt == 0) {
|
|
return 0;
|
|
}
|
|
|
|
virtqueue->req_cnt += virtqueue->used_req_cnt;
|
|
virtqueue->used_req_cnt = 0;
|
|
|
|
SPDK_DEBUGLOG(vhost_ring,
|
|
"Queue %td - USED RING: sending IRQ: last used %"PRIu16"\n",
|
|
virtqueue - vsession->virtqueue, virtqueue->last_used_idx);
|
|
|
|
if (rte_vhost_vring_call(vsession->vid, virtqueue->vring_idx) == 0) {
|
|
/* interrupt signalled */
|
|
return 1;
|
|
} else {
|
|
/* interrupt not signalled */
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static void
|
|
session_vq_io_stats_update(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue, uint64_t now)
|
|
{
|
|
uint32_t irq_delay_base = vsession->coalescing_delay_time_base;
|
|
uint32_t io_threshold = vsession->coalescing_io_rate_threshold;
|
|
int32_t irq_delay;
|
|
uint32_t req_cnt;
|
|
|
|
req_cnt = virtqueue->req_cnt + virtqueue->used_req_cnt;
|
|
if (req_cnt <= io_threshold) {
|
|
return;
|
|
}
|
|
|
|
irq_delay = (irq_delay_base * (req_cnt - io_threshold)) / io_threshold;
|
|
virtqueue->irq_delay_time = (uint32_t) spdk_max(0, irq_delay);
|
|
|
|
virtqueue->req_cnt = 0;
|
|
virtqueue->next_event_time = now;
|
|
}
|
|
|
|
static void
|
|
check_session_vq_io_stats(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue, uint64_t now)
|
|
{
|
|
if (now < vsession->next_stats_check_time) {
|
|
return;
|
|
}
|
|
|
|
vsession->next_stats_check_time = now + vsession->stats_check_interval;
|
|
session_vq_io_stats_update(vsession, virtqueue, now);
|
|
}
|
|
|
|
static inline bool
|
|
vhost_vq_event_is_suppressed(struct spdk_vhost_virtqueue *vq)
|
|
{
|
|
if (spdk_unlikely(vq->packed.packed_ring)) {
|
|
if (vq->vring.driver_event->flags & VRING_PACKED_EVENT_FLAG_DISABLE) {
|
|
return true;
|
|
}
|
|
} else {
|
|
if (vq->vring.avail->flags & VRING_AVAIL_F_NO_INTERRUPT) {
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
void
|
|
vhost_session_vq_used_signal(struct spdk_vhost_virtqueue *virtqueue)
|
|
{
|
|
struct spdk_vhost_session *vsession = virtqueue->vsession;
|
|
uint64_t now;
|
|
|
|
if (vsession->coalescing_delay_time_base == 0) {
|
|
if (virtqueue->vring.desc == NULL) {
|
|
return;
|
|
}
|
|
|
|
if (vhost_vq_event_is_suppressed(virtqueue)) {
|
|
return;
|
|
}
|
|
|
|
vhost_vq_used_signal(vsession, virtqueue);
|
|
} else {
|
|
now = spdk_get_ticks();
|
|
check_session_vq_io_stats(vsession, virtqueue, now);
|
|
|
|
/* No need for event right now */
|
|
if (now < virtqueue->next_event_time) {
|
|
return;
|
|
}
|
|
|
|
if (vhost_vq_event_is_suppressed(virtqueue)) {
|
|
return;
|
|
}
|
|
|
|
if (!vhost_vq_used_signal(vsession, virtqueue)) {
|
|
return;
|
|
}
|
|
|
|
/* Syscall is quite long so update time */
|
|
now = spdk_get_ticks();
|
|
virtqueue->next_event_time = now + virtqueue->irq_delay_time;
|
|
}
|
|
}
|
|
|
|
void
|
|
vhost_session_used_signal(struct spdk_vhost_session *vsession)
|
|
{
|
|
struct spdk_vhost_virtqueue *virtqueue;
|
|
uint16_t q_idx;
|
|
|
|
for (q_idx = 0; q_idx < vsession->max_queues; q_idx++) {
|
|
virtqueue = &vsession->virtqueue[q_idx];
|
|
vhost_session_vq_used_signal(virtqueue);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Enqueue id and len to used ring.
|
|
*/
|
|
void
|
|
vhost_vq_used_ring_enqueue(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue,
|
|
uint16_t id, uint32_t len)
|
|
{
|
|
struct rte_vhost_vring *vring = &virtqueue->vring;
|
|
struct vring_used *used = vring->used;
|
|
uint16_t last_idx = virtqueue->last_used_idx & (vring->size - 1);
|
|
uint16_t vq_idx = virtqueue->vring_idx;
|
|
|
|
SPDK_DEBUGLOG(vhost_ring,
|
|
"Queue %td - USED RING: last_idx=%"PRIu16" req id=%"PRIu16" len=%"PRIu32"\n",
|
|
virtqueue - vsession->virtqueue, virtqueue->last_used_idx, id, len);
|
|
|
|
vhost_log_req_desc(vsession, virtqueue, id);
|
|
|
|
virtqueue->last_used_idx++;
|
|
used->ring[last_idx].id = id;
|
|
used->ring[last_idx].len = len;
|
|
|
|
/* Ensure the used ring is updated before we log it or increment used->idx. */
|
|
spdk_smp_wmb();
|
|
|
|
rte_vhost_set_last_inflight_io_split(vsession->vid, vq_idx, id);
|
|
|
|
vhost_log_used_vring_elem(vsession, virtqueue, last_idx);
|
|
* (volatile uint16_t *) &used->idx = virtqueue->last_used_idx;
|
|
vhost_log_used_vring_idx(vsession, virtqueue);
|
|
|
|
rte_vhost_clr_inflight_desc_split(vsession->vid, vq_idx, virtqueue->last_used_idx, id);
|
|
|
|
virtqueue->used_req_cnt++;
|
|
|
|
if (vsession->interrupt_mode) {
|
|
if (virtqueue->vring.desc == NULL || vhost_vq_event_is_suppressed(virtqueue)) {
|
|
return;
|
|
}
|
|
|
|
vhost_vq_used_signal(vsession, virtqueue);
|
|
}
|
|
}
|
|
|
|
void
|
|
vhost_vq_packed_ring_enqueue(struct spdk_vhost_session *vsession,
|
|
struct spdk_vhost_virtqueue *virtqueue,
|
|
uint16_t num_descs, uint16_t buffer_id,
|
|
uint32_t length, uint16_t inflight_head)
|
|
{
|
|
struct vring_packed_desc *desc = &virtqueue->vring.desc_packed[virtqueue->last_used_idx];
|
|
bool used, avail;
|
|
|
|
SPDK_DEBUGLOG(vhost_ring,
|
|
"Queue %td - RING: buffer_id=%"PRIu16"\n",
|
|
virtqueue - vsession->virtqueue, buffer_id);
|
|
|
|
/* When the descriptor is used, two flags in descriptor
|
|
* avail flag and used flag are set to equal
|
|
* and used flag value == used_wrap_counter.
|
|
*/
|
|
used = !!(desc->flags & VRING_DESC_F_USED);
|
|
avail = !!(desc->flags & VRING_DESC_F_AVAIL);
|
|
if (spdk_unlikely(used == virtqueue->packed.used_phase && used == avail)) {
|
|
SPDK_ERRLOG("descriptor has been used before\n");
|
|
return;
|
|
}
|
|
|
|
/* In used desc addr is unused and len specifies the buffer length
|
|
* that has been written to by the device.
|
|
*/
|
|
desc->addr = 0;
|
|
desc->len = length;
|
|
|
|
/* This bit specifies whether any data has been written by the device */
|
|
if (length != 0) {
|
|
desc->flags |= VRING_DESC_F_WRITE;
|
|
}
|
|
|
|
/* Buffer ID is included in the last descriptor in the list.
|
|
* The driver needs to keep track of the size of the list corresponding
|
|
* to each buffer ID.
|
|
*/
|
|
desc->id = buffer_id;
|
|
|
|
/* A device MUST NOT make the descriptor used before buffer_id is
|
|
* written to the descriptor.
|
|
*/
|
|
spdk_smp_wmb();
|
|
|
|
rte_vhost_set_last_inflight_io_packed(vsession->vid, virtqueue->vring_idx, inflight_head);
|
|
/* To mark a desc as used, the device sets the F_USED bit in flags to match
|
|
* the internal Device ring wrap counter. It also sets the F_AVAIL bit to
|
|
* match the same value.
|
|
*/
|
|
if (virtqueue->packed.used_phase) {
|
|
desc->flags |= VRING_DESC_F_AVAIL_USED;
|
|
} else {
|
|
desc->flags &= ~VRING_DESC_F_AVAIL_USED;
|
|
}
|
|
rte_vhost_clr_inflight_desc_packed(vsession->vid, virtqueue->vring_idx, inflight_head);
|
|
|
|
vhost_log_used_vring_elem(vsession, virtqueue, virtqueue->last_used_idx);
|
|
virtqueue->last_used_idx += num_descs;
|
|
if (virtqueue->last_used_idx >= virtqueue->vring.size) {
|
|
virtqueue->last_used_idx -= virtqueue->vring.size;
|
|
virtqueue->packed.used_phase = !virtqueue->packed.used_phase;
|
|
}
|
|
|
|
virtqueue->used_req_cnt++;
|
|
}
|
|
|
|
bool
|
|
vhost_vq_packed_ring_is_avail(struct spdk_vhost_virtqueue *virtqueue)
|
|
{
|
|
uint16_t flags = virtqueue->vring.desc_packed[virtqueue->last_avail_idx].flags;
|
|
|
|
/* To mark a desc as available, the driver sets the F_AVAIL bit in flags
|
|
* to match the internal avail wrap counter. It also sets the F_USED bit to
|
|
* match the inverse value but it's not mandatory.
|
|
*/
|
|
return (!!(flags & VRING_DESC_F_AVAIL) == virtqueue->packed.avail_phase);
|
|
}
|
|
|
|
bool
|
|
vhost_vring_packed_desc_is_wr(struct vring_packed_desc *cur_desc)
|
|
{
|
|
return (cur_desc->flags & VRING_DESC_F_WRITE) != 0;
|
|
}
|
|
|
|
bool
|
|
vhost_vring_inflight_desc_is_wr(spdk_vhost_inflight_desc *cur_desc)
|
|
{
|
|
return (cur_desc->flags & VRING_DESC_F_WRITE) != 0;
|
|
}
|
|
|
|
int
|
|
vhost_vring_packed_desc_get_next(struct vring_packed_desc **desc, uint16_t *req_idx,
|
|
struct spdk_vhost_virtqueue *vq,
|
|
struct vring_packed_desc *desc_table,
|
|
uint32_t desc_table_size)
|
|
{
|
|
if (desc_table != NULL) {
|
|
/* When the desc_table isn't NULL means it's indirect and we get the next
|
|
* desc by req_idx and desc_table_size. The return value is NULL means
|
|
* we reach the last desc of this request.
|
|
*/
|
|
(*req_idx)++;
|
|
if (*req_idx < desc_table_size) {
|
|
*desc = &desc_table[*req_idx];
|
|
} else {
|
|
*desc = NULL;
|
|
}
|
|
} else {
|
|
/* When the desc_table is NULL means it's non-indirect and we get the next
|
|
* desc by req_idx and F_NEXT in flags. The return value is NULL means
|
|
* we reach the last desc of this request. When return new desc
|
|
* we update the req_idx too.
|
|
*/
|
|
if (((*desc)->flags & VRING_DESC_F_NEXT) == 0) {
|
|
*desc = NULL;
|
|
return 0;
|
|
}
|
|
|
|
*req_idx = (*req_idx + 1) % vq->vring.size;
|
|
*desc = &vq->vring.desc_packed[*req_idx];
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
vhost_vring_desc_payload_to_iov(struct spdk_vhost_session *vsession, struct iovec *iov,
|
|
uint16_t *iov_index, uintptr_t payload, uint64_t remaining)
|
|
{
|
|
uintptr_t vva;
|
|
uint64_t len;
|
|
|
|
do {
|
|
if (*iov_index >= SPDK_VHOST_IOVS_MAX) {
|
|
SPDK_ERRLOG("SPDK_VHOST_IOVS_MAX(%d) reached\n", SPDK_VHOST_IOVS_MAX);
|
|
return -1;
|
|
}
|
|
len = remaining;
|
|
vva = (uintptr_t)rte_vhost_va_from_guest_pa(vsession->mem, payload, &len);
|
|
if (vva == 0 || len == 0) {
|
|
SPDK_ERRLOG("gpa_to_vva(%p) == NULL\n", (void *)payload);
|
|
return -1;
|
|
}
|
|
iov[*iov_index].iov_base = (void *)vva;
|
|
iov[*iov_index].iov_len = len;
|
|
remaining -= len;
|
|
payload += len;
|
|
(*iov_index)++;
|
|
} while (remaining);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
vhost_vring_packed_desc_to_iov(struct spdk_vhost_session *vsession, struct iovec *iov,
|
|
uint16_t *iov_index, const struct vring_packed_desc *desc)
|
|
{
|
|
return vhost_vring_desc_payload_to_iov(vsession, iov, iov_index,
|
|
desc->addr, desc->len);
|
|
}
|
|
|
|
int
|
|
vhost_vring_inflight_desc_to_iov(struct spdk_vhost_session *vsession, struct iovec *iov,
|
|
uint16_t *iov_index, const spdk_vhost_inflight_desc *desc)
|
|
{
|
|
return vhost_vring_desc_payload_to_iov(vsession, iov, iov_index,
|
|
desc->addr, desc->len);
|
|
}
|
|
|
|
/* 1, Traverse the desc chain to get the buffer_id and return buffer_id as task_idx.
|
|
* 2, Update the vq->last_avail_idx to point next available desc chain.
|
|
* 3, Update the avail_wrap_counter if last_avail_idx overturn.
|
|
*/
|
|
uint16_t
|
|
vhost_vring_packed_desc_get_buffer_id(struct spdk_vhost_virtqueue *vq, uint16_t req_idx,
|
|
uint16_t *num_descs)
|
|
{
|
|
struct vring_packed_desc *desc;
|
|
uint16_t desc_head = req_idx;
|
|
|
|
*num_descs = 1;
|
|
|
|
desc = &vq->vring.desc_packed[req_idx];
|
|
if (!vhost_vring_packed_desc_is_indirect(desc)) {
|
|
while ((desc->flags & VRING_DESC_F_NEXT) != 0) {
|
|
req_idx = (req_idx + 1) % vq->vring.size;
|
|
desc = &vq->vring.desc_packed[req_idx];
|
|
(*num_descs)++;
|
|
}
|
|
}
|
|
|
|
/* Queue Size doesn't have to be a power of 2
|
|
* Device maintains last_avail_idx so we can make sure
|
|
* the value is valid(0 ~ vring.size - 1)
|
|
*/
|
|
vq->last_avail_idx = (req_idx + 1) % vq->vring.size;
|
|
if (vq->last_avail_idx < desc_head) {
|
|
vq->packed.avail_phase = !vq->packed.avail_phase;
|
|
}
|
|
|
|
return desc->id;
|
|
}
|
|
|
|
int
|
|
vhost_vring_desc_get_next(struct vring_desc **desc,
|
|
struct vring_desc *desc_table, uint32_t desc_table_size)
|
|
{
|
|
struct vring_desc *old_desc = *desc;
|
|
uint16_t next_idx;
|
|
|
|
if ((old_desc->flags & VRING_DESC_F_NEXT) == 0) {
|
|
*desc = NULL;
|
|
return 0;
|
|
}
|
|
|
|
next_idx = old_desc->next;
|
|
if (spdk_unlikely(next_idx >= desc_table_size)) {
|
|
*desc = NULL;
|
|
return -1;
|
|
}
|
|
|
|
*desc = &desc_table[next_idx];
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
vhost_vring_desc_to_iov(struct spdk_vhost_session *vsession, struct iovec *iov,
|
|
uint16_t *iov_index, const struct vring_desc *desc)
|
|
{
|
|
return vhost_vring_desc_payload_to_iov(vsession, iov, iov_index,
|
|
desc->addr, desc->len);
|
|
}
|
|
|
|
static inline void
|
|
vhost_session_mem_region_calc(uint64_t *previous_start, uint64_t *start, uint64_t *end,
|
|
uint64_t *len, struct rte_vhost_mem_region *region)
|
|
{
|
|
*start = FLOOR_2MB(region->mmap_addr);
|
|
*end = CEIL_2MB(region->mmap_addr + region->mmap_size);
|
|
if (*start == *previous_start) {
|
|
*start += (size_t) VALUE_2MB;
|
|
}
|
|
*previous_start = *start;
|
|
*len = *end - *start;
|
|
}
|
|
|
|
void
|
|
vhost_session_mem_register(struct rte_vhost_memory *mem)
|
|
{
|
|
uint64_t start, end, len;
|
|
uint32_t i;
|
|
uint64_t previous_start = UINT64_MAX;
|
|
|
|
|
|
for (i = 0; i < mem->nregions; i++) {
|
|
vhost_session_mem_region_calc(&previous_start, &start, &end, &len, &mem->regions[i]);
|
|
SPDK_INFOLOG(vhost, "Registering VM memory for vtophys translation - 0x%jx len:0x%jx\n",
|
|
start, len);
|
|
|
|
if (spdk_mem_register((void *)start, len) != 0) {
|
|
SPDK_WARNLOG("Failed to register memory region %"PRIu32". Future vtophys translation might fail.\n",
|
|
i);
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
void
|
|
vhost_session_mem_unregister(struct rte_vhost_memory *mem)
|
|
{
|
|
uint64_t start, end, len;
|
|
uint32_t i;
|
|
uint64_t previous_start = UINT64_MAX;
|
|
|
|
for (i = 0; i < mem->nregions; i++) {
|
|
vhost_session_mem_region_calc(&previous_start, &start, &end, &len, &mem->regions[i]);
|
|
if (spdk_vtophys((void *) start, NULL) == SPDK_VTOPHYS_ERROR) {
|
|
continue; /* region has not been registered */
|
|
}
|
|
|
|
if (spdk_mem_unregister((void *)start, len) != 0) {
|
|
assert(false);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int
|
|
_stop_session(struct spdk_vhost_session *vsession)
|
|
{
|
|
struct spdk_vhost_dev *vdev = vsession->vdev;
|
|
struct spdk_vhost_user_dev *user_vdev = to_user_dev(vdev);
|
|
struct spdk_vhost_virtqueue *q;
|
|
int rc;
|
|
uint16_t i;
|
|
|
|
rc = user_vdev->user_backend->stop_session(vsession);
|
|
if (rc != 0) {
|
|
SPDK_ERRLOG("Couldn't stop device with vid %d.\n", vsession->vid);
|
|
return rc;
|
|
}
|
|
|
|
for (i = 0; i < vsession->max_queues; i++) {
|
|
q = &vsession->virtqueue[i];
|
|
|
|
/* vring.desc and vring.desc_packed are in a union struct
|
|
* so q->vring.desc can replace q->vring.desc_packed.
|
|
*/
|
|
if (q->vring.desc == NULL) {
|
|
continue;
|
|
}
|
|
|
|
/* Packed virtqueues support up to 2^15 entries each
|
|
* so left one bit can be used as wrap counter.
|
|
*/
|
|
if (q->packed.packed_ring) {
|
|
q->last_avail_idx = q->last_avail_idx |
|
|
((uint16_t)q->packed.avail_phase << 15);
|
|
q->last_used_idx = q->last_used_idx |
|
|
((uint16_t)q->packed.used_phase << 15);
|
|
}
|
|
|
|
rte_vhost_set_vring_base(vsession->vid, i, q->last_avail_idx, q->last_used_idx);
|
|
}
|
|
|
|
vhost_session_mem_unregister(vsession->mem);
|
|
free(vsession->mem);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
new_connection(int vid)
|
|
{
|
|
struct spdk_vhost_dev *vdev;
|
|
struct spdk_vhost_user_dev *user_dev;
|
|
struct spdk_vhost_session *vsession;
|
|
size_t dev_dirname_len;
|
|
char ifname[PATH_MAX];
|
|
char *ctrlr_name;
|
|
|
|
if (rte_vhost_get_ifname(vid, ifname, PATH_MAX) < 0) {
|
|
SPDK_ERRLOG("Couldn't get a valid ifname for device with vid %d\n", vid);
|
|
return -1;
|
|
}
|
|
|
|
spdk_vhost_lock();
|
|
|
|
ctrlr_name = &ifname[0];
|
|
dev_dirname_len = strlen(g_vhost_user_dev_dirname);
|
|
if (strncmp(ctrlr_name, g_vhost_user_dev_dirname, dev_dirname_len) == 0) {
|
|
ctrlr_name += dev_dirname_len;
|
|
}
|
|
|
|
vdev = spdk_vhost_dev_find(ctrlr_name);
|
|
if (vdev == NULL) {
|
|
SPDK_ERRLOG("Couldn't find device with vid %d to create connection for.\n", vid);
|
|
spdk_vhost_unlock();
|
|
return -1;
|
|
}
|
|
user_dev = to_user_dev(vdev);
|
|
|
|
/* We expect sessions inside user_dev->vsessions to be sorted in ascending
|
|
* order in regard of vsession->id. For now we always set id = vsessions_cnt++
|
|
* and append each session to the very end of the vsessions list.
|
|
* This is required for vhost_user_dev_foreach_session() to work.
|
|
*/
|
|
if (user_dev->vsessions_num == UINT_MAX) {
|
|
assert(false);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (posix_memalign((void **)&vsession, SPDK_CACHE_LINE_SIZE, sizeof(*vsession) +
|
|
user_dev->user_backend->session_ctx_size)) {
|
|
SPDK_ERRLOG("vsession alloc failed\n");
|
|
spdk_vhost_unlock();
|
|
return -1;
|
|
}
|
|
memset(vsession, 0, sizeof(*vsession) + user_dev->user_backend->session_ctx_size);
|
|
|
|
vsession->vdev = vdev;
|
|
vsession->vid = vid;
|
|
vsession->id = user_dev->vsessions_num++;
|
|
vsession->name = spdk_sprintf_alloc("%ss%u", vdev->name, vsession->vid);
|
|
if (vsession->name == NULL) {
|
|
SPDK_ERRLOG("vsession alloc failed\n");
|
|
spdk_vhost_unlock();
|
|
free(vsession);
|
|
return -1;
|
|
}
|
|
vsession->started = false;
|
|
vsession->initialized = false;
|
|
vsession->next_stats_check_time = 0;
|
|
vsession->stats_check_interval = SPDK_VHOST_STATS_CHECK_INTERVAL_MS *
|
|
spdk_get_ticks_hz() / 1000UL;
|
|
TAILQ_INSERT_TAIL(&user_dev->vsessions, vsession, tailq);
|
|
|
|
vhost_session_install_rte_compat_hooks(vsession);
|
|
spdk_vhost_unlock();
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
start_device(int vid)
|
|
{
|
|
struct spdk_vhost_dev *vdev;
|
|
struct spdk_vhost_session *vsession;
|
|
int rc = -1;
|
|
uint16_t i;
|
|
bool packed_ring;
|
|
|
|
spdk_vhost_lock();
|
|
|
|
vsession = vhost_session_find_by_vid(vid);
|
|
if (vsession == NULL) {
|
|
SPDK_ERRLOG("Couldn't find session with vid %d.\n", vid);
|
|
goto out;
|
|
}
|
|
|
|
vdev = vsession->vdev;
|
|
if (vsession->started) {
|
|
/* already started, nothing to do */
|
|
rc = 0;
|
|
goto out;
|
|
}
|
|
|
|
if (vhost_get_negotiated_features(vid, &vsession->negotiated_features) != 0) {
|
|
SPDK_ERRLOG("vhost device %d: Failed to get negotiated driver features\n", vid);
|
|
goto out;
|
|
}
|
|
|
|
packed_ring = ((vsession->negotiated_features & (1ULL << VIRTIO_F_RING_PACKED)) != 0);
|
|
|
|
vsession->max_queues = 0;
|
|
memset(vsession->virtqueue, 0, sizeof(vsession->virtqueue));
|
|
for (i = 0; i < SPDK_VHOST_MAX_VQUEUES; i++) {
|
|
struct spdk_vhost_virtqueue *q = &vsession->virtqueue[i];
|
|
|
|
q->vsession = vsession;
|
|
q->vring_idx = -1;
|
|
if (rte_vhost_get_vhost_vring(vid, i, &q->vring)) {
|
|
continue;
|
|
}
|
|
q->vring_idx = i;
|
|
rte_vhost_get_vhost_ring_inflight(vid, i, &q->vring_inflight);
|
|
|
|
/* vring.desc and vring.desc_packed are in a union struct
|
|
* so q->vring.desc can replace q->vring.desc_packed.
|
|
*/
|
|
if (q->vring.desc == NULL || q->vring.size == 0) {
|
|
continue;
|
|
}
|
|
|
|
if (rte_vhost_get_vring_base(vsession->vid, i, &q->last_avail_idx, &q->last_used_idx)) {
|
|
q->vring.desc = NULL;
|
|
continue;
|
|
}
|
|
|
|
if (packed_ring) {
|
|
/* Use the inflight mem to restore the last_avail_idx and last_used_idx.
|
|
* When the vring format is packed, there is no used_idx in the
|
|
* used ring, so VM can't resend the used_idx to VHOST when reconnect.
|
|
* QEMU version 5.2.0 supports the packed inflight before that it only
|
|
* supports split ring inflight because it doesn't send negotiated features
|
|
* before get inflight fd. Users can use RPC to enable this function.
|
|
*/
|
|
if (spdk_unlikely(vdev->packed_ring_recovery)) {
|
|
rte_vhost_get_vring_base_from_inflight(vsession->vid, i,
|
|
&q->last_avail_idx,
|
|
&q->last_used_idx);
|
|
}
|
|
|
|
/* Packed virtqueues support up to 2^15 entries each
|
|
* so left one bit can be used as wrap counter.
|
|
*/
|
|
q->packed.avail_phase = q->last_avail_idx >> 15;
|
|
q->last_avail_idx = q->last_avail_idx & 0x7FFF;
|
|
q->packed.used_phase = q->last_used_idx >> 15;
|
|
q->last_used_idx = q->last_used_idx & 0x7FFF;
|
|
|
|
if (!vsession->interrupt_mode) {
|
|
/* Disable I/O submission notifications, we'll be polling. */
|
|
q->vring.device_event->flags = VRING_PACKED_EVENT_FLAG_DISABLE;
|
|
}
|
|
} else {
|
|
if (!vsession->interrupt_mode) {
|
|
/* Disable I/O submission notifications, we'll be polling. */
|
|
q->vring.used->flags = VRING_USED_F_NO_NOTIFY;
|
|
}
|
|
}
|
|
|
|
q->packed.packed_ring = packed_ring;
|
|
vsession->max_queues = i + 1;
|
|
}
|
|
|
|
if (vhost_get_mem_table(vid, &vsession->mem) != 0) {
|
|
SPDK_ERRLOG("vhost device %d: Failed to get guest memory table\n", vid);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Not sure right now but this look like some kind of QEMU bug and guest IO
|
|
* might be frozed without kicking all queues after live-migration. This look like
|
|
* the previous vhost instance failed to effectively deliver all interrupts before
|
|
* the GET_VRING_BASE message. This shouldn't harm guest since spurious interrupts
|
|
* should be ignored by guest virtio driver.
|
|
*
|
|
* Tested on QEMU 2.10.91 and 2.11.50.
|
|
*/
|
|
for (i = 0; i < vsession->max_queues; i++) {
|
|
struct spdk_vhost_virtqueue *q = &vsession->virtqueue[i];
|
|
|
|
/* vring.desc and vring.desc_packed are in a union struct
|
|
* so q->vring.desc can replace q->vring.desc_packed.
|
|
*/
|
|
if (q->vring.desc != NULL && q->vring.size > 0) {
|
|
rte_vhost_vring_call(vsession->vid, q->vring_idx);
|
|
}
|
|
}
|
|
|
|
vhost_user_session_set_coalescing(vdev, vsession, NULL);
|
|
vhost_session_mem_register(vsession->mem);
|
|
vsession->initialized = true;
|
|
rc = to_user_dev(vdev)->user_backend->start_session(vsession);
|
|
if (rc != 0) {
|
|
vhost_session_mem_unregister(vsession->mem);
|
|
free(vsession->mem);
|
|
goto out;
|
|
}
|
|
|
|
out:
|
|
spdk_vhost_unlock();
|
|
return rc;
|
|
}
|
|
|
|
static void
|
|
stop_device(int vid)
|
|
{
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
spdk_vhost_lock();
|
|
vsession = vhost_session_find_by_vid(vid);
|
|
if (vsession == NULL) {
|
|
SPDK_ERRLOG("Couldn't find session with vid %d.\n", vid);
|
|
spdk_vhost_unlock();
|
|
return;
|
|
}
|
|
|
|
if (!vsession->started) {
|
|
/* already stopped, nothing to do */
|
|
spdk_vhost_unlock();
|
|
return;
|
|
}
|
|
|
|
_stop_session(vsession);
|
|
spdk_vhost_unlock();
|
|
|
|
return;
|
|
}
|
|
|
|
static void
|
|
destroy_connection(int vid)
|
|
{
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
spdk_vhost_lock();
|
|
vsession = vhost_session_find_by_vid(vid);
|
|
if (vsession == NULL) {
|
|
SPDK_ERRLOG("Couldn't find session with vid %d.\n", vid);
|
|
spdk_vhost_unlock();
|
|
return;
|
|
}
|
|
|
|
if (vsession->started) {
|
|
if (_stop_session(vsession) != 0) {
|
|
spdk_vhost_unlock();
|
|
return;
|
|
}
|
|
}
|
|
|
|
TAILQ_REMOVE(&to_user_dev(vsession->vdev)->vsessions, vsession, tailq);
|
|
free(vsession->name);
|
|
free(vsession);
|
|
spdk_vhost_unlock();
|
|
}
|
|
|
|
#if RTE_VERSION >= RTE_VERSION_NUM(21, 11, 0, 0)
|
|
static const struct rte_vhost_device_ops g_spdk_vhost_ops = {
|
|
#else
|
|
static const struct vhost_device_ops g_spdk_vhost_ops = {
|
|
#endif
|
|
.new_device = start_device,
|
|
.destroy_device = stop_device,
|
|
.new_connection = new_connection,
|
|
.destroy_connection = destroy_connection,
|
|
};
|
|
|
|
static struct spdk_vhost_session *
|
|
vhost_session_find_by_id(struct spdk_vhost_dev *vdev, unsigned id)
|
|
{
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
TAILQ_FOREACH(vsession, &to_user_dev(vdev)->vsessions, tailq) {
|
|
if (vsession->id == id) {
|
|
return vsession;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct spdk_vhost_session *
|
|
vhost_session_find_by_vid(int vid)
|
|
{
|
|
struct spdk_vhost_dev *vdev;
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
for (vdev = spdk_vhost_dev_next(NULL); vdev != NULL;
|
|
vdev = spdk_vhost_dev_next(vdev)) {
|
|
TAILQ_FOREACH(vsession, &to_user_dev(vdev)->vsessions, tailq) {
|
|
if (vsession->vid == vid) {
|
|
return vsession;
|
|
}
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
wait_for_semaphore(int timeout_sec, const char *errmsg)
|
|
{
|
|
struct timespec timeout;
|
|
int rc;
|
|
|
|
clock_gettime(CLOCK_REALTIME, &timeout);
|
|
timeout.tv_sec += timeout_sec;
|
|
rc = sem_timedwait(&g_dpdk_sem, &timeout);
|
|
if (rc != 0) {
|
|
SPDK_ERRLOG("Timeout waiting for event: %s.\n", errmsg);
|
|
sem_wait(&g_dpdk_sem);
|
|
}
|
|
}
|
|
|
|
static void
|
|
vhost_session_cb_done(int rc)
|
|
{
|
|
g_dpdk_response = rc;
|
|
sem_post(&g_dpdk_sem);
|
|
}
|
|
|
|
void
|
|
vhost_user_session_start_done(struct spdk_vhost_session *vsession, int response)
|
|
{
|
|
struct spdk_vhost_user_dev *user_dev = to_user_dev(vsession->vdev);
|
|
if (response == 0) {
|
|
vsession->started = true;
|
|
|
|
assert(user_dev->active_session_num < UINT32_MAX);
|
|
user_dev->active_session_num++;
|
|
}
|
|
|
|
vhost_session_cb_done(response);
|
|
}
|
|
|
|
void
|
|
vhost_user_session_stop_done(struct spdk_vhost_session *vsession, int response)
|
|
{
|
|
struct spdk_vhost_user_dev *user_dev = to_user_dev(vsession->vdev);
|
|
|
|
if (response == 0) {
|
|
vsession->started = false;
|
|
|
|
assert(user_dev->active_session_num > 0);
|
|
user_dev->active_session_num--;
|
|
}
|
|
|
|
vhost_session_cb_done(response);
|
|
}
|
|
|
|
static void
|
|
vhost_event_cb(void *arg1)
|
|
{
|
|
struct vhost_session_fn_ctx *ctx = arg1;
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
if (spdk_vhost_trylock() != 0) {
|
|
spdk_thread_send_msg(spdk_get_thread(), vhost_event_cb, arg1);
|
|
return;
|
|
}
|
|
|
|
vsession = vhost_session_find_by_id(ctx->vdev, ctx->vsession_id);
|
|
ctx->cb_fn(ctx->vdev, vsession, NULL);
|
|
spdk_vhost_unlock();
|
|
}
|
|
|
|
int
|
|
vhost_user_session_send_event(struct spdk_vhost_session *vsession,
|
|
spdk_vhost_session_fn cb_fn, unsigned timeout_sec,
|
|
const char *errmsg)
|
|
{
|
|
struct vhost_session_fn_ctx ev_ctx = {0};
|
|
struct spdk_vhost_dev *vdev = vsession->vdev;
|
|
|
|
ev_ctx.vdev = vdev;
|
|
ev_ctx.vsession_id = vsession->id;
|
|
ev_ctx.cb_fn = cb_fn;
|
|
|
|
spdk_thread_send_msg(vdev->thread, vhost_event_cb, &ev_ctx);
|
|
|
|
spdk_vhost_unlock();
|
|
wait_for_semaphore(timeout_sec, errmsg);
|
|
spdk_vhost_lock();
|
|
|
|
return g_dpdk_response;
|
|
}
|
|
|
|
static void
|
|
foreach_session_finish_cb(void *arg1)
|
|
{
|
|
struct vhost_session_fn_ctx *ev_ctx = arg1;
|
|
struct spdk_vhost_dev *vdev = ev_ctx->vdev;
|
|
struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
|
|
|
|
if (spdk_vhost_trylock() != 0) {
|
|
spdk_thread_send_msg(spdk_get_thread(),
|
|
foreach_session_finish_cb, arg1);
|
|
return;
|
|
}
|
|
|
|
assert(user_dev->pending_async_op_num > 0);
|
|
user_dev->pending_async_op_num--;
|
|
if (ev_ctx->cpl_fn != NULL) {
|
|
ev_ctx->cpl_fn(vdev, ev_ctx->user_ctx);
|
|
}
|
|
|
|
spdk_vhost_unlock();
|
|
free(ev_ctx);
|
|
}
|
|
|
|
static void
|
|
foreach_session(void *arg1)
|
|
{
|
|
struct vhost_session_fn_ctx *ev_ctx = arg1;
|
|
struct spdk_vhost_session *vsession;
|
|
struct spdk_vhost_dev *vdev = ev_ctx->vdev;
|
|
int rc;
|
|
|
|
if (spdk_vhost_trylock() != 0) {
|
|
spdk_thread_send_msg(spdk_get_thread(), foreach_session, arg1);
|
|
return;
|
|
}
|
|
|
|
TAILQ_FOREACH(vsession, &to_user_dev(vdev)->vsessions, tailq) {
|
|
if (vsession->initialized) {
|
|
rc = ev_ctx->cb_fn(vdev, vsession, ev_ctx->user_ctx);
|
|
if (rc < 0) {
|
|
goto out;
|
|
}
|
|
}
|
|
}
|
|
|
|
out:
|
|
spdk_vhost_unlock();
|
|
|
|
spdk_thread_send_msg(g_vhost_user_init_thread, foreach_session_finish_cb, arg1);
|
|
}
|
|
|
|
void
|
|
vhost_user_dev_foreach_session(struct spdk_vhost_dev *vdev,
|
|
spdk_vhost_session_fn fn,
|
|
spdk_vhost_dev_fn cpl_fn,
|
|
void *arg)
|
|
{
|
|
struct vhost_session_fn_ctx *ev_ctx;
|
|
struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
|
|
|
|
ev_ctx = calloc(1, sizeof(*ev_ctx));
|
|
if (ev_ctx == NULL) {
|
|
SPDK_ERRLOG("Failed to alloc vhost event.\n");
|
|
assert(false);
|
|
return;
|
|
}
|
|
|
|
ev_ctx->vdev = vdev;
|
|
ev_ctx->cb_fn = fn;
|
|
ev_ctx->cpl_fn = cpl_fn;
|
|
ev_ctx->user_ctx = arg;
|
|
|
|
assert(user_dev->pending_async_op_num < UINT32_MAX);
|
|
user_dev->pending_async_op_num++;
|
|
|
|
spdk_thread_send_msg(vdev->thread, foreach_session, ev_ctx);
|
|
}
|
|
|
|
void
|
|
vhost_user_session_set_interrupt_mode(struct spdk_vhost_session *vsession, bool interrupt_mode)
|
|
{
|
|
uint16_t i;
|
|
bool packed_ring;
|
|
int rc = 0;
|
|
|
|
packed_ring = ((vsession->negotiated_features & (1ULL << VIRTIO_F_RING_PACKED)) != 0);
|
|
|
|
for (i = 0; i < vsession->max_queues; i++) {
|
|
struct spdk_vhost_virtqueue *q = &vsession->virtqueue[i];
|
|
uint64_t num_events = 1;
|
|
|
|
/* vring.desc and vring.desc_packed are in a union struct
|
|
* so q->vring.desc can replace q->vring.desc_packed.
|
|
*/
|
|
if (q->vring.desc == NULL || q->vring.size == 0) {
|
|
continue;
|
|
}
|
|
|
|
if (interrupt_mode) {
|
|
/* Enable I/O submission notifications, we'll be interrupting. */
|
|
if (packed_ring) {
|
|
* (volatile uint16_t *) &q->vring.device_event->flags = VRING_PACKED_EVENT_FLAG_ENABLE;
|
|
} else {
|
|
* (volatile uint16_t *) &q->vring.used->flags = 0;
|
|
}
|
|
|
|
/* In case of race condition, always kick vring when switch to intr */
|
|
rc = write(q->vring.kickfd, &num_events, sizeof(num_events));
|
|
if (rc < 0) {
|
|
SPDK_ERRLOG("failed to kick vring: %s.\n", spdk_strerror(errno));
|
|
}
|
|
|
|
vsession->interrupt_mode = true;
|
|
} else {
|
|
/* Disable I/O submission notifications, we'll be polling. */
|
|
if (packed_ring) {
|
|
* (volatile uint16_t *) &q->vring.device_event->flags = VRING_PACKED_EVENT_FLAG_DISABLE;
|
|
} else {
|
|
* (volatile uint16_t *) &q->vring.used->flags = VRING_USED_F_NO_NOTIFY;
|
|
}
|
|
|
|
vsession->interrupt_mode = false;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
static enum rte_vhost_msg_result
|
|
extern_vhost_pre_msg_handler(int vid, void *_msg)
|
|
{
|
|
struct vhost_user_msg *msg = _msg;
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
vsession = vhost_session_find_by_vid(vid);
|
|
if (vsession == NULL) {
|
|
SPDK_ERRLOG("Received a message to unitialized session (vid %d).\n", vid);
|
|
assert(false);
|
|
return RTE_VHOST_MSG_RESULT_ERR;
|
|
}
|
|
|
|
switch (msg->request) {
|
|
case VHOST_USER_GET_VRING_BASE:
|
|
if (vsession->forced_polling && vsession->started) {
|
|
/* Our queue is stopped for whatever reason, but we may still
|
|
* need to poll it after it's initialized again.
|
|
*/
|
|
g_spdk_vhost_ops.destroy_device(vid);
|
|
}
|
|
break;
|
|
case VHOST_USER_SET_VRING_BASE:
|
|
case VHOST_USER_SET_VRING_ADDR:
|
|
case VHOST_USER_SET_VRING_NUM:
|
|
if (vsession->forced_polling && vsession->started) {
|
|
/* Additional queues are being initialized, so we either processed
|
|
* enough I/Os and are switching from SeaBIOS to the OS now, or
|
|
* we were never in SeaBIOS in the first place. Either way, we
|
|
* don't need our workaround anymore.
|
|
*/
|
|
g_spdk_vhost_ops.destroy_device(vid);
|
|
vsession->forced_polling = false;
|
|
}
|
|
break;
|
|
case VHOST_USER_SET_VRING_KICK:
|
|
/* rte_vhost(after 20.08) will call new_device after one active vring is
|
|
* configured, we will start the session before all vrings are available,
|
|
* so for each new vring, if the session is started, we need to restart it
|
|
* again.
|
|
*/
|
|
case VHOST_USER_SET_VRING_CALL:
|
|
/* rte_vhost will close the previous callfd and won't notify
|
|
* us about any change. This will effectively make SPDK fail
|
|
* to deliver any subsequent interrupts until a session is
|
|
* restarted. We stop the session here before closing the previous
|
|
* fd (so that all interrupts must have been delivered by the
|
|
* time the descriptor is closed) and start right after (which
|
|
* will make SPDK retrieve the latest, up-to-date callfd from
|
|
* rte_vhost.
|
|
*/
|
|
case VHOST_USER_SET_MEM_TABLE:
|
|
/* rte_vhost will unmap previous memory that SPDK may still
|
|
* have pending DMA operations on. We can't let that happen,
|
|
* so stop the device before letting rte_vhost unmap anything.
|
|
* This will block until all pending I/Os are finished.
|
|
* We will start the device again from the post-processing
|
|
* message handler.
|
|
*/
|
|
if (vsession->started) {
|
|
g_spdk_vhost_ops.destroy_device(vid);
|
|
vsession->needs_restart = true;
|
|
}
|
|
break;
|
|
case VHOST_USER_GET_CONFIG: {
|
|
int rc = 0;
|
|
|
|
spdk_vhost_lock();
|
|
if (vsession->vdev->backend->vhost_get_config) {
|
|
rc = vsession->vdev->backend->vhost_get_config(vsession->vdev,
|
|
msg->payload.cfg.region, msg->payload.cfg.size);
|
|
if (rc != 0) {
|
|
msg->size = 0;
|
|
}
|
|
}
|
|
spdk_vhost_unlock();
|
|
|
|
return RTE_VHOST_MSG_RESULT_REPLY;
|
|
}
|
|
case VHOST_USER_SET_CONFIG: {
|
|
int rc = 0;
|
|
|
|
spdk_vhost_lock();
|
|
if (vsession->vdev->backend->vhost_set_config) {
|
|
rc = vsession->vdev->backend->vhost_set_config(vsession->vdev,
|
|
msg->payload.cfg.region, msg->payload.cfg.offset,
|
|
msg->payload.cfg.size, msg->payload.cfg.flags);
|
|
}
|
|
spdk_vhost_unlock();
|
|
|
|
return rc == 0 ? RTE_VHOST_MSG_RESULT_OK : RTE_VHOST_MSG_RESULT_ERR;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return RTE_VHOST_MSG_RESULT_NOT_HANDLED;
|
|
}
|
|
|
|
static enum rte_vhost_msg_result
|
|
extern_vhost_post_msg_handler(int vid, void *_msg)
|
|
{
|
|
struct vhost_user_msg *msg = _msg;
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
vsession = vhost_session_find_by_vid(vid);
|
|
if (vsession == NULL) {
|
|
SPDK_ERRLOG("Received a message to unitialized session (vid %d).\n", vid);
|
|
assert(false);
|
|
return RTE_VHOST_MSG_RESULT_ERR;
|
|
}
|
|
|
|
if (vsession->needs_restart) {
|
|
g_spdk_vhost_ops.new_device(vid);
|
|
vsession->needs_restart = false;
|
|
return RTE_VHOST_MSG_RESULT_NOT_HANDLED;
|
|
}
|
|
|
|
switch (msg->request) {
|
|
case VHOST_USER_SET_FEATURES:
|
|
/* rte_vhost requires all queues to be fully initialized in order
|
|
* to start I/O processing. This behavior is not compliant with the
|
|
* vhost-user specification and doesn't work with QEMU 2.12+, which
|
|
* will only initialize 1 I/O queue for the SeaBIOS boot.
|
|
* Theoretically, we should start polling each virtqueue individually
|
|
* after receiving its SET_VRING_KICK message, but rte_vhost is not
|
|
* designed to poll individual queues. So here we use a workaround
|
|
* to detect when the vhost session could be potentially at that SeaBIOS
|
|
* stage and we mark it to start polling as soon as its first virtqueue
|
|
* gets initialized. This doesn't hurt any non-QEMU vhost slaves
|
|
* and allows QEMU 2.12+ to boot correctly. SET_FEATURES could be sent
|
|
* at any time, but QEMU will send it at least once on SeaBIOS
|
|
* initialization - whenever powered-up or rebooted.
|
|
*/
|
|
vsession->forced_polling = true;
|
|
break;
|
|
case VHOST_USER_SET_VRING_KICK:
|
|
/* vhost-user spec tells us to start polling a queue after receiving
|
|
* its SET_VRING_KICK message. Let's do it!
|
|
*/
|
|
if (vsession->forced_polling && !vsession->started) {
|
|
g_spdk_vhost_ops.new_device(vid);
|
|
}
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return RTE_VHOST_MSG_RESULT_NOT_HANDLED;
|
|
}
|
|
|
|
struct rte_vhost_user_extern_ops g_spdk_extern_vhost_ops = {
|
|
.pre_msg_handle = extern_vhost_pre_msg_handler,
|
|
.post_msg_handle = extern_vhost_post_msg_handler,
|
|
};
|
|
|
|
void
|
|
vhost_session_install_rte_compat_hooks(struct spdk_vhost_session *vsession)
|
|
{
|
|
int rc;
|
|
|
|
rc = rte_vhost_extern_callback_register(vsession->vid, &g_spdk_extern_vhost_ops, NULL);
|
|
if (rc != 0) {
|
|
SPDK_ERRLOG("rte_vhost_extern_callback_register() failed for vid = %d\n",
|
|
vsession->vid);
|
|
return;
|
|
}
|
|
}
|
|
|
|
int
|
|
vhost_register_unix_socket(const char *path, const char *ctrl_name,
|
|
uint64_t virtio_features, uint64_t disabled_features, uint64_t protocol_features)
|
|
{
|
|
struct stat file_stat;
|
|
uint64_t features = 0;
|
|
|
|
/* Register vhost driver to handle vhost messages. */
|
|
if (stat(path, &file_stat) != -1) {
|
|
if (!S_ISSOCK(file_stat.st_mode)) {
|
|
SPDK_ERRLOG("Cannot create a domain socket at path \"%s\": "
|
|
"The file already exists and is not a socket.\n",
|
|
path);
|
|
return -EIO;
|
|
} else if (unlink(path) != 0) {
|
|
SPDK_ERRLOG("Cannot create a domain socket at path \"%s\": "
|
|
"The socket already exists and failed to unlink.\n",
|
|
path);
|
|
return -EIO;
|
|
}
|
|
}
|
|
|
|
#if RTE_VERSION < RTE_VERSION_NUM(20, 8, 0, 0)
|
|
if (rte_vhost_driver_register(path, 0) != 0) {
|
|
#else
|
|
if (rte_vhost_driver_register(path, RTE_VHOST_USER_ASYNC_COPY) != 0) {
|
|
#endif
|
|
SPDK_ERRLOG("Could not register controller %s with vhost library\n", ctrl_name);
|
|
SPDK_ERRLOG("Check if domain socket %s already exists\n", path);
|
|
return -EIO;
|
|
}
|
|
if (rte_vhost_driver_set_features(path, virtio_features) ||
|
|
rte_vhost_driver_disable_features(path, disabled_features)) {
|
|
SPDK_ERRLOG("Couldn't set vhost features for controller %s\n", ctrl_name);
|
|
|
|
rte_vhost_driver_unregister(path);
|
|
return -EIO;
|
|
}
|
|
|
|
if (rte_vhost_driver_callback_register(path, &g_spdk_vhost_ops) != 0) {
|
|
rte_vhost_driver_unregister(path);
|
|
SPDK_ERRLOG("Couldn't register callbacks for controller %s\n", ctrl_name);
|
|
return -EIO;
|
|
}
|
|
|
|
rte_vhost_driver_get_protocol_features(path, &features);
|
|
features |= protocol_features;
|
|
rte_vhost_driver_set_protocol_features(path, features);
|
|
|
|
if (rte_vhost_driver_start(path) != 0) {
|
|
SPDK_ERRLOG("Failed to start vhost driver for controller %s (%d): %s\n",
|
|
ctrl_name, errno, spdk_strerror(errno));
|
|
rte_vhost_driver_unregister(path);
|
|
return -EIO;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
vhost_get_mem_table(int vid, struct rte_vhost_memory **mem)
|
|
{
|
|
return rte_vhost_get_mem_table(vid, mem);
|
|
}
|
|
|
|
int
|
|
vhost_driver_unregister(const char *path)
|
|
{
|
|
return rte_vhost_driver_unregister(path);
|
|
}
|
|
|
|
int
|
|
vhost_get_negotiated_features(int vid, uint64_t *negotiated_features)
|
|
{
|
|
return rte_vhost_get_negotiated_features(vid, negotiated_features);
|
|
}
|
|
|
|
int
|
|
vhost_user_dev_set_coalescing(struct spdk_vhost_user_dev *user_dev, uint32_t delay_base_us,
|
|
uint32_t iops_threshold)
|
|
{
|
|
uint64_t delay_time_base = delay_base_us * spdk_get_ticks_hz() / 1000000ULL;
|
|
uint32_t io_rate = iops_threshold * SPDK_VHOST_STATS_CHECK_INTERVAL_MS / 1000U;
|
|
|
|
if (delay_time_base >= UINT32_MAX) {
|
|
SPDK_ERRLOG("Delay time of %"PRIu32" is to big\n", delay_base_us);
|
|
return -EINVAL;
|
|
} else if (io_rate == 0) {
|
|
SPDK_ERRLOG("IOPS rate of %"PRIu32" is too low. Min is %u\n", io_rate,
|
|
1000U / SPDK_VHOST_STATS_CHECK_INTERVAL_MS);
|
|
return -EINVAL;
|
|
}
|
|
|
|
user_dev->coalescing_delay_us = delay_base_us;
|
|
user_dev->coalescing_iops_threshold = iops_threshold;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
vhost_user_session_set_coalescing(struct spdk_vhost_dev *vdev,
|
|
struct spdk_vhost_session *vsession, void *ctx)
|
|
{
|
|
vsession->coalescing_delay_time_base =
|
|
to_user_dev(vdev)->coalescing_delay_us * spdk_get_ticks_hz() / 1000000ULL;
|
|
vsession->coalescing_io_rate_threshold =
|
|
to_user_dev(vdev)->coalescing_iops_threshold * SPDK_VHOST_STATS_CHECK_INTERVAL_MS / 1000U;
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
spdk_vhost_set_coalescing(struct spdk_vhost_dev *vdev, uint32_t delay_base_us,
|
|
uint32_t iops_threshold)
|
|
{
|
|
int rc;
|
|
|
|
rc = vhost_user_dev_set_coalescing(to_user_dev(vdev), delay_base_us, iops_threshold);
|
|
if (rc != 0) {
|
|
return rc;
|
|
}
|
|
|
|
vhost_user_dev_foreach_session(vdev, vhost_user_session_set_coalescing, NULL, NULL);
|
|
return 0;
|
|
}
|
|
|
|
void
|
|
spdk_vhost_get_coalescing(struct spdk_vhost_dev *vdev, uint32_t *delay_base_us,
|
|
uint32_t *iops_threshold)
|
|
{
|
|
struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
|
|
|
|
if (delay_base_us) {
|
|
*delay_base_us = user_dev->coalescing_delay_us;
|
|
}
|
|
|
|
if (iops_threshold) {
|
|
*iops_threshold = user_dev->coalescing_iops_threshold;
|
|
}
|
|
}
|
|
|
|
int
|
|
spdk_vhost_set_socket_path(const char *basename)
|
|
{
|
|
int ret;
|
|
|
|
if (basename && strlen(basename) > 0) {
|
|
ret = snprintf(g_vhost_user_dev_dirname, sizeof(g_vhost_user_dev_dirname) - 2, "%s", basename);
|
|
if (ret <= 0) {
|
|
return -EINVAL;
|
|
}
|
|
if ((size_t)ret >= sizeof(g_vhost_user_dev_dirname) - 2) {
|
|
SPDK_ERRLOG("Char dev dir path length %d is too long\n", ret);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (g_vhost_user_dev_dirname[ret - 1] != '/') {
|
|
g_vhost_user_dev_dirname[ret] = '/';
|
|
g_vhost_user_dev_dirname[ret + 1] = '\0';
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
vhost_dev_thread_exit(void *arg1)
|
|
{
|
|
spdk_thread_exit(spdk_get_thread());
|
|
}
|
|
|
|
int
|
|
vhost_user_dev_register(struct spdk_vhost_dev *vdev, const char *name, struct spdk_cpuset *cpumask,
|
|
const struct spdk_vhost_user_dev_backend *user_backend)
|
|
{
|
|
char path[PATH_MAX];
|
|
struct spdk_vhost_user_dev *user_dev;
|
|
|
|
if (snprintf(path, sizeof(path), "%s%s", g_vhost_user_dev_dirname, name) >= (int)sizeof(path)) {
|
|
SPDK_ERRLOG("Resulting socket path for controller %s is too long: %s%s\n",
|
|
name,g_vhost_user_dev_dirname, name);
|
|
return -EINVAL;
|
|
}
|
|
|
|
vdev->path = strdup(path);
|
|
if (vdev->path == NULL) {
|
|
return -EIO;
|
|
}
|
|
|
|
user_dev = calloc(1, sizeof(*user_dev));
|
|
if (user_dev == NULL) {
|
|
free(vdev->path);
|
|
return -ENOMEM;
|
|
}
|
|
vdev->ctxt = user_dev;
|
|
|
|
vdev->thread = spdk_thread_create(vdev->name, cpumask);
|
|
if (vdev->thread == NULL) {
|
|
free(user_dev);
|
|
free(vdev->path);
|
|
SPDK_ERRLOG("Failed to create thread for vhost controller %s.\n", name);
|
|
return -EIO;
|
|
}
|
|
|
|
vdev->registered = true;
|
|
user_dev->user_backend = user_backend;
|
|
user_dev->vdev = vdev;
|
|
TAILQ_INIT(&user_dev->vsessions);
|
|
|
|
vhost_user_dev_set_coalescing(user_dev, SPDK_VHOST_COALESCING_DELAY_BASE_US,
|
|
SPDK_VHOST_VQ_IOPS_COALESCING_THRESHOLD);
|
|
|
|
if (vhost_register_unix_socket(path, name, vdev->virtio_features, vdev->disabled_features,
|
|
vdev->protocol_features)) {
|
|
spdk_thread_send_msg(vdev->thread, vhost_dev_thread_exit, NULL);
|
|
free(user_dev);
|
|
free(vdev->path);
|
|
return -EIO;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
vhost_user_dev_unregister(struct spdk_vhost_dev *vdev)
|
|
{
|
|
struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
|
|
|
|
if (user_dev->pending_async_op_num) {
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (!TAILQ_EMPTY(&user_dev->vsessions)) {
|
|
SPDK_ERRLOG("Controller %s has still valid connection.\n", vdev->name);
|
|
return -EBUSY;
|
|
}
|
|
|
|
if (vdev->registered && vhost_driver_unregister(vdev->path) != 0) {
|
|
SPDK_ERRLOG("Could not unregister controller %s with vhost library\n"
|
|
"Check if domain socket %s still exists\n",
|
|
vdev->name, vdev->path);
|
|
return -EIO;
|
|
}
|
|
|
|
spdk_thread_send_msg(vdev->thread, vhost_dev_thread_exit, NULL);
|
|
free(user_dev);
|
|
free(vdev->path);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool g_vhost_user_started = false;
|
|
|
|
int
|
|
vhost_user_init(void)
|
|
{
|
|
size_t len;
|
|
|
|
if (g_vhost_user_started) {
|
|
return 0;
|
|
}
|
|
|
|
if (g_vhost_user_dev_dirname[0] == '\0') {
|
|
if (getcwd(g_vhost_user_dev_dirname, sizeof(g_vhost_user_dev_dirname) - 1) == NULL) {
|
|
SPDK_ERRLOG("getcwd failed (%d): %s\n", errno, spdk_strerror(errno));
|
|
return -1;
|
|
}
|
|
|
|
len = strlen(g_vhost_user_dev_dirname);
|
|
if (g_vhost_user_dev_dirname[len - 1] != '/') {
|
|
g_vhost_user_dev_dirname[len] = '/';
|
|
g_vhost_user_dev_dirname[len + 1] = '\0';
|
|
}
|
|
}
|
|
|
|
g_vhost_user_started = true;
|
|
|
|
g_vhost_user_init_thread = spdk_get_thread();
|
|
assert(g_vhost_user_init_thread != NULL);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
vhost_user_session_shutdown_on_init(void *vhost_cb)
|
|
{
|
|
spdk_vhost_fini_cb fn = vhost_cb;
|
|
|
|
fn();
|
|
}
|
|
|
|
static void *
|
|
vhost_user_session_shutdown(void *vhost_cb)
|
|
{
|
|
struct spdk_vhost_dev *vdev = NULL;
|
|
struct spdk_vhost_session *vsession;
|
|
|
|
for (vdev = spdk_vhost_dev_next(NULL); vdev != NULL;
|
|
vdev = spdk_vhost_dev_next(vdev)) {
|
|
spdk_vhost_lock();
|
|
TAILQ_FOREACH(vsession, &to_user_dev(vdev)->vsessions, tailq) {
|
|
if (vsession->started) {
|
|
_stop_session(vsession);
|
|
}
|
|
}
|
|
spdk_vhost_unlock();
|
|
vhost_driver_unregister(vdev->path);
|
|
vdev->registered = false;
|
|
}
|
|
|
|
SPDK_INFOLOG(vhost, "Exiting\n");
|
|
spdk_thread_send_msg(g_vhost_user_init_thread, vhost_user_session_shutdown_on_init, vhost_cb);
|
|
return NULL;
|
|
}
|
|
|
|
void
|
|
vhost_user_fini(spdk_vhost_fini_cb vhost_cb)
|
|
{
|
|
pthread_t tid;
|
|
int rc;
|
|
|
|
if (!g_vhost_user_started) {
|
|
vhost_cb();
|
|
return;
|
|
}
|
|
|
|
g_vhost_user_started = false;
|
|
|
|
/* rte_vhost API for removing sockets is not asynchronous. Since it may call SPDK
|
|
* ops for stopping a device or removing a connection, we need to call it from
|
|
* a separate thread to avoid deadlock.
|
|
*/
|
|
rc = pthread_create(&tid, NULL, &vhost_user_session_shutdown, vhost_cb);
|
|
if (rc < 0) {
|
|
SPDK_ERRLOG("Failed to start session shutdown thread (%d): %s\n", rc, spdk_strerror(rc));
|
|
abort();
|
|
}
|
|
pthread_detach(tid);
|
|
}
|