2018-02-16 18:14:32 +00:00
|
|
|
/*-
|
|
|
|
* BSD LICENSE
|
|
|
|
*
|
2020-01-28 11:54:47 +00:00
|
|
|
* Copyright (c) Intel Corporation. All rights reserved.
|
|
|
|
* Copyright (c) 2020 Mellanox Technologies LTD. All rights reserved.
|
2018-02-16 18:14:32 +00:00
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
*
|
|
|
|
* * Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* * Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in
|
|
|
|
* the documentation and/or other materials provided with the
|
|
|
|
* distribution.
|
|
|
|
* * Neither the name of Intel Corporation nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived
|
|
|
|
* from this software without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/** \file
|
|
|
|
* TCP network implementation abstraction layer
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef SPDK_INTERNAL_SOCK_H
|
|
|
|
#define SPDK_INTERNAL_SOCK_H
|
|
|
|
|
|
|
|
#include "spdk/stdinc.h"
|
|
|
|
#include "spdk/sock.h"
|
|
|
|
#include "spdk/queue.h"
|
|
|
|
|
|
|
|
#ifdef __cplusplus
|
|
|
|
extern "C" {
|
|
|
|
#endif
|
|
|
|
|
2018-03-12 15:42:04 +00:00
|
|
|
#define MAX_EVENTS_PER_POLL 32
|
2020-01-07 14:23:16 +00:00
|
|
|
#define DEFAULT_SOCK_PRIORITY 0
|
2020-05-29 12:42:38 +00:00
|
|
|
#define MIN_SOCK_PIPE_SIZE 1024
|
2020-09-10 14:57:52 +00:00
|
|
|
#define MIN_SO_RCVBUF_SIZE (2 * 1024 * 1024)
|
|
|
|
#define MIN_SO_SNDBUF_SIZE (2 * 1024 * 1024)
|
2018-03-12 15:42:04 +00:00
|
|
|
|
|
|
|
struct spdk_sock {
|
2019-09-30 19:38:26 +00:00
|
|
|
struct spdk_net_impl *net_impl;
|
2020-02-19 11:18:51 +00:00
|
|
|
struct spdk_sock_opts opts;
|
2019-09-30 19:38:26 +00:00
|
|
|
struct spdk_sock_group_impl *group_impl;
|
|
|
|
TAILQ_ENTRY(spdk_sock) link;
|
2019-08-26 22:03:07 +00:00
|
|
|
|
|
|
|
TAILQ_HEAD(, spdk_sock_request) queued_reqs;
|
2019-10-18 17:50:27 +00:00
|
|
|
TAILQ_HEAD(, spdk_sock_request) pending_reqs;
|
2019-08-26 22:03:07 +00:00
|
|
|
int queued_iovcnt;
|
2020-09-01 11:03:34 +00:00
|
|
|
int cb_cnt;
|
|
|
|
spdk_sock_cb cb_fn;
|
|
|
|
void *cb_arg;
|
2020-08-31 16:47:15 +00:00
|
|
|
int placement_id;
|
2019-08-26 22:03:07 +00:00
|
|
|
struct {
|
|
|
|
uint8_t closed : 1;
|
|
|
|
uint8_t reserved : 7;
|
|
|
|
} flags;
|
2018-03-12 15:42:04 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
struct spdk_sock_group {
|
|
|
|
STAILQ_HEAD(, spdk_sock_group_impl) group_impls;
|
2019-05-14 18:40:20 +00:00
|
|
|
void *ctx;
|
2018-03-12 15:42:04 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
struct spdk_sock_group_impl {
|
|
|
|
struct spdk_net_impl *net_impl;
|
|
|
|
TAILQ_HEAD(, spdk_sock) socks;
|
|
|
|
STAILQ_ENTRY(spdk_sock_group_impl) link;
|
sock: keep track of removed sockets during call to poll
We have been intermittently hitting the assert where
we check sock->cb_fn != NULL in spdk_sock_group_impl_poll_count.
The only way we could be hitting this specific error is if we
wereremoving a socket from a sock group within after receiving
an event for it.
Specifically, we are seeing this error on the NVMe-oF TCP target
which relies on posix sockets using epoll.
The man page for epoll states the following:
If you use an event cache or store all the file descriptors
returned from epoll_wait(2), then make sure to provide
a way to mark its closure dynamically (i.e., caused by
a previous event's processing). Suppose you receive 100 events
from epoll_wait(2), and in event #47 a condition causes event
#13 to be closed. If you remove the structure and close(2)
the file descriptor for event #13, then your event cache might
still say there are events waiting for that file descriptor
causing confusion.
One solution for this is to call, during the processing
of event 47, epoll_ctl(EPOLL_CTL_DEL) to delete file
descriptor 13 and close(2), then mark its associated data
structure as removed and link it to a cleanup list. If
you find another event for file descriptor 13 in your batch
processing, you will discover the file descriptor had
been previously removed and there will be no confusion.
Since we do store all of the file descriptors returned from
epoll_wait, we need to implement the tracking mentioned above.
fixes issue #1294
Signed-off-by: Seth Howell <seth.howell@intel.com>
Change-Id: Ib592ce19e3f0b691e3a825d02ebb42d7338e3ceb
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/1589
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com>
Reviewed-by: Aleksey Marchuk <alexeymar@mellanox.com>
2020-03-30 21:54:02 +00:00
|
|
|
/* List of removed sockets. refreshed each time we poll the sock group. */
|
|
|
|
int num_removed_socks;
|
|
|
|
/* Unfortunately, we can't just keep a tailq of the sockets in case they are freed
|
|
|
|
* or added to another poll group later.
|
|
|
|
*/
|
|
|
|
uintptr_t removed_socks[MAX_EVENTS_PER_POLL];
|
2018-03-12 15:42:04 +00:00
|
|
|
};
|
2018-02-16 18:14:32 +00:00
|
|
|
|
|
|
|
struct spdk_net_impl {
|
|
|
|
const char *name;
|
2020-01-07 14:23:16 +00:00
|
|
|
int priority;
|
2018-02-16 18:14:32 +00:00
|
|
|
|
2018-10-08 01:51:03 +00:00
|
|
|
int (*getaddr)(struct spdk_sock *sock, char *saddr, int slen, uint16_t *sport, char *caddr,
|
|
|
|
int clen, uint16_t *cport);
|
2020-02-19 11:18:51 +00:00
|
|
|
struct spdk_sock *(*connect)(const char *ip, int port, struct spdk_sock_opts *opts);
|
|
|
|
struct spdk_sock *(*listen)(const char *ip, int port, struct spdk_sock_opts *opts);
|
2018-02-16 18:14:32 +00:00
|
|
|
struct spdk_sock *(*accept)(struct spdk_sock *sock);
|
|
|
|
int (*close)(struct spdk_sock *sock);
|
|
|
|
ssize_t (*recv)(struct spdk_sock *sock, void *buf, size_t len);
|
2019-02-26 23:58:24 +00:00
|
|
|
ssize_t (*readv)(struct spdk_sock *sock, struct iovec *iov, int iovcnt);
|
2018-02-16 18:14:32 +00:00
|
|
|
ssize_t (*writev)(struct spdk_sock *sock, struct iovec *iov, int iovcnt);
|
|
|
|
|
2019-08-26 22:03:07 +00:00
|
|
|
void (*writev_async)(struct spdk_sock *sock, struct spdk_sock_request *req);
|
2019-11-21 15:16:18 +00:00
|
|
|
int (*flush)(struct spdk_sock *sock);
|
2019-08-26 22:03:07 +00:00
|
|
|
|
2018-02-16 18:14:32 +00:00
|
|
|
int (*set_recvlowat)(struct spdk_sock *sock, int nbytes);
|
|
|
|
int (*set_recvbuf)(struct spdk_sock *sock, int sz);
|
|
|
|
int (*set_sendbuf)(struct spdk_sock *sock, int sz);
|
|
|
|
|
|
|
|
bool (*is_ipv6)(struct spdk_sock *sock);
|
|
|
|
bool (*is_ipv4)(struct spdk_sock *sock);
|
2019-10-04 17:48:32 +00:00
|
|
|
bool (*is_connected)(struct spdk_sock *sock);
|
2018-02-16 18:14:32 +00:00
|
|
|
|
2019-05-14 12:50:15 +00:00
|
|
|
int (*get_placement_id)(struct spdk_sock *sock, int *placement_id);
|
2018-02-19 04:52:43 +00:00
|
|
|
struct spdk_sock_group_impl *(*group_impl_create)(void);
|
|
|
|
int (*group_impl_add_sock)(struct spdk_sock_group_impl *group, struct spdk_sock *sock);
|
|
|
|
int (*group_impl_remove_sock)(struct spdk_sock_group_impl *group, struct spdk_sock *sock);
|
|
|
|
int (*group_impl_poll)(struct spdk_sock_group_impl *group, int max_events,
|
|
|
|
struct spdk_sock **socks);
|
|
|
|
int (*group_impl_close)(struct spdk_sock_group_impl *group);
|
2018-02-16 18:14:32 +00:00
|
|
|
|
2020-01-28 11:54:47 +00:00
|
|
|
int (*get_opts)(struct spdk_sock_impl_opts *opts, size_t *len);
|
|
|
|
int (*set_opts)(const struct spdk_sock_impl_opts *opts, size_t len);
|
|
|
|
|
2018-02-16 18:14:32 +00:00
|
|
|
STAILQ_ENTRY(spdk_net_impl) link;
|
|
|
|
};
|
|
|
|
|
2020-01-07 14:23:16 +00:00
|
|
|
void spdk_net_impl_register(struct spdk_net_impl *impl, int priority);
|
2018-02-16 18:14:32 +00:00
|
|
|
|
2020-01-07 14:23:16 +00:00
|
|
|
#define SPDK_NET_IMPL_REGISTER(name, impl, priority) \
|
2018-02-16 18:14:32 +00:00
|
|
|
static void __attribute__((constructor)) net_impl_register_##name(void) \
|
|
|
|
{ \
|
2020-01-07 14:23:16 +00:00
|
|
|
spdk_net_impl_register(impl, priority); \
|
2018-02-16 18:14:32 +00:00
|
|
|
}
|
|
|
|
|
2019-08-26 22:03:07 +00:00
|
|
|
static inline void
|
|
|
|
spdk_sock_request_queue(struct spdk_sock *sock, struct spdk_sock_request *req)
|
|
|
|
{
|
|
|
|
TAILQ_INSERT_TAIL(&sock->queued_reqs, req, internal.link);
|
|
|
|
sock->queued_iovcnt += req->iovcnt;
|
|
|
|
}
|
|
|
|
|
2019-10-18 17:50:27 +00:00
|
|
|
static inline void
|
|
|
|
spdk_sock_request_pend(struct spdk_sock *sock, struct spdk_sock_request *req)
|
|
|
|
{
|
|
|
|
TAILQ_REMOVE(&sock->queued_reqs, req, internal.link);
|
|
|
|
assert(sock->queued_iovcnt >= req->iovcnt);
|
|
|
|
sock->queued_iovcnt -= req->iovcnt;
|
|
|
|
TAILQ_INSERT_TAIL(&sock->pending_reqs, req, internal.link);
|
|
|
|
}
|
|
|
|
|
2019-08-26 22:03:07 +00:00
|
|
|
static inline int
|
|
|
|
spdk_sock_request_put(struct spdk_sock *sock, struct spdk_sock_request *req, int err)
|
|
|
|
{
|
|
|
|
bool closed;
|
|
|
|
int rc = 0;
|
|
|
|
|
2019-10-18 17:50:27 +00:00
|
|
|
TAILQ_REMOVE(&sock->pending_reqs, req, internal.link);
|
2019-08-26 22:03:07 +00:00
|
|
|
|
2019-10-09 21:06:15 +00:00
|
|
|
req->internal.offset = 0;
|
|
|
|
|
2019-08-26 22:03:07 +00:00
|
|
|
closed = sock->flags.closed;
|
|
|
|
sock->cb_cnt++;
|
|
|
|
req->cb_fn(req->cb_arg, err);
|
|
|
|
assert(sock->cb_cnt > 0);
|
|
|
|
sock->cb_cnt--;
|
|
|
|
|
|
|
|
if (sock->cb_cnt == 0 && !closed && sock->flags.closed) {
|
|
|
|
/* The user closed the socket in response to a callback above. */
|
|
|
|
rc = -1;
|
|
|
|
spdk_sock_close(&sock);
|
|
|
|
}
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
spdk_sock_abort_requests(struct spdk_sock *sock)
|
|
|
|
{
|
|
|
|
struct spdk_sock_request *req;
|
|
|
|
bool closed;
|
|
|
|
int rc = 0;
|
|
|
|
|
|
|
|
closed = sock->flags.closed;
|
|
|
|
sock->cb_cnt++;
|
|
|
|
|
2019-10-18 17:50:27 +00:00
|
|
|
req = TAILQ_FIRST(&sock->pending_reqs);
|
|
|
|
while (req) {
|
|
|
|
TAILQ_REMOVE(&sock->pending_reqs, req, internal.link);
|
|
|
|
|
|
|
|
req->cb_fn(req->cb_arg, -ECANCELED);
|
|
|
|
|
|
|
|
req = TAILQ_FIRST(&sock->pending_reqs);
|
|
|
|
}
|
|
|
|
|
2019-08-26 22:03:07 +00:00
|
|
|
req = TAILQ_FIRST(&sock->queued_reqs);
|
|
|
|
while (req) {
|
|
|
|
TAILQ_REMOVE(&sock->queued_reqs, req, internal.link);
|
|
|
|
|
|
|
|
assert(sock->queued_iovcnt >= req->iovcnt);
|
|
|
|
sock->queued_iovcnt -= req->iovcnt;
|
|
|
|
|
|
|
|
req->cb_fn(req->cb_arg, -ECANCELED);
|
|
|
|
|
|
|
|
req = TAILQ_FIRST(&sock->queued_reqs);
|
|
|
|
}
|
|
|
|
assert(sock->cb_cnt > 0);
|
|
|
|
sock->cb_cnt--;
|
|
|
|
|
|
|
|
assert(TAILQ_EMPTY(&sock->queued_reqs));
|
2019-10-18 17:50:27 +00:00
|
|
|
assert(TAILQ_EMPTY(&sock->pending_reqs));
|
2019-08-26 22:03:07 +00:00
|
|
|
|
|
|
|
if (sock->cb_cnt == 0 && !closed && sock->flags.closed) {
|
|
|
|
/* The user closed the socket in response to a callback above. */
|
|
|
|
rc = -1;
|
|
|
|
spdk_sock_close(&sock);
|
|
|
|
}
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2018-02-16 18:14:32 +00:00
|
|
|
#ifdef __cplusplus
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#endif /* SPDK_INTERNAL_SOCK_H */
|