Many open source projects have moved to using SPDX identifiers to specify license information, reducing the amount of boilerplate code in every source file. This patch replaces the bulk of SPDK .c, .cpp and Makefiles with the BSD-3-Clause identifier. Almost all of these files share the exact same license text, and this patch only modifies the files that contain the most common license text. There can be slight variations because the third clause contains company names - most say "Intel Corporation", but there are instances for Nvidia, Samsung, Eideticom and even "the copyright holder". Used a bash script to automate replacement of the license text with SPDX identifier which is checked into scripts/spdx.sh. Signed-off-by: Jim Harris <james.r.harris@intel.com> Change-Id: Iaa88ab5e92ea471691dc298cfe41ebfb5d169780 Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/12904 Community-CI: Broadcom CI <spdk-ci.pdl@broadcom.com> Community-CI: Mellanox Build Bot Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Aleksey Marchuk <alexeymar@nvidia.com> Reviewed-by: Changpeng Liu <changpeng.liu@intel.com> Reviewed-by: Dong Yi <dongx.yi@intel.com> Reviewed-by: Konrad Sztyber <konrad.sztyber@intel.com> Reviewed-by: Paul Luse <paul.e.luse@intel.com> Reviewed-by: <qun.wan@intel.com>
536 lines
11 KiB
C
536 lines
11 KiB
C
/* SPDX-License-Identifier: BSD-3-Clause
|
|
* Copyright (c) Intel Corporation.
|
|
* All rights reserved.
|
|
*/
|
|
|
|
#include "spdk/stdinc.h"
|
|
#include "spdk/ftl.h"
|
|
#include "spdk/likely.h"
|
|
#include "spdk/util.h"
|
|
|
|
#include "ftl_io.h"
|
|
#include "ftl_core.h"
|
|
#include "ftl_band.h"
|
|
#include "ftl_debug.h"
|
|
|
|
void
|
|
ftl_io_inc_req(struct ftl_io *io)
|
|
{
|
|
struct ftl_band *band = io->band;
|
|
|
|
if (!(io->flags & FTL_IO_CACHE) && io->type != FTL_IO_READ && io->type != FTL_IO_ERASE) {
|
|
ftl_band_acquire_lba_map(band);
|
|
}
|
|
|
|
__atomic_fetch_add(&io->dev->num_inflight, 1, __ATOMIC_SEQ_CST);
|
|
|
|
++io->req_cnt;
|
|
}
|
|
|
|
void
|
|
ftl_io_dec_req(struct ftl_io *io)
|
|
{
|
|
struct ftl_band *band = io->band;
|
|
unsigned long num_inflight __attribute__((unused));
|
|
|
|
if (!(io->flags & FTL_IO_CACHE) && io->type != FTL_IO_READ && io->type != FTL_IO_ERASE) {
|
|
ftl_band_release_lba_map(band);
|
|
}
|
|
|
|
num_inflight = __atomic_fetch_sub(&io->dev->num_inflight, 1, __ATOMIC_SEQ_CST);
|
|
|
|
assert(num_inflight > 0);
|
|
assert(io->req_cnt > 0);
|
|
|
|
--io->req_cnt;
|
|
}
|
|
|
|
struct iovec *
|
|
ftl_io_iovec(struct ftl_io *io)
|
|
{
|
|
return &io->iov[0];
|
|
}
|
|
|
|
uint64_t
|
|
ftl_io_get_lba(const struct ftl_io *io, size_t offset)
|
|
{
|
|
assert(offset < io->num_blocks);
|
|
|
|
if (io->flags & FTL_IO_VECTOR_LBA) {
|
|
return io->lba.vector[offset];
|
|
} else {
|
|
return io->lba.single + offset;
|
|
}
|
|
}
|
|
|
|
uint64_t
|
|
ftl_io_current_lba(const struct ftl_io *io)
|
|
{
|
|
return ftl_io_get_lba(io, io->pos);
|
|
}
|
|
|
|
void
|
|
ftl_io_advance(struct ftl_io *io, size_t num_blocks)
|
|
{
|
|
struct iovec *iov = ftl_io_iovec(io);
|
|
size_t iov_blocks, block_left = num_blocks;
|
|
|
|
io->pos += num_blocks;
|
|
|
|
if (io->iov_cnt != 0) {
|
|
while (block_left > 0) {
|
|
assert(io->iov_pos < io->iov_cnt);
|
|
iov_blocks = iov[io->iov_pos].iov_len / FTL_BLOCK_SIZE;
|
|
|
|
if (io->iov_off + block_left < iov_blocks) {
|
|
io->iov_off += block_left;
|
|
break;
|
|
}
|
|
|
|
assert(iov_blocks > io->iov_off);
|
|
block_left -= (iov_blocks - io->iov_off);
|
|
io->iov_off = 0;
|
|
io->iov_pos++;
|
|
}
|
|
}
|
|
|
|
if (io->parent) {
|
|
ftl_io_advance(io->parent, num_blocks);
|
|
}
|
|
}
|
|
|
|
size_t
|
|
ftl_iovec_num_blocks(struct iovec *iov, size_t iov_cnt)
|
|
{
|
|
size_t num_blocks = 0, i = 0;
|
|
|
|
for (; i < iov_cnt; ++i) {
|
|
num_blocks += iov[i].iov_len / FTL_BLOCK_SIZE;
|
|
}
|
|
|
|
return num_blocks;
|
|
}
|
|
|
|
void *
|
|
ftl_io_iovec_addr(struct ftl_io *io)
|
|
{
|
|
assert(io->iov_pos < io->iov_cnt);
|
|
assert(io->iov_off * FTL_BLOCK_SIZE < ftl_io_iovec(io)[io->iov_pos].iov_len);
|
|
|
|
return (char *)ftl_io_iovec(io)[io->iov_pos].iov_base +
|
|
io->iov_off * FTL_BLOCK_SIZE;
|
|
}
|
|
|
|
size_t
|
|
ftl_io_iovec_len_left(struct ftl_io *io)
|
|
{
|
|
struct iovec *iov = ftl_io_iovec(io);
|
|
return iov[io->iov_pos].iov_len / FTL_BLOCK_SIZE - io->iov_off;
|
|
}
|
|
|
|
static void
|
|
ftl_io_init_iovec(struct ftl_io *io, const struct iovec *iov, size_t iov_cnt, size_t iov_off,
|
|
size_t num_blocks)
|
|
{
|
|
size_t offset = 0, num_left;
|
|
|
|
io->iov_pos = 0;
|
|
io->iov_cnt = 0;
|
|
io->num_blocks = num_blocks;
|
|
|
|
while (offset < num_blocks) {
|
|
assert(io->iov_cnt < FTL_IO_MAX_IOVEC && io->iov_cnt < iov_cnt);
|
|
|
|
num_left = spdk_min(iov[io->iov_cnt].iov_len / FTL_BLOCK_SIZE - iov_off,
|
|
num_blocks);
|
|
io->iov[io->iov_cnt].iov_base = (char *)iov[io->iov_cnt].iov_base +
|
|
iov_off * FTL_BLOCK_SIZE;
|
|
io->iov[io->iov_cnt].iov_len = num_left * FTL_BLOCK_SIZE;
|
|
|
|
offset += num_left;
|
|
io->iov_cnt++;
|
|
iov_off = 0;
|
|
}
|
|
}
|
|
|
|
void
|
|
ftl_io_shrink_iovec(struct ftl_io *io, size_t num_blocks)
|
|
{
|
|
size_t iov_off = 0, block_off = 0;
|
|
|
|
assert(io->num_blocks >= num_blocks);
|
|
assert(io->pos == 0 && io->iov_pos == 0 && io->iov_off == 0);
|
|
|
|
for (; iov_off < io->iov_cnt; ++iov_off) {
|
|
size_t num_iov = io->iov[iov_off].iov_len / FTL_BLOCK_SIZE;
|
|
size_t num_left = num_blocks - block_off;
|
|
|
|
if (num_iov >= num_left) {
|
|
io->iov[iov_off].iov_len = num_left * FTL_BLOCK_SIZE;
|
|
io->iov_cnt = iov_off + 1;
|
|
io->num_blocks = num_blocks;
|
|
break;
|
|
}
|
|
|
|
block_off += num_iov;
|
|
}
|
|
}
|
|
|
|
static void
|
|
ftl_io_init(struct ftl_io *io, struct spdk_ftl_dev *dev,
|
|
ftl_io_fn fn, void *ctx, int flags, int type)
|
|
{
|
|
io->flags |= flags | FTL_IO_INITIALIZED;
|
|
io->type = type;
|
|
io->dev = dev;
|
|
io->lba.single = FTL_LBA_INVALID;
|
|
io->addr.offset = FTL_ADDR_INVALID;
|
|
io->cb_fn = fn;
|
|
io->cb_ctx = ctx;
|
|
io->trace = ftl_trace_alloc_id(dev);
|
|
}
|
|
|
|
struct ftl_io *
|
|
ftl_io_init_internal(const struct ftl_io_init_opts *opts)
|
|
{
|
|
struct ftl_io *io = opts->io;
|
|
struct ftl_io *parent = opts->parent;
|
|
struct spdk_ftl_dev *dev = opts->dev;
|
|
const struct iovec *iov;
|
|
size_t iov_cnt, iov_off;
|
|
|
|
if (!io) {
|
|
if (parent) {
|
|
io = ftl_io_alloc_child(parent);
|
|
} else {
|
|
io = ftl_io_alloc(ftl_get_io_channel(dev));
|
|
}
|
|
|
|
if (!io) {
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
ftl_io_clear(io);
|
|
ftl_io_init(io, dev, opts->cb_fn, opts->cb_ctx, opts->flags | FTL_IO_INTERNAL, opts->type);
|
|
|
|
io->batch = opts->batch;
|
|
io->band = opts->band;
|
|
io->md = opts->md;
|
|
io->iov = &io->iov_buf[0];
|
|
|
|
if (parent) {
|
|
if (parent->flags & FTL_IO_VECTOR_LBA) {
|
|
io->lba.vector = parent->lba.vector + parent->pos;
|
|
} else {
|
|
io->lba.single = parent->lba.single + parent->pos;
|
|
}
|
|
|
|
iov = &parent->iov[parent->iov_pos];
|
|
iov_cnt = parent->iov_cnt - parent->iov_pos;
|
|
iov_off = parent->iov_off;
|
|
} else {
|
|
iov = &opts->iovs[0];
|
|
iov_cnt = opts->iovcnt;
|
|
iov_off = 0;
|
|
}
|
|
|
|
/* Some requests (zone resets) do not use iovecs */
|
|
if (iov_cnt > 0) {
|
|
ftl_io_init_iovec(io, iov, iov_cnt, iov_off, opts->num_blocks);
|
|
}
|
|
|
|
if (opts->flags & FTL_IO_VECTOR_LBA) {
|
|
io->lba.vector = calloc(io->num_blocks, sizeof(uint64_t));
|
|
if (!io->lba.vector) {
|
|
ftl_io_free(io);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
return io;
|
|
}
|
|
|
|
struct ftl_io *
|
|
ftl_io_wbuf_init(struct spdk_ftl_dev *dev, struct ftl_addr addr, struct ftl_band *band,
|
|
struct ftl_batch *batch, ftl_io_fn cb)
|
|
{
|
|
struct ftl_io *io;
|
|
struct ftl_io_init_opts opts = {
|
|
.dev = dev,
|
|
.io = NULL,
|
|
.batch = batch,
|
|
.band = band,
|
|
.size = sizeof(struct ftl_io),
|
|
.flags = 0,
|
|
.type = FTL_IO_WRITE,
|
|
.num_blocks = dev->xfer_size,
|
|
.cb_fn = cb,
|
|
.iovcnt = dev->xfer_size,
|
|
.md = batch->metadata,
|
|
};
|
|
|
|
memcpy(opts.iovs, batch->iov, sizeof(struct iovec) * dev->xfer_size);
|
|
|
|
io = ftl_io_init_internal(&opts);
|
|
if (!io) {
|
|
return NULL;
|
|
}
|
|
|
|
io->addr = addr;
|
|
|
|
return io;
|
|
}
|
|
|
|
struct ftl_io *
|
|
ftl_io_erase_init(struct ftl_band *band, size_t num_blocks, ftl_io_fn cb)
|
|
{
|
|
struct ftl_io *io;
|
|
struct ftl_io_init_opts opts = {
|
|
.dev = band->dev,
|
|
.io = NULL,
|
|
.band = band,
|
|
.size = sizeof(struct ftl_io),
|
|
.flags = FTL_IO_PHYSICAL_MODE,
|
|
.type = FTL_IO_ERASE,
|
|
.num_blocks = 1,
|
|
.cb_fn = cb,
|
|
.iovcnt = 0,
|
|
.md = NULL,
|
|
};
|
|
|
|
io = ftl_io_init_internal(&opts);
|
|
if (!io) {
|
|
return NULL;
|
|
}
|
|
|
|
io->num_blocks = num_blocks;
|
|
|
|
return io;
|
|
}
|
|
|
|
static void
|
|
_ftl_user_cb(struct ftl_io *io, void *arg, int status)
|
|
{
|
|
io->user_fn(arg, status);
|
|
}
|
|
|
|
struct ftl_io *
|
|
ftl_io_user_init(struct spdk_io_channel *_ioch, uint64_t lba, size_t num_blocks, struct iovec *iov,
|
|
size_t iov_cnt, spdk_ftl_fn cb_fn, void *cb_ctx, int type)
|
|
{
|
|
struct ftl_io_channel *ioch = ftl_io_channel_get_ctx(_ioch);
|
|
struct spdk_ftl_dev *dev = ioch->dev;
|
|
struct ftl_io *io;
|
|
|
|
io = ftl_io_alloc(_ioch);
|
|
if (spdk_unlikely(!io)) {
|
|
return NULL;
|
|
}
|
|
|
|
ftl_io_init(io, dev, _ftl_user_cb, cb_ctx, 0, type);
|
|
io->lba.single = lba;
|
|
io->user_fn = cb_fn;
|
|
io->iov = iov;
|
|
io->iov_cnt = iov_cnt;
|
|
io->num_blocks = num_blocks;
|
|
|
|
ftl_trace_lba_io_init(io->dev, io);
|
|
return io;
|
|
}
|
|
|
|
static void
|
|
_ftl_io_free(struct ftl_io *io)
|
|
{
|
|
struct ftl_io_channel *ioch;
|
|
|
|
assert(LIST_EMPTY(&io->children));
|
|
|
|
if (io->flags & FTL_IO_VECTOR_LBA) {
|
|
free(io->lba.vector);
|
|
}
|
|
|
|
if (pthread_spin_destroy(&io->lock)) {
|
|
SPDK_ERRLOG("pthread_spin_destroy failed\n");
|
|
}
|
|
|
|
ioch = ftl_io_channel_get_ctx(io->ioch);
|
|
spdk_mempool_put(ioch->io_pool, io);
|
|
}
|
|
|
|
static bool
|
|
ftl_io_remove_child(struct ftl_io *io)
|
|
{
|
|
struct ftl_io *parent = io->parent;
|
|
bool parent_done;
|
|
|
|
pthread_spin_lock(&parent->lock);
|
|
LIST_REMOVE(io, child_entry);
|
|
parent_done = parent->done && LIST_EMPTY(&parent->children);
|
|
parent->status = parent->status ? : io->status;
|
|
pthread_spin_unlock(&parent->lock);
|
|
|
|
return parent_done;
|
|
}
|
|
|
|
void
|
|
ftl_io_complete(struct ftl_io *io)
|
|
{
|
|
struct ftl_io *parent = io->parent;
|
|
bool complete;
|
|
|
|
io->flags &= ~FTL_IO_INITIALIZED;
|
|
|
|
pthread_spin_lock(&io->lock);
|
|
complete = LIST_EMPTY(&io->children);
|
|
io->done = true;
|
|
pthread_spin_unlock(&io->lock);
|
|
|
|
if (complete) {
|
|
if (io->cb_fn) {
|
|
io->cb_fn(io, io->cb_ctx, io->status);
|
|
}
|
|
|
|
if (parent && ftl_io_remove_child(io)) {
|
|
ftl_io_complete(parent);
|
|
}
|
|
|
|
_ftl_io_free(io);
|
|
}
|
|
}
|
|
|
|
struct ftl_io *
|
|
ftl_io_alloc_child(struct ftl_io *parent)
|
|
{
|
|
struct ftl_io *io;
|
|
|
|
io = ftl_io_alloc(parent->ioch);
|
|
if (spdk_unlikely(!io)) {
|
|
return NULL;
|
|
}
|
|
|
|
ftl_io_init(io, parent->dev, NULL, NULL, parent->flags, parent->type);
|
|
io->parent = parent;
|
|
|
|
pthread_spin_lock(&parent->lock);
|
|
LIST_INSERT_HEAD(&parent->children, io, child_entry);
|
|
pthread_spin_unlock(&parent->lock);
|
|
|
|
return io;
|
|
}
|
|
|
|
void ftl_io_fail(struct ftl_io *io, int status)
|
|
{
|
|
io->status = status;
|
|
ftl_io_advance(io, io->num_blocks - io->pos);
|
|
}
|
|
|
|
void *
|
|
ftl_io_get_md(const struct ftl_io *io)
|
|
{
|
|
if (!io->md) {
|
|
return NULL;
|
|
}
|
|
|
|
return (char *)io->md + io->pos * io->dev->md_size;
|
|
}
|
|
|
|
struct ftl_io *
|
|
ftl_io_alloc(struct spdk_io_channel *ch)
|
|
{
|
|
struct ftl_io *io;
|
|
struct ftl_io_channel *ioch = ftl_io_channel_get_ctx(ch);
|
|
|
|
io = spdk_mempool_get(ioch->io_pool);
|
|
if (!io) {
|
|
return NULL;
|
|
}
|
|
|
|
memset(io, 0, ioch->elem_size);
|
|
io->ioch = ch;
|
|
|
|
if (pthread_spin_init(&io->lock, PTHREAD_PROCESS_PRIVATE)) {
|
|
SPDK_ERRLOG("pthread_spin_init failed\n");
|
|
spdk_mempool_put(ioch->io_pool, io);
|
|
return NULL;
|
|
}
|
|
|
|
return io;
|
|
}
|
|
|
|
void
|
|
ftl_io_reinit(struct ftl_io *io, ftl_io_fn cb, void *ctx, int flags, int type)
|
|
{
|
|
ftl_io_clear(io);
|
|
ftl_io_init(io, io->dev, cb, ctx, flags, type);
|
|
}
|
|
|
|
void
|
|
ftl_io_clear(struct ftl_io *io)
|
|
{
|
|
ftl_io_reset(io);
|
|
|
|
io->flags = 0;
|
|
io->batch = NULL;
|
|
io->band = NULL;
|
|
}
|
|
|
|
void
|
|
ftl_io_reset(struct ftl_io *io)
|
|
{
|
|
io->req_cnt = io->pos = io->iov_pos = io->iov_off = 0;
|
|
io->done = false;
|
|
}
|
|
|
|
void
|
|
ftl_io_free(struct ftl_io *io)
|
|
{
|
|
struct ftl_io *parent;
|
|
|
|
if (!io) {
|
|
return;
|
|
}
|
|
|
|
parent = io->parent;
|
|
if (parent && ftl_io_remove_child(io)) {
|
|
ftl_io_complete(parent);
|
|
}
|
|
|
|
_ftl_io_free(io);
|
|
}
|
|
|
|
void
|
|
ftl_io_call_foreach_child(struct ftl_io *io, int (*callback)(struct ftl_io *))
|
|
{
|
|
struct ftl_io *child, *tmp;
|
|
|
|
assert(!io->done);
|
|
|
|
/*
|
|
* If the IO doesn't have any children, it means that it directly describes a request (i.e.
|
|
* all of the buffers, LBAs, etc. are filled). Otherwise the IO only groups together several
|
|
* requests and may be partially filled, so the callback needs to be called on all of its
|
|
* children instead.
|
|
*/
|
|
if (LIST_EMPTY(&io->children)) {
|
|
callback(io);
|
|
return;
|
|
}
|
|
|
|
LIST_FOREACH_SAFE(child, &io->children, child_entry, tmp) {
|
|
int rc = callback(child);
|
|
if (rc) {
|
|
assert(rc != -EAGAIN);
|
|
ftl_io_fail(io, rc);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If all the callbacks were processed or an error occurred, treat this IO as completed.
|
|
* Multiple calls to ftl_io_call_foreach_child are not supported, resubmissions are supposed
|
|
* to be handled in the callback.
|
|
*/
|
|
ftl_io_complete(io);
|
|
}
|