2022-06-21 11:04:00 +00:00
|
|
|
/* SPDX-License-Identifier: BSD-3-Clause
|
2022-11-01 20:26:26 +00:00
|
|
|
* Copyright (C) 2018 Intel Corporation.
|
2022-06-21 11:04:00 +00:00
|
|
|
* All rights reserved.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "spdk/stdinc.h"
|
|
|
|
#include "spdk/ftl.h"
|
|
|
|
#include "spdk/likely.h"
|
|
|
|
#include "spdk/util.h"
|
|
|
|
|
|
|
|
#include "ftl_io.h"
|
|
|
|
#include "ftl_core.h"
|
2022-06-10 11:40:26 +00:00
|
|
|
#include "ftl_band.h"
|
2022-06-21 11:04:00 +00:00
|
|
|
#include "ftl_debug.h"
|
|
|
|
|
|
|
|
void
|
|
|
|
ftl_io_inc_req(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
io->dev->num_inflight++;
|
|
|
|
io->req_cnt++;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
ftl_io_dec_req(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
assert(io->dev->num_inflight > 0);
|
|
|
|
assert(io->req_cnt > 0);
|
|
|
|
|
|
|
|
io->dev->num_inflight--;
|
|
|
|
io->req_cnt--;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct iovec *
|
|
|
|
ftl_io_iovec(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
return &io->iov[0];
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
ftl_io_get_lba(const struct ftl_io *io, size_t offset)
|
|
|
|
{
|
|
|
|
assert(offset < io->num_blocks);
|
|
|
|
return io->lba + offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t
|
|
|
|
ftl_io_current_lba(const struct ftl_io *io)
|
|
|
|
{
|
|
|
|
return ftl_io_get_lba(io, io->pos);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
ftl_io_advance(struct ftl_io *io, size_t num_blocks)
|
|
|
|
{
|
|
|
|
struct iovec *iov = ftl_io_iovec(io);
|
|
|
|
size_t iov_blocks, block_left = num_blocks;
|
|
|
|
|
|
|
|
io->pos += num_blocks;
|
|
|
|
|
|
|
|
if (io->iov_cnt == 0) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
while (block_left > 0) {
|
|
|
|
assert(io->iov_pos < io->iov_cnt);
|
|
|
|
iov_blocks = iov[io->iov_pos].iov_len / FTL_BLOCK_SIZE;
|
|
|
|
|
|
|
|
if (io->iov_off + block_left < iov_blocks) {
|
|
|
|
io->iov_off += block_left;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(iov_blocks > io->iov_off);
|
|
|
|
block_left -= (iov_blocks - io->iov_off);
|
|
|
|
io->iov_off = 0;
|
|
|
|
io->iov_pos++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
size_t
|
|
|
|
ftl_iovec_num_blocks(struct iovec *iov, size_t iov_cnt)
|
|
|
|
{
|
|
|
|
size_t num_blocks = 0, i = 0;
|
|
|
|
|
|
|
|
for (; i < iov_cnt; ++i) {
|
|
|
|
if (iov[i].iov_len & (FTL_BLOCK_SIZE - 1)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
num_blocks += iov[i].iov_len / FTL_BLOCK_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
return num_blocks;
|
|
|
|
}
|
|
|
|
|
|
|
|
void *
|
|
|
|
ftl_io_iovec_addr(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
assert(io->iov_pos < io->iov_cnt);
|
|
|
|
assert(io->iov_off * FTL_BLOCK_SIZE < ftl_io_iovec(io)[io->iov_pos].iov_len);
|
|
|
|
|
|
|
|
return (char *)ftl_io_iovec(io)[io->iov_pos].iov_base +
|
|
|
|
io->iov_off * FTL_BLOCK_SIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
size_t
|
|
|
|
ftl_io_iovec_len_left(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
if (io->iov_pos < io->iov_cnt) {
|
|
|
|
struct iovec *iov = ftl_io_iovec(io);
|
|
|
|
return iov[io->iov_pos].iov_len / FTL_BLOCK_SIZE - io->iov_off;
|
|
|
|
} else {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
ftl_io_cb(struct ftl_io *io, void *arg, int status)
|
|
|
|
{
|
2022-06-21 11:03:31 +00:00
|
|
|
struct ftl_io_channel *ioch = ftl_io_channel_get_ctx(io->ioch);
|
2022-08-03 04:06:48 +00:00
|
|
|
size_t result __attribute__((unused));
|
2022-06-21 11:03:31 +00:00
|
|
|
|
2022-06-21 11:04:00 +00:00
|
|
|
if (spdk_unlikely(status)) {
|
|
|
|
io->status = status;
|
|
|
|
|
|
|
|
if (-EAGAIN == status) {
|
|
|
|
/* IO has to be rescheduled again */
|
|
|
|
switch (io->type) {
|
|
|
|
case FTL_IO_READ:
|
|
|
|
ftl_io_clear(io);
|
|
|
|
TAILQ_INSERT_HEAD(&io->dev->rd_sq, io, queue_entry);
|
|
|
|
break;
|
|
|
|
case FTL_IO_WRITE:
|
|
|
|
ftl_io_clear(io);
|
|
|
|
TAILQ_INSERT_HEAD(&io->dev->wr_sq, io, queue_entry);
|
|
|
|
break;
|
2022-05-27 10:31:42 +00:00
|
|
|
case FTL_IO_UNMAP:
|
|
|
|
ftl_io_clear(io);
|
|
|
|
TAILQ_INSERT_HEAD(&io->dev->unmap_sq, io, queue_entry);
|
|
|
|
break;
|
2022-06-21 11:04:00 +00:00
|
|
|
default:
|
|
|
|
/* Unknown IO type, complete to the user */
|
|
|
|
assert(0);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!io->status) {
|
|
|
|
/* IO rescheduled, return from the function */
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-06-21 11:03:31 +00:00
|
|
|
if (io->map) {
|
|
|
|
ftl_mempool_put(ioch->map_pool, io->map);
|
|
|
|
}
|
|
|
|
|
|
|
|
result = spdk_ring_enqueue(ioch->cq, (void **)&io, 1, NULL);
|
|
|
|
assert(result != 0);
|
2022-06-21 11:04:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
ftl_io_init(struct spdk_io_channel *_ioch, struct ftl_io *io, uint64_t lba, size_t num_blocks,
|
|
|
|
struct iovec *iov, size_t iov_cnt, spdk_ftl_fn cb_fn, void *cb_ctx, int type)
|
|
|
|
{
|
2022-06-21 11:03:31 +00:00
|
|
|
struct ftl_io_channel *ioch = ftl_io_channel_get_ctx(_ioch);
|
|
|
|
struct spdk_ftl_dev *dev = ioch->dev;
|
2022-06-21 11:04:00 +00:00
|
|
|
|
|
|
|
memset(io, 0, sizeof(struct ftl_io));
|
|
|
|
io->ioch = _ioch;
|
|
|
|
|
|
|
|
io->flags |= FTL_IO_INITIALIZED;
|
|
|
|
io->type = type;
|
|
|
|
io->dev = dev;
|
|
|
|
io->lba = FTL_LBA_INVALID;
|
|
|
|
io->addr = FTL_ADDR_INVALID;
|
|
|
|
io->cb_ctx = cb_ctx;
|
|
|
|
io->lba = lba;
|
|
|
|
io->user_fn = cb_fn;
|
|
|
|
io->iov = iov;
|
|
|
|
io->iov_cnt = iov_cnt;
|
|
|
|
io->num_blocks = num_blocks;
|
2022-05-25 13:12:52 +00:00
|
|
|
io->trace = ftl_trace_alloc_id(dev);
|
2022-06-21 11:04:00 +00:00
|
|
|
|
2022-05-25 13:12:52 +00:00
|
|
|
ftl_trace_lba_io_init(io->dev, io);
|
2022-06-21 11:04:00 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2022-07-07 17:53:08 +00:00
|
|
|
static void
|
|
|
|
ftl_io_complete_verify(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
struct spdk_ftl_dev *dev = io->dev;
|
|
|
|
uint64_t i;
|
|
|
|
uint64_t lba = io->lba;
|
|
|
|
|
|
|
|
assert(io->num_blocks <= dev->xfer_size);
|
|
|
|
|
|
|
|
if (FTL_IO_WRITE == io->type) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (spdk_unlikely(io->status)) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < io->num_blocks; i++, lba++) {
|
|
|
|
ftl_addr current_addr = ftl_l2p_get(dev, lba);
|
|
|
|
|
|
|
|
/* If user read request gets stuck for whatever reason, then it's possible the LBA
|
|
|
|
* has been relocated by GC or compaction and it may no longer be safe to return data
|
|
|
|
* from that address */
|
|
|
|
if (spdk_unlikely(current_addr != io->map[i])) {
|
|
|
|
io->status = -EAGAIN;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-06-21 11:04:00 +00:00
|
|
|
void
|
|
|
|
ftl_io_complete(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
io->flags &= ~FTL_IO_INITIALIZED;
|
|
|
|
io->done = true;
|
|
|
|
|
2022-07-07 17:53:08 +00:00
|
|
|
if (io->flags & FTL_IO_PINNED) {
|
|
|
|
ftl_io_complete_verify(io);
|
|
|
|
ftl_l2p_unpin(io->dev, io->lba, io->num_blocks);
|
|
|
|
}
|
|
|
|
|
2022-06-21 11:04:00 +00:00
|
|
|
ftl_io_cb(io, io->cb_ctx, io->status);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
ftl_io_fail(struct ftl_io *io, int status)
|
|
|
|
{
|
|
|
|
io->status = status;
|
|
|
|
ftl_io_advance(io, io->num_blocks - io->pos);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
ftl_io_clear(struct ftl_io *io)
|
|
|
|
{
|
|
|
|
io->req_cnt = io->pos = io->iov_pos = io->iov_off = 0;
|
|
|
|
io->done = false;
|
|
|
|
io->status = 0;
|
|
|
|
io->flags = 0;
|
2022-06-10 11:40:26 +00:00
|
|
|
io->band = NULL;
|
2022-06-21 11:04:00 +00:00
|
|
|
}
|