ioat: add spdk_ prefixes
Change-Id: Ic42f204a9a4e65661931af2220e8f8ac77242a69 Signed-off-by: Daniel Verkamp <daniel.verkamp@intel.com>
This commit is contained in:
parent
c7150a5611
commit
5cab054f4e
@ -41,8 +41,10 @@
|
||||
|
||||
\section key_functions Key Functions
|
||||
|
||||
- ioat_attach() \copybrief ioat_attach()
|
||||
- ioat_register_thread() \copybrief ioat_register_thread()
|
||||
- ioat_submit_copy() \copybrief ioat_submit_copy()
|
||||
- spdk_ioat_probe() \copybrief spdk_ioat_probe()
|
||||
- spdk_ioat_register_thread() \copybrief spdk_ioat_register_thread()
|
||||
- spdk_ioat_get_dma_capabilities() \copybrief spdk_ioat_get_dma_capabilities()
|
||||
- spdk_ioat_submit_copy() \copybrief spdk_ioat_submit_copy()
|
||||
- spdk_ioat_submit_fill() \copybrief spdk_ioat_submit_fill()
|
||||
|
||||
*/
|
||||
|
@ -56,7 +56,7 @@ struct user_config {
|
||||
};
|
||||
|
||||
struct ioat_device {
|
||||
struct ioat_channel *ioat;
|
||||
struct spdk_ioat_chan *ioat;
|
||||
TAILQ_ENTRY(ioat_device) tailq;
|
||||
};
|
||||
|
||||
@ -113,7 +113,7 @@ ioat_exit(void)
|
||||
dev = TAILQ_FIRST(&g_devices);
|
||||
TAILQ_REMOVE(&g_devices, dev, tailq);
|
||||
if (dev->ioat) {
|
||||
ioat_detach(dev->ioat);
|
||||
spdk_ioat_detach(dev->ioat);
|
||||
}
|
||||
rte_free(dev);
|
||||
}
|
||||
@ -161,7 +161,7 @@ probe_cb(void *cb_ctx, struct spdk_pci_device *pci_dev)
|
||||
}
|
||||
|
||||
static void
|
||||
attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct ioat_channel *ioat)
|
||||
attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
struct ioat_device *dev;
|
||||
|
||||
@ -180,7 +180,7 @@ ioat_init(void)
|
||||
{
|
||||
TAILQ_INIT(&g_devices);
|
||||
|
||||
if (ioat_probe(NULL, probe_cb, attach_cb) != 0) {
|
||||
if (spdk_ioat_probe(NULL, probe_cb, attach_cb) != 0) {
|
||||
fprintf(stderr, "ioat_probe() failed\n");
|
||||
return 1;
|
||||
}
|
||||
@ -244,7 +244,7 @@ static void
|
||||
drain_io(struct thread_entry *thread_entry)
|
||||
{
|
||||
while (thread_entry->current_queue_depth > 0) {
|
||||
ioat_process_events();
|
||||
spdk_ioat_process_events();
|
||||
}
|
||||
}
|
||||
|
||||
@ -256,7 +256,7 @@ submit_single_xfer(struct thread_entry *thread_entry, struct ioat_task *ioat_tas
|
||||
ioat_task->src = src;
|
||||
ioat_task->dst = dst;
|
||||
|
||||
ioat_submit_copy(ioat_task, ioat_done, dst, src, g_user_config.xfer_size_bytes);
|
||||
spdk_ioat_submit_copy(ioat_task, ioat_done, dst, src, g_user_config.xfer_size_bytes);
|
||||
|
||||
thread_entry->current_queue_depth++;
|
||||
}
|
||||
@ -297,7 +297,7 @@ work_fn(void *arg)
|
||||
return 1;
|
||||
}
|
||||
|
||||
if (ioat_register_thread() != 0) {
|
||||
if (spdk_ioat_register_thread() != 0) {
|
||||
fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n",
|
||||
rte_lcore_id());
|
||||
return 0;
|
||||
@ -308,13 +308,13 @@ work_fn(void *arg)
|
||||
// begin to submit transfers
|
||||
submit_xfers(t, g_user_config.queue_depth);
|
||||
while (rte_get_timer_cycles() < tsc_end) {
|
||||
ioat_process_events();
|
||||
spdk_ioat_process_events();
|
||||
}
|
||||
|
||||
// begin to drain io
|
||||
t->is_draining = true;
|
||||
drain_io(t);
|
||||
ioat_unregister_thread();
|
||||
spdk_ioat_unregister_thread();
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -61,7 +61,7 @@ struct user_config {
|
||||
};
|
||||
|
||||
struct ioat_device {
|
||||
struct ioat_channel *ioat;
|
||||
struct spdk_ioat_chan *ioat;
|
||||
TAILQ_ENTRY(ioat_device) tailq;
|
||||
};
|
||||
|
||||
@ -123,7 +123,7 @@ ioat_exit(void)
|
||||
dev = TAILQ_FIRST(&g_devices);
|
||||
TAILQ_REMOVE(&g_devices, dev, tailq);
|
||||
if (dev->ioat) {
|
||||
ioat_detach(dev->ioat);
|
||||
spdk_ioat_detach(dev->ioat);
|
||||
}
|
||||
free(dev);
|
||||
}
|
||||
@ -217,7 +217,7 @@ probe_cb(void *cb_ctx, struct spdk_pci_device *pci_dev)
|
||||
}
|
||||
|
||||
static void
|
||||
attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct ioat_channel *ioat)
|
||||
attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
struct ioat_device *dev;
|
||||
|
||||
@ -237,7 +237,7 @@ ioat_init(void)
|
||||
{
|
||||
TAILQ_INIT(&g_devices);
|
||||
|
||||
if (ioat_probe(NULL, probe_cb, attach_cb) != 0) {
|
||||
if (spdk_ioat_probe(NULL, probe_cb, attach_cb) != 0) {
|
||||
fprintf(stderr, "ioat_probe() failed\n");
|
||||
return 1;
|
||||
}
|
||||
@ -292,7 +292,7 @@ static void
|
||||
drain_xfers(struct thread_entry *thread_entry)
|
||||
{
|
||||
while (thread_entry->current_queue_depth > 0) {
|
||||
ioat_process_events();
|
||||
spdk_ioat_process_events();
|
||||
}
|
||||
}
|
||||
|
||||
@ -300,9 +300,10 @@ static void
|
||||
submit_single_xfer(struct ioat_task *ioat_task)
|
||||
{
|
||||
if (ioat_task->type == IOAT_FILL_TYPE)
|
||||
ioat_submit_fill(ioat_task, ioat_done, ioat_task->dst, ioat_task->fill_pattern, ioat_task->len);
|
||||
spdk_ioat_submit_fill(ioat_task, ioat_done, ioat_task->dst, ioat_task->fill_pattern,
|
||||
ioat_task->len);
|
||||
else
|
||||
ioat_submit_copy(ioat_task, ioat_done, ioat_task->dst, ioat_task->src, ioat_task->len);
|
||||
spdk_ioat_submit_copy(ioat_task, ioat_done, ioat_task->dst, ioat_task->src, ioat_task->len);
|
||||
ioat_task->thread_entry->current_queue_depth++;
|
||||
}
|
||||
|
||||
@ -315,7 +316,7 @@ submit_xfers(struct thread_entry *thread_entry, uint64_t queue_depth)
|
||||
rte_mempool_get(thread_entry->data_pool, &(ioat_task->buffer));
|
||||
|
||||
ioat_task->type = IOAT_COPY_TYPE;
|
||||
if (ioat_get_dma_capabilities() & IOAT_ENGINE_FILL_SUPPORTED) {
|
||||
if (spdk_ioat_get_dma_capabilities() & SPDK_IOAT_ENGINE_FILL_SUPPORTED) {
|
||||
if (queue_depth % 2)
|
||||
ioat_task->type = IOAT_FILL_TYPE;
|
||||
}
|
||||
@ -346,7 +347,7 @@ work_fn(void *arg)
|
||||
return 1;
|
||||
}
|
||||
|
||||
if (ioat_register_thread() != 0) {
|
||||
if (spdk_ioat_register_thread() != 0) {
|
||||
fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n",
|
||||
rte_lcore_id());
|
||||
return 0;
|
||||
@ -356,13 +357,13 @@ work_fn(void *arg)
|
||||
|
||||
submit_xfers(t, g_user_config.queue_depth);
|
||||
while (rte_get_timer_cycles() < tsc_end) {
|
||||
ioat_process_events();
|
||||
spdk_ioat_process_events();
|
||||
}
|
||||
|
||||
t->is_draining = true;
|
||||
drain_xfers(t);
|
||||
|
||||
ioat_unregister_thread();
|
||||
spdk_ioat_unregister_thread();
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -35,8 +35,8 @@
|
||||
* This file defines the public interface to the I/OAT DMA engine driver.
|
||||
*/
|
||||
|
||||
#ifndef __IOAT_H__
|
||||
#define __IOAT_H__
|
||||
#ifndef SPDK_IOAT_H
|
||||
#define SPDK_IOAT_H
|
||||
|
||||
#include <inttypes.h>
|
||||
#include <stdbool.h>
|
||||
@ -45,27 +45,27 @@
|
||||
#include "spdk/pci.h"
|
||||
|
||||
/**
|
||||
* Opaque handle for a single I/OAT channel returned by \ref ioat_probe().
|
||||
* Opaque handle for a single I/OAT channel returned by \ref spdk_ioat_probe().
|
||||
*/
|
||||
struct ioat_channel;
|
||||
struct spdk_ioat_chan;
|
||||
|
||||
/**
|
||||
* Signature for callback function invoked when a request is completed.
|
||||
*/
|
||||
typedef void (*ioat_callback_t)(void *arg);
|
||||
typedef void (*spdk_ioat_req_cb)(void *arg);
|
||||
|
||||
/**
|
||||
* Callback for ioat_probe() enumeration.
|
||||
* Callback for spdk_ioat_probe() enumeration.
|
||||
*
|
||||
* \return true to attach to this device.
|
||||
*/
|
||||
typedef bool (*ioat_probe_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev);
|
||||
typedef bool (*spdk_ioat_probe_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev);
|
||||
|
||||
/**
|
||||
* Callback for ioat_probe() to report a device that has been attached to the userspace I/OAT driver.
|
||||
* Callback for spdk_ioat_probe() to report a device that has been attached to the userspace I/OAT driver.
|
||||
*/
|
||||
typedef void (*ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev,
|
||||
struct ioat_channel *ioat);
|
||||
typedef void (*spdk_ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev,
|
||||
struct spdk_ioat_chan *ioat);
|
||||
|
||||
/**
|
||||
* \brief Enumerate the I/OAT devices attached to the system and attach the userspace I/OAT driver
|
||||
@ -79,74 +79,74 @@ typedef void (*ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev,
|
||||
* will be reported.
|
||||
*
|
||||
* To stop using the the controller and release its associated resources,
|
||||
* call \ref ioat_detach with the ioat_channel instance returned by this function.
|
||||
* call \ref spdk_ioat_detach with the ioat_channel instance returned by this function.
|
||||
*/
|
||||
int ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb);
|
||||
int spdk_ioat_probe(void *cb_ctx, spdk_ioat_probe_cb probe_cb, spdk_ioat_attach_cb attach_cb);
|
||||
|
||||
/**
|
||||
* Detaches specified device returned by \ref ioat_probe() from the I/OAT driver.
|
||||
* Detaches specified device returned by \ref spdk_ioat_probe() from the I/OAT driver.
|
||||
*/
|
||||
int ioat_detach(struct ioat_channel *ioat);
|
||||
int spdk_ioat_detach(struct spdk_ioat_chan *ioat);
|
||||
|
||||
/**
|
||||
* Request a DMA engine channel for the calling thread.
|
||||
*
|
||||
* Must be called before submitting any requests from a thread.
|
||||
*
|
||||
* The \ref ioat_unregister_thread() function can be called to release the channel.
|
||||
* The \ref spdk_ioat_unregister_thread() function can be called to release the channel.
|
||||
*/
|
||||
int ioat_register_thread(void);
|
||||
int spdk_ioat_register_thread(void);
|
||||
|
||||
/**
|
||||
* Unregister the current thread's I/OAT channel.
|
||||
*
|
||||
* This function can be called after \ref ioat_register_thread() to release the thread's
|
||||
* This function can be called after \ref spdk_ioat_register_thread() to release the thread's
|
||||
* DMA engine channel for use by other threads.
|
||||
*/
|
||||
void ioat_unregister_thread(void);
|
||||
void spdk_ioat_unregister_thread(void);
|
||||
|
||||
/**
|
||||
* Submit a DMA engine memory copy request.
|
||||
*
|
||||
* Before submitting any requests on a thread, the thread must be registered
|
||||
* using the \ref ioat_register_thread() function.
|
||||
* using the \ref spdk_ioat_register_thread() function.
|
||||
*/
|
||||
int64_t ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn,
|
||||
void *dst, const void *src, uint64_t nbytes);
|
||||
int64_t spdk_ioat_submit_copy(void *cb_arg, spdk_ioat_req_cb cb_fn,
|
||||
void *dst, const void *src, uint64_t nbytes);
|
||||
|
||||
/**
|
||||
* Submit a DMA engine memory fill request.
|
||||
*
|
||||
* Before submitting any requests on a thread, the thread must be registered
|
||||
* using the \ref ioat_register_thread() function.
|
||||
* using the \ref spdk_ioat_register_thread() function.
|
||||
*/
|
||||
int64_t ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn,
|
||||
void *dst, uint64_t fill_pattern, uint64_t nbytes);
|
||||
int64_t spdk_ioat_submit_fill(void *cb_arg, spdk_ioat_req_cb cb_fn,
|
||||
void *dst, uint64_t fill_pattern, uint64_t nbytes);
|
||||
|
||||
/**
|
||||
* Check for completed requests on the current thread.
|
||||
*
|
||||
* Before submitting any requests on a thread, the thread must be registered
|
||||
* using the \ref ioat_register_thread() function.
|
||||
* using the \ref spdk_ioat_register_thread() function.
|
||||
*
|
||||
* \returns 0 on success or negative if something went wrong.
|
||||
*/
|
||||
int ioat_process_events(void);
|
||||
int spdk_ioat_process_events(void);
|
||||
|
||||
/**
|
||||
* DMA engine capability flags
|
||||
*/
|
||||
enum ioat_dma_capability_flags {
|
||||
IOAT_ENGINE_COPY_SUPPORTED = 0x1, /**< The memory copy is supported */
|
||||
IOAT_ENGINE_FILL_SUPPORTED = 0x2, /**< The memory fill is supported */
|
||||
enum spdk_ioat_dma_capability_flags {
|
||||
SPDK_IOAT_ENGINE_COPY_SUPPORTED = 0x1, /**< The memory copy is supported */
|
||||
SPDK_IOAT_ENGINE_FILL_SUPPORTED = 0x2, /**< The memory fill is supported */
|
||||
};
|
||||
|
||||
/**
|
||||
* Get the DMA engine capabilities.
|
||||
*
|
||||
* Before submitting any requests on a thread, the thread must be registered
|
||||
* using the \ref ioat_register_thread() function.
|
||||
* using the \ref spdk_ioat_register_thread() function.
|
||||
*/
|
||||
uint32_t ioat_get_dma_capabilities(void);
|
||||
uint32_t spdk_ioat_get_dma_capabilities(void);
|
||||
|
||||
#endif
|
||||
|
@ -38,14 +38,14 @@
|
||||
*
|
||||
* Must hold g_ioat_driver.lock while manipulating this list.
|
||||
*/
|
||||
static SLIST_HEAD(, ioat_channel) ioat_free_channels;
|
||||
static SLIST_HEAD(, spdk_ioat_chan) ioat_free_channels;
|
||||
|
||||
/** IOAT channel assigned to this thread (or NULL if not assigned yet). */
|
||||
static __thread struct ioat_channel *ioat_thread_channel;
|
||||
static __thread struct spdk_ioat_chan *ioat_thread_channel;
|
||||
|
||||
struct ioat_driver {
|
||||
ioat_mutex_t lock;
|
||||
TAILQ_HEAD(, ioat_channel) attached_chans;
|
||||
TAILQ_HEAD(, spdk_ioat_chan) attached_chans;
|
||||
};
|
||||
|
||||
static struct ioat_driver g_ioat_driver = {
|
||||
@ -54,37 +54,37 @@ static struct ioat_driver g_ioat_driver = {
|
||||
};
|
||||
|
||||
static uint64_t
|
||||
ioat_get_chansts(struct ioat_channel *ioat)
|
||||
ioat_get_chansts(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
return spdk_mmio_read_8(&ioat->regs->chansts);
|
||||
}
|
||||
|
||||
static void
|
||||
ioat_write_chancmp(struct ioat_channel *ioat, uint64_t addr)
|
||||
ioat_write_chancmp(struct spdk_ioat_chan *ioat, uint64_t addr)
|
||||
{
|
||||
spdk_mmio_write_8(&ioat->regs->chancmp, addr);
|
||||
}
|
||||
|
||||
static void
|
||||
ioat_write_chainaddr(struct ioat_channel *ioat, uint64_t addr)
|
||||
ioat_write_chainaddr(struct spdk_ioat_chan *ioat, uint64_t addr)
|
||||
{
|
||||
spdk_mmio_write_8(&ioat->regs->chainaddr, addr);
|
||||
}
|
||||
|
||||
static inline void
|
||||
ioat_suspend(struct ioat_channel *ioat)
|
||||
ioat_suspend(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
ioat->regs->chancmd = SPDK_IOAT_CHANCMD_SUSPEND;
|
||||
}
|
||||
|
||||
static inline void
|
||||
ioat_reset(struct ioat_channel *ioat)
|
||||
ioat_reset(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
ioat->regs->chancmd = SPDK_IOAT_CHANCMD_RESET;
|
||||
}
|
||||
|
||||
static inline uint32_t
|
||||
ioat_reset_pending(struct ioat_channel *ioat)
|
||||
ioat_reset_pending(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
uint8_t cmd;
|
||||
|
||||
@ -93,7 +93,7 @@ ioat_reset_pending(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
static int
|
||||
ioat_map_pci_bar(struct ioat_channel *ioat)
|
||||
ioat_map_pci_bar(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
int regs_bar, rc;
|
||||
void *addr;
|
||||
@ -112,7 +112,7 @@ ioat_map_pci_bar(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
static int
|
||||
ioat_unmap_pci_bar(struct ioat_channel *ioat)
|
||||
ioat_unmap_pci_bar(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
int rc = 0;
|
||||
void *addr = (void *)ioat->regs;
|
||||
@ -125,25 +125,25 @@ ioat_unmap_pci_bar(struct ioat_channel *ioat)
|
||||
|
||||
|
||||
static inline uint32_t
|
||||
ioat_get_active(struct ioat_channel *ioat)
|
||||
ioat_get_active(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
return (ioat->head - ioat->tail) & ((1 << ioat->ring_size_order) - 1);
|
||||
}
|
||||
|
||||
static inline uint32_t
|
||||
ioat_get_ring_space(struct ioat_channel *ioat)
|
||||
ioat_get_ring_space(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
return (1 << ioat->ring_size_order) - ioat_get_active(ioat) - 1;
|
||||
}
|
||||
|
||||
static uint32_t
|
||||
ioat_get_ring_index(struct ioat_channel *ioat, uint32_t index)
|
||||
ioat_get_ring_index(struct spdk_ioat_chan *ioat, uint32_t index)
|
||||
{
|
||||
return index & ((1 << ioat->ring_size_order) - 1);
|
||||
}
|
||||
|
||||
static void
|
||||
ioat_get_ring_entry(struct ioat_channel *ioat, uint32_t index,
|
||||
ioat_get_ring_entry(struct spdk_ioat_chan *ioat, uint32_t index,
|
||||
struct ioat_descriptor **desc,
|
||||
union spdk_ioat_hw_desc **hw_desc)
|
||||
{
|
||||
@ -154,26 +154,26 @@ ioat_get_ring_entry(struct ioat_channel *ioat, uint32_t index,
|
||||
}
|
||||
|
||||
static uint64_t
|
||||
ioat_get_desc_phys_addr(struct ioat_channel *ioat, uint32_t index)
|
||||
ioat_get_desc_phys_addr(struct spdk_ioat_chan *ioat, uint32_t index)
|
||||
{
|
||||
return ioat->hw_ring_phys_addr +
|
||||
ioat_get_ring_index(ioat, index) * sizeof(union spdk_ioat_hw_desc);
|
||||
}
|
||||
|
||||
static void
|
||||
ioat_submit_single(struct ioat_channel *ioat)
|
||||
ioat_submit_single(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
ioat->head++;
|
||||
}
|
||||
|
||||
static void
|
||||
ioat_flush(struct ioat_channel *ioat)
|
||||
ioat_flush(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
ioat->regs->dmacount = (uint16_t)ioat->head;
|
||||
}
|
||||
|
||||
static struct ioat_descriptor *
|
||||
ioat_prep_null(struct ioat_channel *ioat)
|
||||
ioat_prep_null(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
struct ioat_descriptor *desc;
|
||||
union spdk_ioat_hw_desc *hw_desc;
|
||||
@ -202,7 +202,7 @@ ioat_prep_null(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
static struct ioat_descriptor *
|
||||
ioat_prep_copy(struct ioat_channel *ioat, uint64_t dst,
|
||||
ioat_prep_copy(struct spdk_ioat_chan *ioat, uint64_t dst,
|
||||
uint64_t src, uint32_t len)
|
||||
{
|
||||
struct ioat_descriptor *desc;
|
||||
@ -233,7 +233,7 @@ ioat_prep_copy(struct ioat_channel *ioat, uint64_t dst,
|
||||
}
|
||||
|
||||
static struct ioat_descriptor *
|
||||
ioat_prep_fill(struct ioat_channel *ioat, uint64_t dst,
|
||||
ioat_prep_fill(struct spdk_ioat_chan *ioat, uint64_t dst,
|
||||
uint64_t fill_pattern, uint32_t len)
|
||||
{
|
||||
struct ioat_descriptor *desc;
|
||||
@ -263,7 +263,7 @@ ioat_prep_fill(struct ioat_channel *ioat, uint64_t dst,
|
||||
return desc;
|
||||
}
|
||||
|
||||
static int ioat_reset_hw(struct ioat_channel *ioat)
|
||||
static int ioat_reset_hw(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
int timeout;
|
||||
uint64_t status;
|
||||
@ -308,7 +308,7 @@ static int ioat_reset_hw(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
static int
|
||||
ioat_process_channel_events(struct ioat_channel *ioat)
|
||||
ioat_process_channel_events(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
struct ioat_descriptor *desc;
|
||||
uint64_t status, completed_descriptor, hw_desc_phys_addr;
|
||||
@ -347,7 +347,7 @@ ioat_process_channel_events(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
static int
|
||||
ioat_channel_destruct(struct ioat_channel *ioat)
|
||||
ioat_channel_destruct(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
ioat_unmap_pci_bar(ioat);
|
||||
|
||||
@ -368,7 +368,7 @@ ioat_channel_destruct(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
static int
|
||||
ioat_channel_start(struct ioat_channel *ioat)
|
||||
ioat_channel_start(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
uint8_t xfercap, version;
|
||||
uint64_t status;
|
||||
@ -388,9 +388,9 @@ ioat_channel_start(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
/* Always support DMA copy */
|
||||
ioat->dma_capabilities = IOAT_ENGINE_COPY_SUPPORTED;
|
||||
ioat->dma_capabilities = SPDK_IOAT_ENGINE_COPY_SUPPORTED;
|
||||
if (ioat->regs->dmacapability & SPDK_IOAT_DMACAP_BFILL)
|
||||
ioat->dma_capabilities |= IOAT_ENGINE_FILL_SUPPORTED;
|
||||
ioat->dma_capabilities |= SPDK_IOAT_ENGINE_FILL_SUPPORTED;
|
||||
xfercap = ioat->regs->xfercap;
|
||||
|
||||
/* Only bits [4:0] are valid. */
|
||||
@ -464,13 +464,13 @@ ioat_channel_start(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
/* Caller must hold g_ioat_driver.lock */
|
||||
static struct ioat_channel *
|
||||
static struct spdk_ioat_chan *
|
||||
ioat_attach(void *device)
|
||||
{
|
||||
struct ioat_channel *ioat;
|
||||
struct spdk_ioat_chan *ioat;
|
||||
uint32_t cmd_reg;
|
||||
|
||||
ioat = calloc(1, sizeof(struct ioat_channel));
|
||||
ioat = calloc(1, sizeof(struct spdk_ioat_chan));
|
||||
if (ioat == NULL) {
|
||||
return NULL;
|
||||
}
|
||||
@ -494,8 +494,8 @@ ioat_attach(void *device)
|
||||
}
|
||||
|
||||
struct ioat_enum_ctx {
|
||||
ioat_probe_cb probe_cb;
|
||||
ioat_attach_cb attach_cb;
|
||||
spdk_ioat_probe_cb probe_cb;
|
||||
spdk_ioat_attach_cb attach_cb;
|
||||
void *cb_ctx;
|
||||
};
|
||||
|
||||
@ -504,7 +504,7 @@ static int
|
||||
ioat_enum_cb(void *ctx, struct spdk_pci_device *pci_dev)
|
||||
{
|
||||
struct ioat_enum_ctx *enum_ctx = ctx;
|
||||
struct ioat_channel *ioat;
|
||||
struct spdk_ioat_chan *ioat;
|
||||
|
||||
/* Verify that this device is not already attached */
|
||||
TAILQ_FOREACH(ioat, &g_ioat_driver.attached_chans, tailq) {
|
||||
@ -538,7 +538,7 @@ ioat_enum_cb(void *ctx, struct spdk_pci_device *pci_dev)
|
||||
}
|
||||
|
||||
int
|
||||
ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb)
|
||||
spdk_ioat_probe(void *cb_ctx, spdk_ioat_probe_cb probe_cb, spdk_ioat_attach_cb attach_cb)
|
||||
{
|
||||
int rc;
|
||||
struct ioat_enum_ctx enum_ctx;
|
||||
@ -557,7 +557,7 @@ ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb)
|
||||
}
|
||||
|
||||
int
|
||||
ioat_detach(struct ioat_channel *ioat)
|
||||
spdk_ioat_detach(struct spdk_ioat_chan *ioat)
|
||||
{
|
||||
struct ioat_driver *driver = &g_ioat_driver;
|
||||
|
||||
@ -565,7 +565,7 @@ ioat_detach(struct ioat_channel *ioat)
|
||||
* when calling ioat_detach().
|
||||
*/
|
||||
ioat_mutex_lock(&driver->lock);
|
||||
SLIST_REMOVE(&ioat_free_channels, ioat, ioat_channel, next);
|
||||
SLIST_REMOVE(&ioat_free_channels, ioat, spdk_ioat_chan, next);
|
||||
TAILQ_REMOVE(&driver->attached_chans, ioat, tailq);
|
||||
ioat_mutex_unlock(&driver->lock);
|
||||
|
||||
@ -576,7 +576,7 @@ ioat_detach(struct ioat_channel *ioat)
|
||||
}
|
||||
|
||||
int
|
||||
ioat_register_thread(void)
|
||||
spdk_ioat_register_thread(void)
|
||||
{
|
||||
struct ioat_driver *driver = &g_ioat_driver;
|
||||
|
||||
@ -598,7 +598,7 @@ ioat_register_thread(void)
|
||||
}
|
||||
|
||||
void
|
||||
ioat_unregister_thread(void)
|
||||
spdk_ioat_unregister_thread(void)
|
||||
{
|
||||
struct ioat_driver *driver = &g_ioat_driver;
|
||||
|
||||
@ -620,10 +620,10 @@ ioat_unregister_thread(void)
|
||||
#define _2MB_OFFSET(ptr) ((ptr) & (0x200000 - 1))
|
||||
|
||||
int64_t
|
||||
ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn,
|
||||
void *dst, const void *src, uint64_t nbytes)
|
||||
spdk_ioat_submit_copy(void *cb_arg, spdk_ioat_req_cb cb_fn,
|
||||
void *dst, const void *src, uint64_t nbytes)
|
||||
{
|
||||
struct ioat_channel *ioat;
|
||||
struct spdk_ioat_chan *ioat;
|
||||
struct ioat_descriptor *last_desc;
|
||||
uint64_t remaining, op_size;
|
||||
uint64_t vdst, vsrc;
|
||||
@ -698,10 +698,10 @@ ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn,
|
||||
}
|
||||
|
||||
int64_t
|
||||
ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn,
|
||||
void *dst, uint64_t fill_pattern, uint64_t nbytes)
|
||||
spdk_ioat_submit_fill(void *cb_arg, spdk_ioat_req_cb cb_fn,
|
||||
void *dst, uint64_t fill_pattern, uint64_t nbytes)
|
||||
{
|
||||
struct ioat_channel *ioat;
|
||||
struct spdk_ioat_chan *ioat;
|
||||
struct ioat_descriptor *last_desc = NULL;
|
||||
uint64_t remaining, op_size;
|
||||
uint64_t vdst;
|
||||
@ -712,7 +712,7 @@ ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn,
|
||||
return -1;
|
||||
}
|
||||
|
||||
if (!(ioat->dma_capabilities & IOAT_ENGINE_FILL_SUPPORTED)) {
|
||||
if (!(ioat->dma_capabilities & SPDK_IOAT_ENGINE_FILL_SUPPORTED)) {
|
||||
ioat_printf(ioat, "Channel does not support memory fill\n");
|
||||
return -1;
|
||||
}
|
||||
@ -755,9 +755,10 @@ ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn,
|
||||
return nbytes;
|
||||
}
|
||||
|
||||
uint32_t ioat_get_dma_capabilities(void)
|
||||
uint32_t
|
||||
spdk_ioat_get_dma_capabilities(void)
|
||||
{
|
||||
struct ioat_channel *ioat;
|
||||
struct spdk_ioat_chan *ioat;
|
||||
|
||||
ioat = ioat_thread_channel;
|
||||
if (!ioat) {
|
||||
@ -766,7 +767,8 @@ uint32_t ioat_get_dma_capabilities(void)
|
||||
return ioat->dma_capabilities;
|
||||
}
|
||||
|
||||
int ioat_process_events(void)
|
||||
int
|
||||
spdk_ioat_process_events(void)
|
||||
{
|
||||
if (!ioat_thread_channel) {
|
||||
return -1;
|
||||
|
@ -50,13 +50,13 @@
|
||||
#define IOAT_DEFAULT_ORDER 15
|
||||
|
||||
struct ioat_descriptor {
|
||||
ioat_callback_t callback_fn;
|
||||
spdk_ioat_req_cb callback_fn;
|
||||
void *callback_arg;
|
||||
};
|
||||
|
||||
/* One of these per allocated PCI device. */
|
||||
struct ioat_channel {
|
||||
SLIST_ENTRY(ioat_channel) next;
|
||||
struct spdk_ioat_chan {
|
||||
SLIST_ENTRY(spdk_ioat_chan) next;
|
||||
|
||||
/* Opaque handle to upper layer */
|
||||
void *device;
|
||||
@ -77,7 +77,7 @@ struct ioat_channel {
|
||||
uint32_t dma_capabilities;
|
||||
|
||||
/* tailq entry for attached_chans */
|
||||
TAILQ_ENTRY(ioat_channel) tailq;
|
||||
TAILQ_ENTRY(spdk_ioat_chan) tailq;
|
||||
};
|
||||
|
||||
static inline uint32_t
|
||||
|
Loading…
Reference in New Issue
Block a user