ioat: add spdk_ prefixes

Change-Id: Ic42f204a9a4e65661931af2220e8f8ac77242a69
Signed-off-by: Daniel Verkamp <daniel.verkamp@intel.com>
This commit is contained in:
Daniel Verkamp 2016-02-11 12:31:35 -07:00
parent c7150a5611
commit 5cab054f4e
6 changed files with 111 additions and 106 deletions

View File

@ -41,8 +41,10 @@
\section key_functions Key Functions \section key_functions Key Functions
- ioat_attach() \copybrief ioat_attach() - spdk_ioat_probe() \copybrief spdk_ioat_probe()
- ioat_register_thread() \copybrief ioat_register_thread() - spdk_ioat_register_thread() \copybrief spdk_ioat_register_thread()
- ioat_submit_copy() \copybrief ioat_submit_copy() - spdk_ioat_get_dma_capabilities() \copybrief spdk_ioat_get_dma_capabilities()
- spdk_ioat_submit_copy() \copybrief spdk_ioat_submit_copy()
- spdk_ioat_submit_fill() \copybrief spdk_ioat_submit_fill()
*/ */

View File

@ -56,7 +56,7 @@ struct user_config {
}; };
struct ioat_device { struct ioat_device {
struct ioat_channel *ioat; struct spdk_ioat_chan *ioat;
TAILQ_ENTRY(ioat_device) tailq; TAILQ_ENTRY(ioat_device) tailq;
}; };
@ -113,7 +113,7 @@ ioat_exit(void)
dev = TAILQ_FIRST(&g_devices); dev = TAILQ_FIRST(&g_devices);
TAILQ_REMOVE(&g_devices, dev, tailq); TAILQ_REMOVE(&g_devices, dev, tailq);
if (dev->ioat) { if (dev->ioat) {
ioat_detach(dev->ioat); spdk_ioat_detach(dev->ioat);
} }
rte_free(dev); rte_free(dev);
} }
@ -161,7 +161,7 @@ probe_cb(void *cb_ctx, struct spdk_pci_device *pci_dev)
} }
static void static void
attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct ioat_channel *ioat) attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct spdk_ioat_chan *ioat)
{ {
struct ioat_device *dev; struct ioat_device *dev;
@ -180,7 +180,7 @@ ioat_init(void)
{ {
TAILQ_INIT(&g_devices); TAILQ_INIT(&g_devices);
if (ioat_probe(NULL, probe_cb, attach_cb) != 0) { if (spdk_ioat_probe(NULL, probe_cb, attach_cb) != 0) {
fprintf(stderr, "ioat_probe() failed\n"); fprintf(stderr, "ioat_probe() failed\n");
return 1; return 1;
} }
@ -244,7 +244,7 @@ static void
drain_io(struct thread_entry *thread_entry) drain_io(struct thread_entry *thread_entry)
{ {
while (thread_entry->current_queue_depth > 0) { while (thread_entry->current_queue_depth > 0) {
ioat_process_events(); spdk_ioat_process_events();
} }
} }
@ -256,7 +256,7 @@ submit_single_xfer(struct thread_entry *thread_entry, struct ioat_task *ioat_tas
ioat_task->src = src; ioat_task->src = src;
ioat_task->dst = dst; ioat_task->dst = dst;
ioat_submit_copy(ioat_task, ioat_done, dst, src, g_user_config.xfer_size_bytes); spdk_ioat_submit_copy(ioat_task, ioat_done, dst, src, g_user_config.xfer_size_bytes);
thread_entry->current_queue_depth++; thread_entry->current_queue_depth++;
} }
@ -297,7 +297,7 @@ work_fn(void *arg)
return 1; return 1;
} }
if (ioat_register_thread() != 0) { if (spdk_ioat_register_thread() != 0) {
fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n", fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n",
rte_lcore_id()); rte_lcore_id());
return 0; return 0;
@ -308,13 +308,13 @@ work_fn(void *arg)
// begin to submit transfers // begin to submit transfers
submit_xfers(t, g_user_config.queue_depth); submit_xfers(t, g_user_config.queue_depth);
while (rte_get_timer_cycles() < tsc_end) { while (rte_get_timer_cycles() < tsc_end) {
ioat_process_events(); spdk_ioat_process_events();
} }
// begin to drain io // begin to drain io
t->is_draining = true; t->is_draining = true;
drain_io(t); drain_io(t);
ioat_unregister_thread(); spdk_ioat_unregister_thread();
return 0; return 0;
} }

View File

@ -61,7 +61,7 @@ struct user_config {
}; };
struct ioat_device { struct ioat_device {
struct ioat_channel *ioat; struct spdk_ioat_chan *ioat;
TAILQ_ENTRY(ioat_device) tailq; TAILQ_ENTRY(ioat_device) tailq;
}; };
@ -123,7 +123,7 @@ ioat_exit(void)
dev = TAILQ_FIRST(&g_devices); dev = TAILQ_FIRST(&g_devices);
TAILQ_REMOVE(&g_devices, dev, tailq); TAILQ_REMOVE(&g_devices, dev, tailq);
if (dev->ioat) { if (dev->ioat) {
ioat_detach(dev->ioat); spdk_ioat_detach(dev->ioat);
} }
free(dev); free(dev);
} }
@ -217,7 +217,7 @@ probe_cb(void *cb_ctx, struct spdk_pci_device *pci_dev)
} }
static void static void
attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct ioat_channel *ioat) attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct spdk_ioat_chan *ioat)
{ {
struct ioat_device *dev; struct ioat_device *dev;
@ -237,7 +237,7 @@ ioat_init(void)
{ {
TAILQ_INIT(&g_devices); TAILQ_INIT(&g_devices);
if (ioat_probe(NULL, probe_cb, attach_cb) != 0) { if (spdk_ioat_probe(NULL, probe_cb, attach_cb) != 0) {
fprintf(stderr, "ioat_probe() failed\n"); fprintf(stderr, "ioat_probe() failed\n");
return 1; return 1;
} }
@ -292,7 +292,7 @@ static void
drain_xfers(struct thread_entry *thread_entry) drain_xfers(struct thread_entry *thread_entry)
{ {
while (thread_entry->current_queue_depth > 0) { while (thread_entry->current_queue_depth > 0) {
ioat_process_events(); spdk_ioat_process_events();
} }
} }
@ -300,9 +300,10 @@ static void
submit_single_xfer(struct ioat_task *ioat_task) submit_single_xfer(struct ioat_task *ioat_task)
{ {
if (ioat_task->type == IOAT_FILL_TYPE) if (ioat_task->type == IOAT_FILL_TYPE)
ioat_submit_fill(ioat_task, ioat_done, ioat_task->dst, ioat_task->fill_pattern, ioat_task->len); spdk_ioat_submit_fill(ioat_task, ioat_done, ioat_task->dst, ioat_task->fill_pattern,
ioat_task->len);
else else
ioat_submit_copy(ioat_task, ioat_done, ioat_task->dst, ioat_task->src, ioat_task->len); spdk_ioat_submit_copy(ioat_task, ioat_done, ioat_task->dst, ioat_task->src, ioat_task->len);
ioat_task->thread_entry->current_queue_depth++; ioat_task->thread_entry->current_queue_depth++;
} }
@ -315,7 +316,7 @@ submit_xfers(struct thread_entry *thread_entry, uint64_t queue_depth)
rte_mempool_get(thread_entry->data_pool, &(ioat_task->buffer)); rte_mempool_get(thread_entry->data_pool, &(ioat_task->buffer));
ioat_task->type = IOAT_COPY_TYPE; ioat_task->type = IOAT_COPY_TYPE;
if (ioat_get_dma_capabilities() & IOAT_ENGINE_FILL_SUPPORTED) { if (spdk_ioat_get_dma_capabilities() & SPDK_IOAT_ENGINE_FILL_SUPPORTED) {
if (queue_depth % 2) if (queue_depth % 2)
ioat_task->type = IOAT_FILL_TYPE; ioat_task->type = IOAT_FILL_TYPE;
} }
@ -346,7 +347,7 @@ work_fn(void *arg)
return 1; return 1;
} }
if (ioat_register_thread() != 0) { if (spdk_ioat_register_thread() != 0) {
fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n", fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n",
rte_lcore_id()); rte_lcore_id());
return 0; return 0;
@ -356,13 +357,13 @@ work_fn(void *arg)
submit_xfers(t, g_user_config.queue_depth); submit_xfers(t, g_user_config.queue_depth);
while (rte_get_timer_cycles() < tsc_end) { while (rte_get_timer_cycles() < tsc_end) {
ioat_process_events(); spdk_ioat_process_events();
} }
t->is_draining = true; t->is_draining = true;
drain_xfers(t); drain_xfers(t);
ioat_unregister_thread(); spdk_ioat_unregister_thread();
return 0; return 0;
} }

View File

@ -35,8 +35,8 @@
* This file defines the public interface to the I/OAT DMA engine driver. * This file defines the public interface to the I/OAT DMA engine driver.
*/ */
#ifndef __IOAT_H__ #ifndef SPDK_IOAT_H
#define __IOAT_H__ #define SPDK_IOAT_H
#include <inttypes.h> #include <inttypes.h>
#include <stdbool.h> #include <stdbool.h>
@ -45,27 +45,27 @@
#include "spdk/pci.h" #include "spdk/pci.h"
/** /**
* Opaque handle for a single I/OAT channel returned by \ref ioat_probe(). * Opaque handle for a single I/OAT channel returned by \ref spdk_ioat_probe().
*/ */
struct ioat_channel; struct spdk_ioat_chan;
/** /**
* Signature for callback function invoked when a request is completed. * Signature for callback function invoked when a request is completed.
*/ */
typedef void (*ioat_callback_t)(void *arg); typedef void (*spdk_ioat_req_cb)(void *arg);
/** /**
* Callback for ioat_probe() enumeration. * Callback for spdk_ioat_probe() enumeration.
* *
* \return true to attach to this device. * \return true to attach to this device.
*/ */
typedef bool (*ioat_probe_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev); typedef bool (*spdk_ioat_probe_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev);
/** /**
* Callback for ioat_probe() to report a device that has been attached to the userspace I/OAT driver. * Callback for spdk_ioat_probe() to report a device that has been attached to the userspace I/OAT driver.
*/ */
typedef void (*ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev, typedef void (*spdk_ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev,
struct ioat_channel *ioat); struct spdk_ioat_chan *ioat);
/** /**
* \brief Enumerate the I/OAT devices attached to the system and attach the userspace I/OAT driver * \brief Enumerate the I/OAT devices attached to the system and attach the userspace I/OAT driver
@ -79,74 +79,74 @@ typedef void (*ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev,
* will be reported. * will be reported.
* *
* To stop using the the controller and release its associated resources, * To stop using the the controller and release its associated resources,
* call \ref ioat_detach with the ioat_channel instance returned by this function. * call \ref spdk_ioat_detach with the ioat_channel instance returned by this function.
*/ */
int ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb); int spdk_ioat_probe(void *cb_ctx, spdk_ioat_probe_cb probe_cb, spdk_ioat_attach_cb attach_cb);
/** /**
* Detaches specified device returned by \ref ioat_probe() from the I/OAT driver. * Detaches specified device returned by \ref spdk_ioat_probe() from the I/OAT driver.
*/ */
int ioat_detach(struct ioat_channel *ioat); int spdk_ioat_detach(struct spdk_ioat_chan *ioat);
/** /**
* Request a DMA engine channel for the calling thread. * Request a DMA engine channel for the calling thread.
* *
* Must be called before submitting any requests from a thread. * Must be called before submitting any requests from a thread.
* *
* The \ref ioat_unregister_thread() function can be called to release the channel. * The \ref spdk_ioat_unregister_thread() function can be called to release the channel.
*/ */
int ioat_register_thread(void); int spdk_ioat_register_thread(void);
/** /**
* Unregister the current thread's I/OAT channel. * Unregister the current thread's I/OAT channel.
* *
* This function can be called after \ref ioat_register_thread() to release the thread's * This function can be called after \ref spdk_ioat_register_thread() to release the thread's
* DMA engine channel for use by other threads. * DMA engine channel for use by other threads.
*/ */
void ioat_unregister_thread(void); void spdk_ioat_unregister_thread(void);
/** /**
* Submit a DMA engine memory copy request. * Submit a DMA engine memory copy request.
* *
* Before submitting any requests on a thread, the thread must be registered * Before submitting any requests on a thread, the thread must be registered
* using the \ref ioat_register_thread() function. * using the \ref spdk_ioat_register_thread() function.
*/ */
int64_t ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn, int64_t spdk_ioat_submit_copy(void *cb_arg, spdk_ioat_req_cb cb_fn,
void *dst, const void *src, uint64_t nbytes); void *dst, const void *src, uint64_t nbytes);
/** /**
* Submit a DMA engine memory fill request. * Submit a DMA engine memory fill request.
* *
* Before submitting any requests on a thread, the thread must be registered * Before submitting any requests on a thread, the thread must be registered
* using the \ref ioat_register_thread() function. * using the \ref spdk_ioat_register_thread() function.
*/ */
int64_t ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn, int64_t spdk_ioat_submit_fill(void *cb_arg, spdk_ioat_req_cb cb_fn,
void *dst, uint64_t fill_pattern, uint64_t nbytes); void *dst, uint64_t fill_pattern, uint64_t nbytes);
/** /**
* Check for completed requests on the current thread. * Check for completed requests on the current thread.
* *
* Before submitting any requests on a thread, the thread must be registered * Before submitting any requests on a thread, the thread must be registered
* using the \ref ioat_register_thread() function. * using the \ref spdk_ioat_register_thread() function.
* *
* \returns 0 on success or negative if something went wrong. * \returns 0 on success or negative if something went wrong.
*/ */
int ioat_process_events(void); int spdk_ioat_process_events(void);
/** /**
* DMA engine capability flags * DMA engine capability flags
*/ */
enum ioat_dma_capability_flags { enum spdk_ioat_dma_capability_flags {
IOAT_ENGINE_COPY_SUPPORTED = 0x1, /**< The memory copy is supported */ SPDK_IOAT_ENGINE_COPY_SUPPORTED = 0x1, /**< The memory copy is supported */
IOAT_ENGINE_FILL_SUPPORTED = 0x2, /**< The memory fill is supported */ SPDK_IOAT_ENGINE_FILL_SUPPORTED = 0x2, /**< The memory fill is supported */
}; };
/** /**
* Get the DMA engine capabilities. * Get the DMA engine capabilities.
* *
* Before submitting any requests on a thread, the thread must be registered * Before submitting any requests on a thread, the thread must be registered
* using the \ref ioat_register_thread() function. * using the \ref spdk_ioat_register_thread() function.
*/ */
uint32_t ioat_get_dma_capabilities(void); uint32_t spdk_ioat_get_dma_capabilities(void);
#endif #endif

View File

@ -38,14 +38,14 @@
* *
* Must hold g_ioat_driver.lock while manipulating this list. * Must hold g_ioat_driver.lock while manipulating this list.
*/ */
static SLIST_HEAD(, ioat_channel) ioat_free_channels; static SLIST_HEAD(, spdk_ioat_chan) ioat_free_channels;
/** IOAT channel assigned to this thread (or NULL if not assigned yet). */ /** IOAT channel assigned to this thread (or NULL if not assigned yet). */
static __thread struct ioat_channel *ioat_thread_channel; static __thread struct spdk_ioat_chan *ioat_thread_channel;
struct ioat_driver { struct ioat_driver {
ioat_mutex_t lock; ioat_mutex_t lock;
TAILQ_HEAD(, ioat_channel) attached_chans; TAILQ_HEAD(, spdk_ioat_chan) attached_chans;
}; };
static struct ioat_driver g_ioat_driver = { static struct ioat_driver g_ioat_driver = {
@ -54,37 +54,37 @@ static struct ioat_driver g_ioat_driver = {
}; };
static uint64_t static uint64_t
ioat_get_chansts(struct ioat_channel *ioat) ioat_get_chansts(struct spdk_ioat_chan *ioat)
{ {
return spdk_mmio_read_8(&ioat->regs->chansts); return spdk_mmio_read_8(&ioat->regs->chansts);
} }
static void static void
ioat_write_chancmp(struct ioat_channel *ioat, uint64_t addr) ioat_write_chancmp(struct spdk_ioat_chan *ioat, uint64_t addr)
{ {
spdk_mmio_write_8(&ioat->regs->chancmp, addr); spdk_mmio_write_8(&ioat->regs->chancmp, addr);
} }
static void static void
ioat_write_chainaddr(struct ioat_channel *ioat, uint64_t addr) ioat_write_chainaddr(struct spdk_ioat_chan *ioat, uint64_t addr)
{ {
spdk_mmio_write_8(&ioat->regs->chainaddr, addr); spdk_mmio_write_8(&ioat->regs->chainaddr, addr);
} }
static inline void static inline void
ioat_suspend(struct ioat_channel *ioat) ioat_suspend(struct spdk_ioat_chan *ioat)
{ {
ioat->regs->chancmd = SPDK_IOAT_CHANCMD_SUSPEND; ioat->regs->chancmd = SPDK_IOAT_CHANCMD_SUSPEND;
} }
static inline void static inline void
ioat_reset(struct ioat_channel *ioat) ioat_reset(struct spdk_ioat_chan *ioat)
{ {
ioat->regs->chancmd = SPDK_IOAT_CHANCMD_RESET; ioat->regs->chancmd = SPDK_IOAT_CHANCMD_RESET;
} }
static inline uint32_t static inline uint32_t
ioat_reset_pending(struct ioat_channel *ioat) ioat_reset_pending(struct spdk_ioat_chan *ioat)
{ {
uint8_t cmd; uint8_t cmd;
@ -93,7 +93,7 @@ ioat_reset_pending(struct ioat_channel *ioat)
} }
static int static int
ioat_map_pci_bar(struct ioat_channel *ioat) ioat_map_pci_bar(struct spdk_ioat_chan *ioat)
{ {
int regs_bar, rc; int regs_bar, rc;
void *addr; void *addr;
@ -112,7 +112,7 @@ ioat_map_pci_bar(struct ioat_channel *ioat)
} }
static int static int
ioat_unmap_pci_bar(struct ioat_channel *ioat) ioat_unmap_pci_bar(struct spdk_ioat_chan *ioat)
{ {
int rc = 0; int rc = 0;
void *addr = (void *)ioat->regs; void *addr = (void *)ioat->regs;
@ -125,25 +125,25 @@ ioat_unmap_pci_bar(struct ioat_channel *ioat)
static inline uint32_t static inline uint32_t
ioat_get_active(struct ioat_channel *ioat) ioat_get_active(struct spdk_ioat_chan *ioat)
{ {
return (ioat->head - ioat->tail) & ((1 << ioat->ring_size_order) - 1); return (ioat->head - ioat->tail) & ((1 << ioat->ring_size_order) - 1);
} }
static inline uint32_t static inline uint32_t
ioat_get_ring_space(struct ioat_channel *ioat) ioat_get_ring_space(struct spdk_ioat_chan *ioat)
{ {
return (1 << ioat->ring_size_order) - ioat_get_active(ioat) - 1; return (1 << ioat->ring_size_order) - ioat_get_active(ioat) - 1;
} }
static uint32_t static uint32_t
ioat_get_ring_index(struct ioat_channel *ioat, uint32_t index) ioat_get_ring_index(struct spdk_ioat_chan *ioat, uint32_t index)
{ {
return index & ((1 << ioat->ring_size_order) - 1); return index & ((1 << ioat->ring_size_order) - 1);
} }
static void static void
ioat_get_ring_entry(struct ioat_channel *ioat, uint32_t index, ioat_get_ring_entry(struct spdk_ioat_chan *ioat, uint32_t index,
struct ioat_descriptor **desc, struct ioat_descriptor **desc,
union spdk_ioat_hw_desc **hw_desc) union spdk_ioat_hw_desc **hw_desc)
{ {
@ -154,26 +154,26 @@ ioat_get_ring_entry(struct ioat_channel *ioat, uint32_t index,
} }
static uint64_t static uint64_t
ioat_get_desc_phys_addr(struct ioat_channel *ioat, uint32_t index) ioat_get_desc_phys_addr(struct spdk_ioat_chan *ioat, uint32_t index)
{ {
return ioat->hw_ring_phys_addr + return ioat->hw_ring_phys_addr +
ioat_get_ring_index(ioat, index) * sizeof(union spdk_ioat_hw_desc); ioat_get_ring_index(ioat, index) * sizeof(union spdk_ioat_hw_desc);
} }
static void static void
ioat_submit_single(struct ioat_channel *ioat) ioat_submit_single(struct spdk_ioat_chan *ioat)
{ {
ioat->head++; ioat->head++;
} }
static void static void
ioat_flush(struct ioat_channel *ioat) ioat_flush(struct spdk_ioat_chan *ioat)
{ {
ioat->regs->dmacount = (uint16_t)ioat->head; ioat->regs->dmacount = (uint16_t)ioat->head;
} }
static struct ioat_descriptor * static struct ioat_descriptor *
ioat_prep_null(struct ioat_channel *ioat) ioat_prep_null(struct spdk_ioat_chan *ioat)
{ {
struct ioat_descriptor *desc; struct ioat_descriptor *desc;
union spdk_ioat_hw_desc *hw_desc; union spdk_ioat_hw_desc *hw_desc;
@ -202,7 +202,7 @@ ioat_prep_null(struct ioat_channel *ioat)
} }
static struct ioat_descriptor * static struct ioat_descriptor *
ioat_prep_copy(struct ioat_channel *ioat, uint64_t dst, ioat_prep_copy(struct spdk_ioat_chan *ioat, uint64_t dst,
uint64_t src, uint32_t len) uint64_t src, uint32_t len)
{ {
struct ioat_descriptor *desc; struct ioat_descriptor *desc;
@ -233,7 +233,7 @@ ioat_prep_copy(struct ioat_channel *ioat, uint64_t dst,
} }
static struct ioat_descriptor * static struct ioat_descriptor *
ioat_prep_fill(struct ioat_channel *ioat, uint64_t dst, ioat_prep_fill(struct spdk_ioat_chan *ioat, uint64_t dst,
uint64_t fill_pattern, uint32_t len) uint64_t fill_pattern, uint32_t len)
{ {
struct ioat_descriptor *desc; struct ioat_descriptor *desc;
@ -263,7 +263,7 @@ ioat_prep_fill(struct ioat_channel *ioat, uint64_t dst,
return desc; return desc;
} }
static int ioat_reset_hw(struct ioat_channel *ioat) static int ioat_reset_hw(struct spdk_ioat_chan *ioat)
{ {
int timeout; int timeout;
uint64_t status; uint64_t status;
@ -308,7 +308,7 @@ static int ioat_reset_hw(struct ioat_channel *ioat)
} }
static int static int
ioat_process_channel_events(struct ioat_channel *ioat) ioat_process_channel_events(struct spdk_ioat_chan *ioat)
{ {
struct ioat_descriptor *desc; struct ioat_descriptor *desc;
uint64_t status, completed_descriptor, hw_desc_phys_addr; uint64_t status, completed_descriptor, hw_desc_phys_addr;
@ -347,7 +347,7 @@ ioat_process_channel_events(struct ioat_channel *ioat)
} }
static int static int
ioat_channel_destruct(struct ioat_channel *ioat) ioat_channel_destruct(struct spdk_ioat_chan *ioat)
{ {
ioat_unmap_pci_bar(ioat); ioat_unmap_pci_bar(ioat);
@ -368,7 +368,7 @@ ioat_channel_destruct(struct ioat_channel *ioat)
} }
static int static int
ioat_channel_start(struct ioat_channel *ioat) ioat_channel_start(struct spdk_ioat_chan *ioat)
{ {
uint8_t xfercap, version; uint8_t xfercap, version;
uint64_t status; uint64_t status;
@ -388,9 +388,9 @@ ioat_channel_start(struct ioat_channel *ioat)
} }
/* Always support DMA copy */ /* Always support DMA copy */
ioat->dma_capabilities = IOAT_ENGINE_COPY_SUPPORTED; ioat->dma_capabilities = SPDK_IOAT_ENGINE_COPY_SUPPORTED;
if (ioat->regs->dmacapability & SPDK_IOAT_DMACAP_BFILL) if (ioat->regs->dmacapability & SPDK_IOAT_DMACAP_BFILL)
ioat->dma_capabilities |= IOAT_ENGINE_FILL_SUPPORTED; ioat->dma_capabilities |= SPDK_IOAT_ENGINE_FILL_SUPPORTED;
xfercap = ioat->regs->xfercap; xfercap = ioat->regs->xfercap;
/* Only bits [4:0] are valid. */ /* Only bits [4:0] are valid. */
@ -464,13 +464,13 @@ ioat_channel_start(struct ioat_channel *ioat)
} }
/* Caller must hold g_ioat_driver.lock */ /* Caller must hold g_ioat_driver.lock */
static struct ioat_channel * static struct spdk_ioat_chan *
ioat_attach(void *device) ioat_attach(void *device)
{ {
struct ioat_channel *ioat; struct spdk_ioat_chan *ioat;
uint32_t cmd_reg; uint32_t cmd_reg;
ioat = calloc(1, sizeof(struct ioat_channel)); ioat = calloc(1, sizeof(struct spdk_ioat_chan));
if (ioat == NULL) { if (ioat == NULL) {
return NULL; return NULL;
} }
@ -494,8 +494,8 @@ ioat_attach(void *device)
} }
struct ioat_enum_ctx { struct ioat_enum_ctx {
ioat_probe_cb probe_cb; spdk_ioat_probe_cb probe_cb;
ioat_attach_cb attach_cb; spdk_ioat_attach_cb attach_cb;
void *cb_ctx; void *cb_ctx;
}; };
@ -504,7 +504,7 @@ static int
ioat_enum_cb(void *ctx, struct spdk_pci_device *pci_dev) ioat_enum_cb(void *ctx, struct spdk_pci_device *pci_dev)
{ {
struct ioat_enum_ctx *enum_ctx = ctx; struct ioat_enum_ctx *enum_ctx = ctx;
struct ioat_channel *ioat; struct spdk_ioat_chan *ioat;
/* Verify that this device is not already attached */ /* Verify that this device is not already attached */
TAILQ_FOREACH(ioat, &g_ioat_driver.attached_chans, tailq) { TAILQ_FOREACH(ioat, &g_ioat_driver.attached_chans, tailq) {
@ -538,7 +538,7 @@ ioat_enum_cb(void *ctx, struct spdk_pci_device *pci_dev)
} }
int int
ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb) spdk_ioat_probe(void *cb_ctx, spdk_ioat_probe_cb probe_cb, spdk_ioat_attach_cb attach_cb)
{ {
int rc; int rc;
struct ioat_enum_ctx enum_ctx; struct ioat_enum_ctx enum_ctx;
@ -557,7 +557,7 @@ ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb)
} }
int int
ioat_detach(struct ioat_channel *ioat) spdk_ioat_detach(struct spdk_ioat_chan *ioat)
{ {
struct ioat_driver *driver = &g_ioat_driver; struct ioat_driver *driver = &g_ioat_driver;
@ -565,7 +565,7 @@ ioat_detach(struct ioat_channel *ioat)
* when calling ioat_detach(). * when calling ioat_detach().
*/ */
ioat_mutex_lock(&driver->lock); ioat_mutex_lock(&driver->lock);
SLIST_REMOVE(&ioat_free_channels, ioat, ioat_channel, next); SLIST_REMOVE(&ioat_free_channels, ioat, spdk_ioat_chan, next);
TAILQ_REMOVE(&driver->attached_chans, ioat, tailq); TAILQ_REMOVE(&driver->attached_chans, ioat, tailq);
ioat_mutex_unlock(&driver->lock); ioat_mutex_unlock(&driver->lock);
@ -576,7 +576,7 @@ ioat_detach(struct ioat_channel *ioat)
} }
int int
ioat_register_thread(void) spdk_ioat_register_thread(void)
{ {
struct ioat_driver *driver = &g_ioat_driver; struct ioat_driver *driver = &g_ioat_driver;
@ -598,7 +598,7 @@ ioat_register_thread(void)
} }
void void
ioat_unregister_thread(void) spdk_ioat_unregister_thread(void)
{ {
struct ioat_driver *driver = &g_ioat_driver; struct ioat_driver *driver = &g_ioat_driver;
@ -620,10 +620,10 @@ ioat_unregister_thread(void)
#define _2MB_OFFSET(ptr) ((ptr) & (0x200000 - 1)) #define _2MB_OFFSET(ptr) ((ptr) & (0x200000 - 1))
int64_t int64_t
ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn, spdk_ioat_submit_copy(void *cb_arg, spdk_ioat_req_cb cb_fn,
void *dst, const void *src, uint64_t nbytes) void *dst, const void *src, uint64_t nbytes)
{ {
struct ioat_channel *ioat; struct spdk_ioat_chan *ioat;
struct ioat_descriptor *last_desc; struct ioat_descriptor *last_desc;
uint64_t remaining, op_size; uint64_t remaining, op_size;
uint64_t vdst, vsrc; uint64_t vdst, vsrc;
@ -698,10 +698,10 @@ ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn,
} }
int64_t int64_t
ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn, spdk_ioat_submit_fill(void *cb_arg, spdk_ioat_req_cb cb_fn,
void *dst, uint64_t fill_pattern, uint64_t nbytes) void *dst, uint64_t fill_pattern, uint64_t nbytes)
{ {
struct ioat_channel *ioat; struct spdk_ioat_chan *ioat;
struct ioat_descriptor *last_desc = NULL; struct ioat_descriptor *last_desc = NULL;
uint64_t remaining, op_size; uint64_t remaining, op_size;
uint64_t vdst; uint64_t vdst;
@ -712,7 +712,7 @@ ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn,
return -1; return -1;
} }
if (!(ioat->dma_capabilities & IOAT_ENGINE_FILL_SUPPORTED)) { if (!(ioat->dma_capabilities & SPDK_IOAT_ENGINE_FILL_SUPPORTED)) {
ioat_printf(ioat, "Channel does not support memory fill\n"); ioat_printf(ioat, "Channel does not support memory fill\n");
return -1; return -1;
} }
@ -755,9 +755,10 @@ ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn,
return nbytes; return nbytes;
} }
uint32_t ioat_get_dma_capabilities(void) uint32_t
spdk_ioat_get_dma_capabilities(void)
{ {
struct ioat_channel *ioat; struct spdk_ioat_chan *ioat;
ioat = ioat_thread_channel; ioat = ioat_thread_channel;
if (!ioat) { if (!ioat) {
@ -766,7 +767,8 @@ uint32_t ioat_get_dma_capabilities(void)
return ioat->dma_capabilities; return ioat->dma_capabilities;
} }
int ioat_process_events(void) int
spdk_ioat_process_events(void)
{ {
if (!ioat_thread_channel) { if (!ioat_thread_channel) {
return -1; return -1;

View File

@ -50,13 +50,13 @@
#define IOAT_DEFAULT_ORDER 15 #define IOAT_DEFAULT_ORDER 15
struct ioat_descriptor { struct ioat_descriptor {
ioat_callback_t callback_fn; spdk_ioat_req_cb callback_fn;
void *callback_arg; void *callback_arg;
}; };
/* One of these per allocated PCI device. */ /* One of these per allocated PCI device. */
struct ioat_channel { struct spdk_ioat_chan {
SLIST_ENTRY(ioat_channel) next; SLIST_ENTRY(spdk_ioat_chan) next;
/* Opaque handle to upper layer */ /* Opaque handle to upper layer */
void *device; void *device;
@ -77,7 +77,7 @@ struct ioat_channel {
uint32_t dma_capabilities; uint32_t dma_capabilities;
/* tailq entry for attached_chans */ /* tailq entry for attached_chans */
TAILQ_ENTRY(ioat_channel) tailq; TAILQ_ENTRY(spdk_ioat_chan) tailq;
}; };
static inline uint32_t static inline uint32_t