diff --git a/doc/ioat.index.txt b/doc/ioat.index.txt index d5b3d5a00..181f5493a 100644 --- a/doc/ioat.index.txt +++ b/doc/ioat.index.txt @@ -41,8 +41,10 @@ \section key_functions Key Functions -- ioat_attach() \copybrief ioat_attach() -- ioat_register_thread() \copybrief ioat_register_thread() -- ioat_submit_copy() \copybrief ioat_submit_copy() +- spdk_ioat_probe() \copybrief spdk_ioat_probe() +- spdk_ioat_register_thread() \copybrief spdk_ioat_register_thread() +- spdk_ioat_get_dma_capabilities() \copybrief spdk_ioat_get_dma_capabilities() +- spdk_ioat_submit_copy() \copybrief spdk_ioat_submit_copy() +- spdk_ioat_submit_fill() \copybrief spdk_ioat_submit_fill() */ diff --git a/examples/ioat/perf/perf.c b/examples/ioat/perf/perf.c index 9d99ff8ff..b81e99573 100644 --- a/examples/ioat/perf/perf.c +++ b/examples/ioat/perf/perf.c @@ -56,7 +56,7 @@ struct user_config { }; struct ioat_device { - struct ioat_channel *ioat; + struct spdk_ioat_chan *ioat; TAILQ_ENTRY(ioat_device) tailq; }; @@ -113,7 +113,7 @@ ioat_exit(void) dev = TAILQ_FIRST(&g_devices); TAILQ_REMOVE(&g_devices, dev, tailq); if (dev->ioat) { - ioat_detach(dev->ioat); + spdk_ioat_detach(dev->ioat); } rte_free(dev); } @@ -161,7 +161,7 @@ probe_cb(void *cb_ctx, struct spdk_pci_device *pci_dev) } static void -attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct ioat_channel *ioat) +attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct spdk_ioat_chan *ioat) { struct ioat_device *dev; @@ -180,7 +180,7 @@ ioat_init(void) { TAILQ_INIT(&g_devices); - if (ioat_probe(NULL, probe_cb, attach_cb) != 0) { + if (spdk_ioat_probe(NULL, probe_cb, attach_cb) != 0) { fprintf(stderr, "ioat_probe() failed\n"); return 1; } @@ -244,7 +244,7 @@ static void drain_io(struct thread_entry *thread_entry) { while (thread_entry->current_queue_depth > 0) { - ioat_process_events(); + spdk_ioat_process_events(); } } @@ -256,7 +256,7 @@ submit_single_xfer(struct thread_entry *thread_entry, struct ioat_task *ioat_tas ioat_task->src = src; ioat_task->dst = dst; - ioat_submit_copy(ioat_task, ioat_done, dst, src, g_user_config.xfer_size_bytes); + spdk_ioat_submit_copy(ioat_task, ioat_done, dst, src, g_user_config.xfer_size_bytes); thread_entry->current_queue_depth++; } @@ -297,7 +297,7 @@ work_fn(void *arg) return 1; } - if (ioat_register_thread() != 0) { + if (spdk_ioat_register_thread() != 0) { fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n", rte_lcore_id()); return 0; @@ -308,13 +308,13 @@ work_fn(void *arg) // begin to submit transfers submit_xfers(t, g_user_config.queue_depth); while (rte_get_timer_cycles() < tsc_end) { - ioat_process_events(); + spdk_ioat_process_events(); } // begin to drain io t->is_draining = true; drain_io(t); - ioat_unregister_thread(); + spdk_ioat_unregister_thread(); return 0; } diff --git a/examples/ioat/verify/verify.c b/examples/ioat/verify/verify.c index 742405841..4294bdce9 100644 --- a/examples/ioat/verify/verify.c +++ b/examples/ioat/verify/verify.c @@ -61,7 +61,7 @@ struct user_config { }; struct ioat_device { - struct ioat_channel *ioat; + struct spdk_ioat_chan *ioat; TAILQ_ENTRY(ioat_device) tailq; }; @@ -123,7 +123,7 @@ ioat_exit(void) dev = TAILQ_FIRST(&g_devices); TAILQ_REMOVE(&g_devices, dev, tailq); if (dev->ioat) { - ioat_detach(dev->ioat); + spdk_ioat_detach(dev->ioat); } free(dev); } @@ -217,7 +217,7 @@ probe_cb(void *cb_ctx, struct spdk_pci_device *pci_dev) } static void -attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct ioat_channel *ioat) +attach_cb(void *cb_ctx, struct spdk_pci_device *pci_dev, struct spdk_ioat_chan *ioat) { struct ioat_device *dev; @@ -237,7 +237,7 @@ ioat_init(void) { TAILQ_INIT(&g_devices); - if (ioat_probe(NULL, probe_cb, attach_cb) != 0) { + if (spdk_ioat_probe(NULL, probe_cb, attach_cb) != 0) { fprintf(stderr, "ioat_probe() failed\n"); return 1; } @@ -292,7 +292,7 @@ static void drain_xfers(struct thread_entry *thread_entry) { while (thread_entry->current_queue_depth > 0) { - ioat_process_events(); + spdk_ioat_process_events(); } } @@ -300,9 +300,10 @@ static void submit_single_xfer(struct ioat_task *ioat_task) { if (ioat_task->type == IOAT_FILL_TYPE) - ioat_submit_fill(ioat_task, ioat_done, ioat_task->dst, ioat_task->fill_pattern, ioat_task->len); + spdk_ioat_submit_fill(ioat_task, ioat_done, ioat_task->dst, ioat_task->fill_pattern, + ioat_task->len); else - ioat_submit_copy(ioat_task, ioat_done, ioat_task->dst, ioat_task->src, ioat_task->len); + spdk_ioat_submit_copy(ioat_task, ioat_done, ioat_task->dst, ioat_task->src, ioat_task->len); ioat_task->thread_entry->current_queue_depth++; } @@ -315,7 +316,7 @@ submit_xfers(struct thread_entry *thread_entry, uint64_t queue_depth) rte_mempool_get(thread_entry->data_pool, &(ioat_task->buffer)); ioat_task->type = IOAT_COPY_TYPE; - if (ioat_get_dma_capabilities() & IOAT_ENGINE_FILL_SUPPORTED) { + if (spdk_ioat_get_dma_capabilities() & SPDK_IOAT_ENGINE_FILL_SUPPORTED) { if (queue_depth % 2) ioat_task->type = IOAT_FILL_TYPE; } @@ -346,7 +347,7 @@ work_fn(void *arg) return 1; } - if (ioat_register_thread() != 0) { + if (spdk_ioat_register_thread() != 0) { fprintf(stderr, "lcore %u: No ioat channels found. Check that ioatdma driver is unloaded.\n", rte_lcore_id()); return 0; @@ -356,13 +357,13 @@ work_fn(void *arg) submit_xfers(t, g_user_config.queue_depth); while (rte_get_timer_cycles() < tsc_end) { - ioat_process_events(); + spdk_ioat_process_events(); } t->is_draining = true; drain_xfers(t); - ioat_unregister_thread(); + spdk_ioat_unregister_thread(); return 0; } diff --git a/include/spdk/ioat.h b/include/spdk/ioat.h index b69cc0ccc..e64caa571 100644 --- a/include/spdk/ioat.h +++ b/include/spdk/ioat.h @@ -35,8 +35,8 @@ * This file defines the public interface to the I/OAT DMA engine driver. */ -#ifndef __IOAT_H__ -#define __IOAT_H__ +#ifndef SPDK_IOAT_H +#define SPDK_IOAT_H #include #include @@ -45,27 +45,27 @@ #include "spdk/pci.h" /** - * Opaque handle for a single I/OAT channel returned by \ref ioat_probe(). + * Opaque handle for a single I/OAT channel returned by \ref spdk_ioat_probe(). */ -struct ioat_channel; +struct spdk_ioat_chan; /** * Signature for callback function invoked when a request is completed. */ -typedef void (*ioat_callback_t)(void *arg); +typedef void (*spdk_ioat_req_cb)(void *arg); /** - * Callback for ioat_probe() enumeration. + * Callback for spdk_ioat_probe() enumeration. * * \return true to attach to this device. */ -typedef bool (*ioat_probe_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev); +typedef bool (*spdk_ioat_probe_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev); /** - * Callback for ioat_probe() to report a device that has been attached to the userspace I/OAT driver. + * Callback for spdk_ioat_probe() to report a device that has been attached to the userspace I/OAT driver. */ -typedef void (*ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev, - struct ioat_channel *ioat); +typedef void (*spdk_ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev, + struct spdk_ioat_chan *ioat); /** * \brief Enumerate the I/OAT devices attached to the system and attach the userspace I/OAT driver @@ -79,74 +79,74 @@ typedef void (*ioat_attach_cb)(void *cb_ctx, struct spdk_pci_device *pci_dev, * will be reported. * * To stop using the the controller and release its associated resources, - * call \ref ioat_detach with the ioat_channel instance returned by this function. + * call \ref spdk_ioat_detach with the ioat_channel instance returned by this function. */ -int ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb); +int spdk_ioat_probe(void *cb_ctx, spdk_ioat_probe_cb probe_cb, spdk_ioat_attach_cb attach_cb); /** - * Detaches specified device returned by \ref ioat_probe() from the I/OAT driver. + * Detaches specified device returned by \ref spdk_ioat_probe() from the I/OAT driver. */ -int ioat_detach(struct ioat_channel *ioat); +int spdk_ioat_detach(struct spdk_ioat_chan *ioat); /** * Request a DMA engine channel for the calling thread. * * Must be called before submitting any requests from a thread. * - * The \ref ioat_unregister_thread() function can be called to release the channel. + * The \ref spdk_ioat_unregister_thread() function can be called to release the channel. */ -int ioat_register_thread(void); +int spdk_ioat_register_thread(void); /** * Unregister the current thread's I/OAT channel. * - * This function can be called after \ref ioat_register_thread() to release the thread's + * This function can be called after \ref spdk_ioat_register_thread() to release the thread's * DMA engine channel for use by other threads. */ -void ioat_unregister_thread(void); +void spdk_ioat_unregister_thread(void); /** * Submit a DMA engine memory copy request. * * Before submitting any requests on a thread, the thread must be registered - * using the \ref ioat_register_thread() function. + * using the \ref spdk_ioat_register_thread() function. */ -int64_t ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn, - void *dst, const void *src, uint64_t nbytes); +int64_t spdk_ioat_submit_copy(void *cb_arg, spdk_ioat_req_cb cb_fn, + void *dst, const void *src, uint64_t nbytes); /** * Submit a DMA engine memory fill request. * * Before submitting any requests on a thread, the thread must be registered - * using the \ref ioat_register_thread() function. + * using the \ref spdk_ioat_register_thread() function. */ -int64_t ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn, - void *dst, uint64_t fill_pattern, uint64_t nbytes); +int64_t spdk_ioat_submit_fill(void *cb_arg, spdk_ioat_req_cb cb_fn, + void *dst, uint64_t fill_pattern, uint64_t nbytes); /** * Check for completed requests on the current thread. * * Before submitting any requests on a thread, the thread must be registered - * using the \ref ioat_register_thread() function. + * using the \ref spdk_ioat_register_thread() function. * * \returns 0 on success or negative if something went wrong. */ -int ioat_process_events(void); +int spdk_ioat_process_events(void); /** * DMA engine capability flags */ -enum ioat_dma_capability_flags { - IOAT_ENGINE_COPY_SUPPORTED = 0x1, /**< The memory copy is supported */ - IOAT_ENGINE_FILL_SUPPORTED = 0x2, /**< The memory fill is supported */ +enum spdk_ioat_dma_capability_flags { + SPDK_IOAT_ENGINE_COPY_SUPPORTED = 0x1, /**< The memory copy is supported */ + SPDK_IOAT_ENGINE_FILL_SUPPORTED = 0x2, /**< The memory fill is supported */ }; /** * Get the DMA engine capabilities. * * Before submitting any requests on a thread, the thread must be registered - * using the \ref ioat_register_thread() function. + * using the \ref spdk_ioat_register_thread() function. */ -uint32_t ioat_get_dma_capabilities(void); +uint32_t spdk_ioat_get_dma_capabilities(void); #endif diff --git a/lib/ioat/ioat.c b/lib/ioat/ioat.c index d2e40c116..bbbf81988 100644 --- a/lib/ioat/ioat.c +++ b/lib/ioat/ioat.c @@ -38,14 +38,14 @@ * * Must hold g_ioat_driver.lock while manipulating this list. */ -static SLIST_HEAD(, ioat_channel) ioat_free_channels; +static SLIST_HEAD(, spdk_ioat_chan) ioat_free_channels; /** IOAT channel assigned to this thread (or NULL if not assigned yet). */ -static __thread struct ioat_channel *ioat_thread_channel; +static __thread struct spdk_ioat_chan *ioat_thread_channel; struct ioat_driver { ioat_mutex_t lock; - TAILQ_HEAD(, ioat_channel) attached_chans; + TAILQ_HEAD(, spdk_ioat_chan) attached_chans; }; static struct ioat_driver g_ioat_driver = { @@ -54,37 +54,37 @@ static struct ioat_driver g_ioat_driver = { }; static uint64_t -ioat_get_chansts(struct ioat_channel *ioat) +ioat_get_chansts(struct spdk_ioat_chan *ioat) { return spdk_mmio_read_8(&ioat->regs->chansts); } static void -ioat_write_chancmp(struct ioat_channel *ioat, uint64_t addr) +ioat_write_chancmp(struct spdk_ioat_chan *ioat, uint64_t addr) { spdk_mmio_write_8(&ioat->regs->chancmp, addr); } static void -ioat_write_chainaddr(struct ioat_channel *ioat, uint64_t addr) +ioat_write_chainaddr(struct spdk_ioat_chan *ioat, uint64_t addr) { spdk_mmio_write_8(&ioat->regs->chainaddr, addr); } static inline void -ioat_suspend(struct ioat_channel *ioat) +ioat_suspend(struct spdk_ioat_chan *ioat) { ioat->regs->chancmd = SPDK_IOAT_CHANCMD_SUSPEND; } static inline void -ioat_reset(struct ioat_channel *ioat) +ioat_reset(struct spdk_ioat_chan *ioat) { ioat->regs->chancmd = SPDK_IOAT_CHANCMD_RESET; } static inline uint32_t -ioat_reset_pending(struct ioat_channel *ioat) +ioat_reset_pending(struct spdk_ioat_chan *ioat) { uint8_t cmd; @@ -93,7 +93,7 @@ ioat_reset_pending(struct ioat_channel *ioat) } static int -ioat_map_pci_bar(struct ioat_channel *ioat) +ioat_map_pci_bar(struct spdk_ioat_chan *ioat) { int regs_bar, rc; void *addr; @@ -112,7 +112,7 @@ ioat_map_pci_bar(struct ioat_channel *ioat) } static int -ioat_unmap_pci_bar(struct ioat_channel *ioat) +ioat_unmap_pci_bar(struct spdk_ioat_chan *ioat) { int rc = 0; void *addr = (void *)ioat->regs; @@ -125,25 +125,25 @@ ioat_unmap_pci_bar(struct ioat_channel *ioat) static inline uint32_t -ioat_get_active(struct ioat_channel *ioat) +ioat_get_active(struct spdk_ioat_chan *ioat) { return (ioat->head - ioat->tail) & ((1 << ioat->ring_size_order) - 1); } static inline uint32_t -ioat_get_ring_space(struct ioat_channel *ioat) +ioat_get_ring_space(struct spdk_ioat_chan *ioat) { return (1 << ioat->ring_size_order) - ioat_get_active(ioat) - 1; } static uint32_t -ioat_get_ring_index(struct ioat_channel *ioat, uint32_t index) +ioat_get_ring_index(struct spdk_ioat_chan *ioat, uint32_t index) { return index & ((1 << ioat->ring_size_order) - 1); } static void -ioat_get_ring_entry(struct ioat_channel *ioat, uint32_t index, +ioat_get_ring_entry(struct spdk_ioat_chan *ioat, uint32_t index, struct ioat_descriptor **desc, union spdk_ioat_hw_desc **hw_desc) { @@ -154,26 +154,26 @@ ioat_get_ring_entry(struct ioat_channel *ioat, uint32_t index, } static uint64_t -ioat_get_desc_phys_addr(struct ioat_channel *ioat, uint32_t index) +ioat_get_desc_phys_addr(struct spdk_ioat_chan *ioat, uint32_t index) { return ioat->hw_ring_phys_addr + ioat_get_ring_index(ioat, index) * sizeof(union spdk_ioat_hw_desc); } static void -ioat_submit_single(struct ioat_channel *ioat) +ioat_submit_single(struct spdk_ioat_chan *ioat) { ioat->head++; } static void -ioat_flush(struct ioat_channel *ioat) +ioat_flush(struct spdk_ioat_chan *ioat) { ioat->regs->dmacount = (uint16_t)ioat->head; } static struct ioat_descriptor * -ioat_prep_null(struct ioat_channel *ioat) +ioat_prep_null(struct spdk_ioat_chan *ioat) { struct ioat_descriptor *desc; union spdk_ioat_hw_desc *hw_desc; @@ -202,7 +202,7 @@ ioat_prep_null(struct ioat_channel *ioat) } static struct ioat_descriptor * -ioat_prep_copy(struct ioat_channel *ioat, uint64_t dst, +ioat_prep_copy(struct spdk_ioat_chan *ioat, uint64_t dst, uint64_t src, uint32_t len) { struct ioat_descriptor *desc; @@ -233,7 +233,7 @@ ioat_prep_copy(struct ioat_channel *ioat, uint64_t dst, } static struct ioat_descriptor * -ioat_prep_fill(struct ioat_channel *ioat, uint64_t dst, +ioat_prep_fill(struct spdk_ioat_chan *ioat, uint64_t dst, uint64_t fill_pattern, uint32_t len) { struct ioat_descriptor *desc; @@ -263,7 +263,7 @@ ioat_prep_fill(struct ioat_channel *ioat, uint64_t dst, return desc; } -static int ioat_reset_hw(struct ioat_channel *ioat) +static int ioat_reset_hw(struct spdk_ioat_chan *ioat) { int timeout; uint64_t status; @@ -308,7 +308,7 @@ static int ioat_reset_hw(struct ioat_channel *ioat) } static int -ioat_process_channel_events(struct ioat_channel *ioat) +ioat_process_channel_events(struct spdk_ioat_chan *ioat) { struct ioat_descriptor *desc; uint64_t status, completed_descriptor, hw_desc_phys_addr; @@ -347,7 +347,7 @@ ioat_process_channel_events(struct ioat_channel *ioat) } static int -ioat_channel_destruct(struct ioat_channel *ioat) +ioat_channel_destruct(struct spdk_ioat_chan *ioat) { ioat_unmap_pci_bar(ioat); @@ -368,7 +368,7 @@ ioat_channel_destruct(struct ioat_channel *ioat) } static int -ioat_channel_start(struct ioat_channel *ioat) +ioat_channel_start(struct spdk_ioat_chan *ioat) { uint8_t xfercap, version; uint64_t status; @@ -388,9 +388,9 @@ ioat_channel_start(struct ioat_channel *ioat) } /* Always support DMA copy */ - ioat->dma_capabilities = IOAT_ENGINE_COPY_SUPPORTED; + ioat->dma_capabilities = SPDK_IOAT_ENGINE_COPY_SUPPORTED; if (ioat->regs->dmacapability & SPDK_IOAT_DMACAP_BFILL) - ioat->dma_capabilities |= IOAT_ENGINE_FILL_SUPPORTED; + ioat->dma_capabilities |= SPDK_IOAT_ENGINE_FILL_SUPPORTED; xfercap = ioat->regs->xfercap; /* Only bits [4:0] are valid. */ @@ -464,13 +464,13 @@ ioat_channel_start(struct ioat_channel *ioat) } /* Caller must hold g_ioat_driver.lock */ -static struct ioat_channel * +static struct spdk_ioat_chan * ioat_attach(void *device) { - struct ioat_channel *ioat; + struct spdk_ioat_chan *ioat; uint32_t cmd_reg; - ioat = calloc(1, sizeof(struct ioat_channel)); + ioat = calloc(1, sizeof(struct spdk_ioat_chan)); if (ioat == NULL) { return NULL; } @@ -494,8 +494,8 @@ ioat_attach(void *device) } struct ioat_enum_ctx { - ioat_probe_cb probe_cb; - ioat_attach_cb attach_cb; + spdk_ioat_probe_cb probe_cb; + spdk_ioat_attach_cb attach_cb; void *cb_ctx; }; @@ -504,7 +504,7 @@ static int ioat_enum_cb(void *ctx, struct spdk_pci_device *pci_dev) { struct ioat_enum_ctx *enum_ctx = ctx; - struct ioat_channel *ioat; + struct spdk_ioat_chan *ioat; /* Verify that this device is not already attached */ TAILQ_FOREACH(ioat, &g_ioat_driver.attached_chans, tailq) { @@ -538,7 +538,7 @@ ioat_enum_cb(void *ctx, struct spdk_pci_device *pci_dev) } int -ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb) +spdk_ioat_probe(void *cb_ctx, spdk_ioat_probe_cb probe_cb, spdk_ioat_attach_cb attach_cb) { int rc; struct ioat_enum_ctx enum_ctx; @@ -557,7 +557,7 @@ ioat_probe(void *cb_ctx, ioat_probe_cb probe_cb, ioat_attach_cb attach_cb) } int -ioat_detach(struct ioat_channel *ioat) +spdk_ioat_detach(struct spdk_ioat_chan *ioat) { struct ioat_driver *driver = &g_ioat_driver; @@ -565,7 +565,7 @@ ioat_detach(struct ioat_channel *ioat) * when calling ioat_detach(). */ ioat_mutex_lock(&driver->lock); - SLIST_REMOVE(&ioat_free_channels, ioat, ioat_channel, next); + SLIST_REMOVE(&ioat_free_channels, ioat, spdk_ioat_chan, next); TAILQ_REMOVE(&driver->attached_chans, ioat, tailq); ioat_mutex_unlock(&driver->lock); @@ -576,7 +576,7 @@ ioat_detach(struct ioat_channel *ioat) } int -ioat_register_thread(void) +spdk_ioat_register_thread(void) { struct ioat_driver *driver = &g_ioat_driver; @@ -598,7 +598,7 @@ ioat_register_thread(void) } void -ioat_unregister_thread(void) +spdk_ioat_unregister_thread(void) { struct ioat_driver *driver = &g_ioat_driver; @@ -620,10 +620,10 @@ ioat_unregister_thread(void) #define _2MB_OFFSET(ptr) ((ptr) & (0x200000 - 1)) int64_t -ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn, - void *dst, const void *src, uint64_t nbytes) +spdk_ioat_submit_copy(void *cb_arg, spdk_ioat_req_cb cb_fn, + void *dst, const void *src, uint64_t nbytes) { - struct ioat_channel *ioat; + struct spdk_ioat_chan *ioat; struct ioat_descriptor *last_desc; uint64_t remaining, op_size; uint64_t vdst, vsrc; @@ -698,10 +698,10 @@ ioat_submit_copy(void *cb_arg, ioat_callback_t cb_fn, } int64_t -ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn, - void *dst, uint64_t fill_pattern, uint64_t nbytes) +spdk_ioat_submit_fill(void *cb_arg, spdk_ioat_req_cb cb_fn, + void *dst, uint64_t fill_pattern, uint64_t nbytes) { - struct ioat_channel *ioat; + struct spdk_ioat_chan *ioat; struct ioat_descriptor *last_desc = NULL; uint64_t remaining, op_size; uint64_t vdst; @@ -712,7 +712,7 @@ ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn, return -1; } - if (!(ioat->dma_capabilities & IOAT_ENGINE_FILL_SUPPORTED)) { + if (!(ioat->dma_capabilities & SPDK_IOAT_ENGINE_FILL_SUPPORTED)) { ioat_printf(ioat, "Channel does not support memory fill\n"); return -1; } @@ -755,9 +755,10 @@ ioat_submit_fill(void *cb_arg, ioat_callback_t cb_fn, return nbytes; } -uint32_t ioat_get_dma_capabilities(void) +uint32_t +spdk_ioat_get_dma_capabilities(void) { - struct ioat_channel *ioat; + struct spdk_ioat_chan *ioat; ioat = ioat_thread_channel; if (!ioat) { @@ -766,7 +767,8 @@ uint32_t ioat_get_dma_capabilities(void) return ioat->dma_capabilities; } -int ioat_process_events(void) +int +spdk_ioat_process_events(void) { if (!ioat_thread_channel) { return -1; diff --git a/lib/ioat/ioat_internal.h b/lib/ioat/ioat_internal.h index 532f83e1e..9b23eef09 100644 --- a/lib/ioat/ioat_internal.h +++ b/lib/ioat/ioat_internal.h @@ -50,13 +50,13 @@ #define IOAT_DEFAULT_ORDER 15 struct ioat_descriptor { - ioat_callback_t callback_fn; + spdk_ioat_req_cb callback_fn; void *callback_arg; }; /* One of these per allocated PCI device. */ -struct ioat_channel { - SLIST_ENTRY(ioat_channel) next; +struct spdk_ioat_chan { + SLIST_ENTRY(spdk_ioat_chan) next; /* Opaque handle to upper layer */ void *device; @@ -77,7 +77,7 @@ struct ioat_channel { uint32_t dma_capabilities; /* tailq entry for attached_chans */ - TAILQ_ENTRY(ioat_channel) tailq; + TAILQ_ENTRY(spdk_ioat_chan) tailq; }; static inline uint32_t