diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-27 18:24:20 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-27 18:24:20 +0000 |
commit | 483eb2f56657e8e7f419ab1a4fab8dce9ade8609 (patch) | |
tree | e5d88d25d870d5dedacb6bbdbe2a966086a0a5cf /src/spdk/lib/event/reactor.c | |
parent | Initial commit. (diff) | |
download | ceph-483eb2f56657e8e7f419ab1a4fab8dce9ade8609.tar.xz ceph-483eb2f56657e8e7f419ab1a4fab8dce9ade8609.zip |
Adding upstream version 14.2.21.upstream/14.2.21upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/spdk/lib/event/reactor.c')
-rw-r--r-- | src/spdk/lib/event/reactor.c | 804 |
1 files changed, 804 insertions, 0 deletions
diff --git a/src/spdk/lib/event/reactor.c b/src/spdk/lib/event/reactor.c new file mode 100644 index 00000000..d9ba9f6b --- /dev/null +++ b/src/spdk/lib/event/reactor.c @@ -0,0 +1,804 @@ +/*- + * BSD LICENSE + * + * Copyright (c) Intel Corporation. + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * + * * Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * * Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in + * the documentation and/or other materials provided with the + * distribution. + * * Neither the name of Intel Corporation nor the names of its + * contributors may be used to endorse or promote products derived + * from this software without specific prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "spdk/stdinc.h" +#include "spdk/likely.h" + +#include "spdk_internal/event.h" +#include "spdk_internal/log.h" + +#include "spdk/log.h" +#include "spdk/thread.h" +#include "spdk/env.h" +#include "spdk/util.h" + +#define SPDK_MAX_SOCKET 64 + +#define SPDK_EVENT_BATCH_SIZE 8 + +enum spdk_poller_state { + /* The poller is registered with a reactor but not currently executing its fn. */ + SPDK_POLLER_STATE_WAITING, + + /* The poller is currently running its fn. */ + SPDK_POLLER_STATE_RUNNING, + + /* The poller was unregistered during the execution of its fn. */ + SPDK_POLLER_STATE_UNREGISTERED, +}; + +struct spdk_poller { + TAILQ_ENTRY(spdk_poller) tailq; + uint32_t lcore; + + /* Current state of the poller; should only be accessed from the poller's thread. */ + enum spdk_poller_state state; + + uint64_t period_ticks; + uint64_t next_run_tick; + spdk_poller_fn fn; + void *arg; +}; + +enum spdk_reactor_state { + SPDK_REACTOR_STATE_INVALID = 0, + SPDK_REACTOR_STATE_INITIALIZED = 1, + SPDK_REACTOR_STATE_RUNNING = 2, + SPDK_REACTOR_STATE_EXITING = 3, + SPDK_REACTOR_STATE_SHUTDOWN = 4, +}; + +struct spdk_reactor { + /* Logical core number for this reactor. */ + uint32_t lcore; + + /* Socket ID for this reactor. */ + uint32_t socket_id; + + /* Poller for get the rusage for the reactor. */ + struct spdk_poller *rusage_poller; + + /* Reactor tsc stats */ + struct spdk_reactor_tsc_stats tsc_stats; + + uint64_t tsc_last; + + /* The last known rusage values */ + struct rusage rusage; + + /* + * Contains pollers actively running on this reactor. Pollers + * are run round-robin. The reactor takes one poller from the head + * of the ring, executes it, then puts it back at the tail of + * the ring. + */ + TAILQ_HEAD(, spdk_poller) active_pollers; + + /** + * Contains pollers running on this reactor with a periodic timer. + */ + TAILQ_HEAD(timer_pollers_head, spdk_poller) timer_pollers; + + struct spdk_ring *events; + + /* Pointer to the per-socket g_spdk_event_mempool for this reactor. */ + struct spdk_mempool *event_mempool; + + uint64_t max_delay_us; +} __attribute__((aligned(64))); + +static struct spdk_reactor *g_reactors; + +static enum spdk_reactor_state g_reactor_state = SPDK_REACTOR_STATE_INVALID; + +static bool g_context_switch_monitor_enabled = true; + +static void spdk_reactor_construct(struct spdk_reactor *w, uint32_t lcore, + uint64_t max_delay_us); + +static struct spdk_mempool *g_spdk_event_mempool[SPDK_MAX_SOCKET]; + +static struct spdk_cpuset *g_spdk_app_core_mask; + +static struct spdk_reactor * +spdk_reactor_get(uint32_t lcore) +{ + struct spdk_reactor *reactor; + reactor = spdk_likely(g_reactors) ? &g_reactors[lcore] : NULL; + return reactor; +} + +struct spdk_event * +spdk_event_allocate(uint32_t lcore, spdk_event_fn fn, void *arg1, void *arg2) +{ + struct spdk_event *event = NULL; + struct spdk_reactor *reactor = spdk_reactor_get(lcore); + + if (!reactor) { + assert(false); + return NULL; + } + + event = spdk_mempool_get(reactor->event_mempool); + if (event == NULL) { + assert(false); + return NULL; + } + + event->lcore = lcore; + event->fn = fn; + event->arg1 = arg1; + event->arg2 = arg2; + + return event; +} + +void +spdk_event_call(struct spdk_event *event) +{ + int rc; + struct spdk_reactor *reactor; + + reactor = spdk_reactor_get(event->lcore); + + assert(reactor->events != NULL); + rc = spdk_ring_enqueue(reactor->events, (void **)&event, 1); + if (rc != 1) { + assert(false); + } +} + +static inline uint32_t +_spdk_event_queue_run_batch(struct spdk_reactor *reactor) +{ + unsigned count, i; + void *events[SPDK_EVENT_BATCH_SIZE]; + +#ifdef DEBUG + /* + * spdk_ring_dequeue() fills events and returns how many entries it wrote, + * so we will never actually read uninitialized data from events, but just to be sure + * (and to silence a static analyzer false positive), initialize the array to NULL pointers. + */ + memset(events, 0, sizeof(events)); +#endif + + count = spdk_ring_dequeue(reactor->events, events, SPDK_EVENT_BATCH_SIZE); + if (count == 0) { + return 0; + } + + for (i = 0; i < count; i++) { + struct spdk_event *event = events[i]; + + assert(event != NULL); + event->fn(event->arg1, event->arg2); + } + + spdk_mempool_put_bulk(reactor->event_mempool, events, count); + + return count; +} + +static void +_spdk_reactor_msg_passed(void *arg1, void *arg2) +{ + spdk_thread_fn fn = arg1; + + fn(arg2); +} + +static void +_spdk_reactor_send_msg(spdk_thread_fn fn, void *ctx, void *thread_ctx) +{ + struct spdk_event *event; + struct spdk_reactor *reactor; + + reactor = thread_ctx; + + event = spdk_event_allocate(reactor->lcore, _spdk_reactor_msg_passed, fn, ctx); + + spdk_event_call(event); +} + +static void +_spdk_poller_insert_timer(struct spdk_reactor *reactor, struct spdk_poller *poller, uint64_t now) +{ + struct spdk_poller *iter; + uint64_t next_run_tick; + + next_run_tick = now + poller->period_ticks; + poller->next_run_tick = next_run_tick; + + /* + * Insert poller in the reactor's timer_pollers list in sorted order by next scheduled + * run time. + */ + TAILQ_FOREACH_REVERSE(iter, &reactor->timer_pollers, timer_pollers_head, tailq) { + if (iter->next_run_tick <= next_run_tick) { + TAILQ_INSERT_AFTER(&reactor->timer_pollers, iter, poller, tailq); + return; + } + } + + /* No earlier pollers were found, so this poller must be the new head */ + TAILQ_INSERT_HEAD(&reactor->timer_pollers, poller, tailq); +} + +static struct spdk_poller * +_spdk_reactor_start_poller(void *thread_ctx, + spdk_poller_fn fn, + void *arg, + uint64_t period_microseconds) +{ + struct spdk_poller *poller; + struct spdk_reactor *reactor; + uint64_t quotient, remainder, ticks; + + reactor = thread_ctx; + + poller = calloc(1, sizeof(*poller)); + if (poller == NULL) { + SPDK_ERRLOG("Poller memory allocation failed\n"); + return NULL; + } + + poller->lcore = reactor->lcore; + poller->state = SPDK_POLLER_STATE_WAITING; + poller->fn = fn; + poller->arg = arg; + + if (period_microseconds) { + quotient = period_microseconds / SPDK_SEC_TO_USEC; + remainder = period_microseconds % SPDK_SEC_TO_USEC; + ticks = spdk_get_ticks_hz(); + + poller->period_ticks = ticks * quotient + (ticks * remainder) / SPDK_SEC_TO_USEC; + } else { + poller->period_ticks = 0; + } + + if (poller->period_ticks) { + _spdk_poller_insert_timer(reactor, poller, spdk_get_ticks()); + } else { + TAILQ_INSERT_TAIL(&reactor->active_pollers, poller, tailq); + } + + return poller; +} + +static void +_spdk_reactor_stop_poller(struct spdk_poller *poller, void *thread_ctx) +{ + struct spdk_reactor *reactor; + + reactor = thread_ctx; + + assert(poller->lcore == spdk_env_get_current_core()); + + if (poller->state == SPDK_POLLER_STATE_RUNNING) { + /* + * We are being called from the poller_fn, so set the state to unregistered + * and let the reactor loop free the poller. + */ + poller->state = SPDK_POLLER_STATE_UNREGISTERED; + } else { + /* Poller is not running currently, so just free it. */ + if (poller->period_ticks) { + TAILQ_REMOVE(&reactor->timer_pollers, poller, tailq); + } else { + TAILQ_REMOVE(&reactor->active_pollers, poller, tailq); + } + + free(poller); + } +} + +static int +get_rusage(void *arg) +{ + struct spdk_reactor *reactor = arg; + struct rusage rusage; + + if (getrusage(RUSAGE_THREAD, &rusage) != 0) { + return -1; + } + + if (rusage.ru_nvcsw != reactor->rusage.ru_nvcsw || rusage.ru_nivcsw != reactor->rusage.ru_nivcsw) { + SPDK_INFOLOG(SPDK_LOG_REACTOR, + "Reactor %d: %ld voluntary context switches and %ld involuntary context switches in the last second.\n", + reactor->lcore, rusage.ru_nvcsw - reactor->rusage.ru_nvcsw, + rusage.ru_nivcsw - reactor->rusage.ru_nivcsw); + } + reactor->rusage = rusage; + + return -1; +} + +static void +_spdk_reactor_context_switch_monitor_start(void *arg1, void *arg2) +{ + struct spdk_reactor *reactor = arg1; + + if (reactor->rusage_poller == NULL) { + getrusage(RUSAGE_THREAD, &reactor->rusage); + reactor->rusage_poller = spdk_poller_register(get_rusage, reactor, 1000000); + } +} + +static void +_spdk_reactor_context_switch_monitor_stop(void *arg1, void *arg2) +{ + struct spdk_reactor *reactor = arg1; + + if (reactor->rusage_poller != NULL) { + spdk_poller_unregister(&reactor->rusage_poller); + } +} + +static size_t +_spdk_reactor_get_max_event_cnt(uint8_t socket_count) +{ + size_t cnt; + + /* Try to make event ring fill at most 2MB of memory, + * as some ring implementations may require physical address + * contingency. We don't want to introduce a requirement of + * at least 2 physically contiguous 2MB hugepages. + */ + cnt = spdk_min(262144 / socket_count, 262144 / 2); + /* Take into account one extra element required by + * some ring implementations. + */ + cnt -= 1; + return cnt; +} + +void +spdk_reactor_enable_context_switch_monitor(bool enable) +{ + struct spdk_reactor *reactor; + spdk_event_fn fn; + uint32_t core; + + if (enable != g_context_switch_monitor_enabled) { + g_context_switch_monitor_enabled = enable; + if (enable) { + fn = _spdk_reactor_context_switch_monitor_start; + } else { + fn = _spdk_reactor_context_switch_monitor_stop; + } + SPDK_ENV_FOREACH_CORE(core) { + reactor = spdk_reactor_get(core); + spdk_event_call(spdk_event_allocate(core, fn, reactor, NULL)); + } + } +} + +bool +spdk_reactor_context_switch_monitor_enabled(void) +{ + return g_context_switch_monitor_enabled; +} + +static void +spdk_reactor_add_tsc_stats(void *arg, int rc, uint64_t now) +{ + struct spdk_reactor *reactor = arg; + struct spdk_reactor_tsc_stats *tsc_stats = &reactor->tsc_stats; + + if (rc == 0) { + /* Poller status idle */ + tsc_stats->idle_tsc += now - reactor->tsc_last; + } else if (rc > 0) { + /* Poller status busy */ + tsc_stats->busy_tsc += now - reactor->tsc_last; + } else { + /* Poller status unknown */ + tsc_stats->unknown_tsc += now - reactor->tsc_last; + } + + reactor->tsc_last = now; +} + +int +spdk_reactor_get_tsc_stats(struct spdk_reactor_tsc_stats *tsc_stats, uint32_t core) +{ + struct spdk_reactor *reactor; + + if (!spdk_cpuset_get_cpu(g_spdk_app_core_mask, core)) { + return -1; + } + + reactor = spdk_reactor_get(core); + *tsc_stats = reactor->tsc_stats; + + return 0; +} + +/** + * + * \brief This is the main function of the reactor thread. + * + * \code + * + * while (1) + * if (events to run) + * dequeue and run a batch of events + * + * if (active pollers) + * run the first poller in the list and move it to the back + * + * if (first timer poller has expired) + * run the first timer poller and reinsert it in the timer list + * + * if (no action taken and sleep enabled) + * sleep until next timer poller is scheduled to expire + * \endcode + * + */ +static int +_spdk_reactor_run(void *arg) +{ + struct spdk_reactor *reactor = arg; + struct spdk_poller *poller; + uint32_t event_count; + uint64_t now; + uint64_t sleep_cycles; + uint32_t sleep_us; + int rc = -1; + char thread_name[32]; + + snprintf(thread_name, sizeof(thread_name), "reactor_%u", reactor->lcore); + if (spdk_allocate_thread(_spdk_reactor_send_msg, + _spdk_reactor_start_poller, + _spdk_reactor_stop_poller, + reactor, thread_name) == NULL) { + return -1; + } + SPDK_NOTICELOG("Reactor started on core %u on socket %u\n", reactor->lcore, + reactor->socket_id); + + sleep_cycles = reactor->max_delay_us * spdk_get_ticks_hz() / SPDK_SEC_TO_USEC; + if (g_context_switch_monitor_enabled) { + _spdk_reactor_context_switch_monitor_start(reactor, NULL); + } + now = spdk_get_ticks(); + reactor->tsc_last = now; + + while (1) { + bool took_action = false; + + event_count = _spdk_event_queue_run_batch(reactor); + if (event_count > 0) { + rc = 1; + now = spdk_get_ticks(); + spdk_reactor_add_tsc_stats(reactor, rc, now); + took_action = true; + } + + poller = TAILQ_FIRST(&reactor->active_pollers); + if (poller) { + TAILQ_REMOVE(&reactor->active_pollers, poller, tailq); + poller->state = SPDK_POLLER_STATE_RUNNING; + rc = poller->fn(poller->arg); + now = spdk_get_ticks(); + spdk_reactor_add_tsc_stats(reactor, rc, now); + if (poller->state == SPDK_POLLER_STATE_UNREGISTERED) { + free(poller); + } else { + poller->state = SPDK_POLLER_STATE_WAITING; + TAILQ_INSERT_TAIL(&reactor->active_pollers, poller, tailq); + } + took_action = true; + } + + poller = TAILQ_FIRST(&reactor->timer_pollers); + if (poller) { + if (took_action == false) { + now = spdk_get_ticks(); + } + + if (now >= poller->next_run_tick) { + uint64_t tmp_timer_tsc; + + TAILQ_REMOVE(&reactor->timer_pollers, poller, tailq); + poller->state = SPDK_POLLER_STATE_RUNNING; + rc = poller->fn(poller->arg); + /* Save the tsc value from before poller->fn was executed. We want to + * use the current time for idle/busy tsc value accounting, but want to + * use the older time to reinsert to the timer poller below. */ + tmp_timer_tsc = now; + now = spdk_get_ticks(); + spdk_reactor_add_tsc_stats(reactor, rc, now); + if (poller->state == SPDK_POLLER_STATE_UNREGISTERED) { + free(poller); + } else { + poller->state = SPDK_POLLER_STATE_WAITING; + _spdk_poller_insert_timer(reactor, poller, tmp_timer_tsc); + } + took_action = true; + } + } + + /* Determine if the thread can sleep */ + if (sleep_cycles && !took_action) { + now = spdk_get_ticks(); + sleep_us = reactor->max_delay_us; + + poller = TAILQ_FIRST(&reactor->timer_pollers); + if (poller) { + /* There are timers registered, so don't sleep beyond + * when the next timer should fire */ + if (poller->next_run_tick < (now + sleep_cycles)) { + if (poller->next_run_tick <= now) { + sleep_us = 0; + } else { + sleep_us = ((poller->next_run_tick - now) * + SPDK_SEC_TO_USEC) / spdk_get_ticks_hz(); + } + } + } + + if (sleep_us > 0) { + usleep(sleep_us); + } + } + + if (g_reactor_state != SPDK_REACTOR_STATE_RUNNING) { + break; + } + } + + _spdk_reactor_context_switch_monitor_stop(reactor, NULL); + spdk_free_thread(); + return 0; +} + +static void +spdk_reactor_construct(struct spdk_reactor *reactor, uint32_t lcore, uint64_t max_delay_us) +{ + reactor->lcore = lcore; + reactor->socket_id = spdk_env_get_socket_id(lcore); + assert(reactor->socket_id < SPDK_MAX_SOCKET); + reactor->max_delay_us = max_delay_us; + + TAILQ_INIT(&reactor->active_pollers); + TAILQ_INIT(&reactor->timer_pollers); + + reactor->events = spdk_ring_create(SPDK_RING_TYPE_MP_SC, 65536, reactor->socket_id); + if (!reactor->events) { + SPDK_NOTICELOG("Ring creation failed on preferred socket %d. Try other sockets.\n", + reactor->socket_id); + + reactor->events = spdk_ring_create(SPDK_RING_TYPE_MP_SC, 65536, + SPDK_ENV_SOCKET_ID_ANY); + } + assert(reactor->events != NULL); + + reactor->event_mempool = g_spdk_event_mempool[reactor->socket_id]; +} + +int +spdk_app_parse_core_mask(const char *mask, struct spdk_cpuset *cpumask) +{ + int ret; + struct spdk_cpuset *validmask; + + ret = spdk_cpuset_parse(cpumask, mask); + if (ret < 0) { + return ret; + } + + validmask = spdk_app_get_core_mask(); + spdk_cpuset_and(cpumask, validmask); + + return 0; +} + +struct spdk_cpuset * +spdk_app_get_core_mask(void) +{ + return g_spdk_app_core_mask; +} + + +static uint64_t +spdk_reactor_get_socket_mask(void) +{ + uint32_t i; + uint32_t socket_id; + uint64_t socket_info = 0; + + SPDK_ENV_FOREACH_CORE(i) { + socket_id = spdk_env_get_socket_id(i); + socket_info |= (1ULL << socket_id); + } + + return socket_info; +} + +void +spdk_reactors_start(void) +{ + struct spdk_reactor *reactor; + uint32_t i, current_core; + int rc; + + g_reactor_state = SPDK_REACTOR_STATE_RUNNING; + g_spdk_app_core_mask = spdk_cpuset_alloc(); + + current_core = spdk_env_get_current_core(); + SPDK_ENV_FOREACH_CORE(i) { + if (i != current_core) { + reactor = spdk_reactor_get(i); + rc = spdk_env_thread_launch_pinned(reactor->lcore, _spdk_reactor_run, reactor); + if (rc < 0) { + SPDK_ERRLOG("Unable to start reactor thread on core %u\n", reactor->lcore); + assert(false); + return; + } + } + spdk_cpuset_set_cpu(g_spdk_app_core_mask, i, true); + } + + /* Start the master reactor */ + reactor = spdk_reactor_get(current_core); + _spdk_reactor_run(reactor); + + spdk_env_thread_wait_all(); + + g_reactor_state = SPDK_REACTOR_STATE_SHUTDOWN; + spdk_cpuset_free(g_spdk_app_core_mask); + g_spdk_app_core_mask = NULL; +} + +void +spdk_reactors_stop(void *arg1, void *arg2) +{ + g_reactor_state = SPDK_REACTOR_STATE_EXITING; +} + +int +spdk_reactors_init(unsigned int max_delay_us) +{ + int rc; + uint32_t i, j, last_core; + struct spdk_reactor *reactor; + uint64_t socket_mask = 0x0; + uint8_t socket_count = 0; + char mempool_name[32]; + + socket_mask = spdk_reactor_get_socket_mask(); + SPDK_NOTICELOG("Occupied cpu socket mask is 0x%lx\n", socket_mask); + + for (i = 0; i < SPDK_MAX_SOCKET; i++) { + if ((1ULL << i) & socket_mask) { + socket_count++; + } + } + if (socket_count == 0) { + SPDK_ERRLOG("No sockets occupied (internal error)\n"); + return -1; + } + + for (i = 0; i < SPDK_MAX_SOCKET; i++) { + if ((1ULL << i) & socket_mask) { + snprintf(mempool_name, sizeof(mempool_name), "evtpool%d_%d", i, getpid()); + g_spdk_event_mempool[i] = spdk_mempool_create(mempool_name, + _spdk_reactor_get_max_event_cnt(socket_count), + sizeof(struct spdk_event), + SPDK_MEMPOOL_DEFAULT_CACHE_SIZE, i); + + if (g_spdk_event_mempool[i] == NULL) { + SPDK_NOTICELOG("Event_mempool creation failed on preferred socket %d.\n", i); + + /* + * Instead of failing the operation directly, try to create + * the mempool on any available sockets in the case that + * memory is not evenly installed on all sockets. If still + * fails, free all allocated memory and exits. + */ + g_spdk_event_mempool[i] = spdk_mempool_create( + mempool_name, + _spdk_reactor_get_max_event_cnt(socket_count), + sizeof(struct spdk_event), + SPDK_MEMPOOL_DEFAULT_CACHE_SIZE, + SPDK_ENV_SOCKET_ID_ANY); + + if (g_spdk_event_mempool[i] == NULL) { + for (j = i - 1; j < i; j--) { + if (g_spdk_event_mempool[j] != NULL) { + spdk_mempool_free(g_spdk_event_mempool[j]); + } + } + SPDK_ERRLOG("spdk_event_mempool creation failed\n"); + return -1; + } + } + } else { + g_spdk_event_mempool[i] = NULL; + } + } + + /* struct spdk_reactor must be aligned on 64 byte boundary */ + last_core = spdk_env_get_last_core(); + rc = posix_memalign((void **)&g_reactors, 64, + (last_core + 1) * sizeof(struct spdk_reactor)); + if (rc != 0) { + SPDK_ERRLOG("Could not allocate array size=%u for g_reactors\n", + last_core + 1); + for (i = 0; i < SPDK_MAX_SOCKET; i++) { + if (g_spdk_event_mempool[i] != NULL) { + spdk_mempool_free(g_spdk_event_mempool[i]); + } + } + return -1; + } + + memset(g_reactors, 0, (last_core + 1) * sizeof(struct spdk_reactor)); + + SPDK_ENV_FOREACH_CORE(i) { + reactor = spdk_reactor_get(i); + spdk_reactor_construct(reactor, i, max_delay_us); + } + + g_reactor_state = SPDK_REACTOR_STATE_INITIALIZED; + + return 0; +} + +void +spdk_reactors_fini(void) +{ + uint32_t i; + struct spdk_reactor *reactor; + + SPDK_ENV_FOREACH_CORE(i) { + reactor = spdk_reactor_get(i); + if (spdk_likely(reactor != NULL) && reactor->events != NULL) { + spdk_ring_free(reactor->events); + } + } + + for (i = 0; i < SPDK_MAX_SOCKET; i++) { + if (g_spdk_event_mempool[i] != NULL) { + spdk_mempool_free(g_spdk_event_mempool[i]); + } + } + + free(g_reactors); + g_reactors = NULL; +} + +SPDK_LOG_REGISTER_COMPONENT("reactor", SPDK_LOG_REACTOR) |