diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-09 13:16:35 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-09 13:16:35 +0000 |
commit | e2bbf175a2184bd76f6c54ccf8456babeb1a46fc (patch) | |
tree | f0b76550d6e6f500ada964a3a4ee933a45e5a6f1 /lib/memory.c | |
parent | Initial commit. (diff) | |
download | frr-e2bbf175a2184bd76f6c54ccf8456babeb1a46fc.tar.xz frr-e2bbf175a2184bd76f6c54ccf8456babeb1a46fc.zip |
Adding upstream version 9.1.upstream/9.1
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'lib/memory.c')
-rw-r--r-- | lib/memory.c | 181 |
1 files changed, 181 insertions, 0 deletions
diff --git a/lib/memory.c b/lib/memory.c new file mode 100644 index 0000000..8fbe5c4 --- /dev/null +++ b/lib/memory.c @@ -0,0 +1,181 @@ +// SPDX-License-Identifier: ISC +/* + * Copyright (c) 2015-16 David Lamparter, for NetDEF, Inc. + */ + +#include <zebra.h> + +#include <stdlib.h> +#ifdef HAVE_MALLOC_H +#include <malloc.h> +#endif +#ifdef HAVE_MALLOC_NP_H +#include <malloc_np.h> +#endif +#ifdef HAVE_MALLOC_MALLOC_H +#include <malloc/malloc.h> +#endif + +#include "memory.h" +#include "log.h" +#include "libfrr_trace.h" + +static struct memgroup *mg_first = NULL; +struct memgroup **mg_insert = &mg_first; + +DEFINE_MGROUP(LIB, "libfrr"); +DEFINE_MTYPE(LIB, TMP, "Temporary memory"); +DEFINE_MTYPE(LIB, BITFIELD, "Bitfield memory"); + +static inline void mt_count_alloc(struct memtype *mt, size_t size, void *ptr) +{ + size_t current; + size_t oldsize; + + current = 1 + atomic_fetch_add_explicit(&mt->n_alloc, 1, + memory_order_relaxed); + + oldsize = atomic_load_explicit(&mt->n_max, memory_order_relaxed); + if (current > oldsize) + /* note that this may fail, but approximation is sufficient */ + atomic_compare_exchange_weak_explicit(&mt->n_max, &oldsize, + current, + memory_order_relaxed, + memory_order_relaxed); + + oldsize = atomic_load_explicit(&mt->size, memory_order_relaxed); + if (oldsize == 0) + oldsize = atomic_exchange_explicit(&mt->size, size, + memory_order_relaxed); + if (oldsize != 0 && oldsize != size && oldsize != SIZE_VAR) + atomic_store_explicit(&mt->size, SIZE_VAR, + memory_order_relaxed); + +#ifdef HAVE_MALLOC_USABLE_SIZE + size_t mallocsz = malloc_usable_size(ptr); + + current = mallocsz + atomic_fetch_add_explicit(&mt->total, mallocsz, + memory_order_relaxed); + oldsize = atomic_load_explicit(&mt->max_size, memory_order_relaxed); + if (current > oldsize) + /* note that this may fail, but approximation is sufficient */ + atomic_compare_exchange_weak_explicit(&mt->max_size, &oldsize, + current, + memory_order_relaxed, + memory_order_relaxed); +#endif +} + +static inline void mt_count_free(struct memtype *mt, void *ptr) +{ + frrtrace(2, frr_libfrr, memfree, mt, ptr); + + assert(mt->n_alloc); + atomic_fetch_sub_explicit(&mt->n_alloc, 1, memory_order_relaxed); + +#ifdef HAVE_MALLOC_USABLE_SIZE + size_t mallocsz = malloc_usable_size(ptr); + + atomic_fetch_sub_explicit(&mt->total, mallocsz, memory_order_relaxed); +#endif +} + +static inline void *mt_checkalloc(struct memtype *mt, void *ptr, size_t size) +{ + frrtrace(3, frr_libfrr, memalloc, mt, ptr, size); + + if (__builtin_expect(ptr == NULL, 0)) { + if (size) { + /* malloc(0) is allowed to return NULL */ + memory_oom(size, mt->name); + } + return NULL; + } + mt_count_alloc(mt, size, ptr); + return ptr; +} + +void *qmalloc(struct memtype *mt, size_t size) +{ + return mt_checkalloc(mt, malloc(size), size); +} + +void *qcalloc(struct memtype *mt, size_t size) +{ + return mt_checkalloc(mt, calloc(size, 1), size); +} + +void *qrealloc(struct memtype *mt, void *ptr, size_t size) +{ + if (ptr) + mt_count_free(mt, ptr); + return mt_checkalloc(mt, ptr ? realloc(ptr, size) : malloc(size), size); +} + +void *qstrdup(struct memtype *mt, const char *str) +{ + return str ? mt_checkalloc(mt, strdup(str), strlen(str) + 1) : NULL; +} + +void qcountfree(struct memtype *mt, void *ptr) +{ + if (ptr) + mt_count_free(mt, ptr); +} + +void qfree(struct memtype *mt, void *ptr) +{ + if (ptr) + mt_count_free(mt, ptr); + free(ptr); +} + +int qmem_walk(qmem_walk_fn *func, void *arg) +{ + struct memgroup *mg; + struct memtype *mt; + int rv; + + for (mg = mg_first; mg; mg = mg->next) { + if ((rv = func(arg, mg, NULL))) + return rv; + for (mt = mg->types; mt; mt = mt->next) + if ((rv = func(arg, mg, mt))) + return rv; + } + return 0; +} + +struct exit_dump_args { + FILE *fp; + const char *prefix; + int error; +}; + +static int qmem_exit_walker(void *arg, struct memgroup *mg, struct memtype *mt) +{ + struct exit_dump_args *eda = arg; + + if (!mt) { + fprintf(eda->fp, + "%s: showing active allocations in memory group %s\n", + eda->prefix, mg->name); + + } else if (mt->n_alloc) { + char size[32]; + if (!mg->active_at_exit) + eda->error++; + snprintf(size, sizeof(size), "%10zu", mt->size); + fprintf(eda->fp, "%s: memstats: %-30s: %6zu * %s\n", + eda->prefix, mt->name, mt->n_alloc, + mt->size == SIZE_VAR ? "(variably sized)" : size); + } + return 0; +} + +int log_memstats(FILE *fp, const char *prefix) +{ + struct exit_dump_args eda = {.fp = fp, .prefix = prefix, .error = 0}; + qmem_walk(qmem_exit_walker, &eda); + return eda.error; +} |