diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-21 11:54:28 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-21 11:54:28 +0000 |
commit | e6918187568dbd01842d8d1d2c808ce16a894239 (patch) | |
tree | 64f88b554b444a49f656b6c656111a145cbbaa28 /src/spdk/dpdk/lib/librte_rib | |
parent | Initial commit. (diff) | |
download | ceph-e6918187568dbd01842d8d1d2c808ce16a894239.tar.xz ceph-e6918187568dbd01842d8d1d2c808ce16a894239.zip |
Adding upstream version 18.2.2.upstream/18.2.2
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/spdk/dpdk/lib/librte_rib')
-rw-r--r-- | src/spdk/dpdk/lib/librte_rib/Makefile | 22 | ||||
-rw-r--r-- | src/spdk/dpdk/lib/librte_rib/meson.build | 9 | ||||
-rw-r--r-- | src/spdk/dpdk/lib/librte_rib/rte_rib.c | 532 | ||||
-rw-r--r-- | src/spdk/dpdk/lib/librte_rib/rte_rib.h | 277 | ||||
-rw-r--r-- | src/spdk/dpdk/lib/librte_rib/rte_rib6.c | 598 | ||||
-rw-r--r-- | src/spdk/dpdk/lib/librte_rib/rte_rib6.h | 334 | ||||
-rw-r--r-- | src/spdk/dpdk/lib/librte_rib/rte_rib_version.map | 35 |
7 files changed, 1807 insertions, 0 deletions
diff --git a/src/spdk/dpdk/lib/librte_rib/Makefile b/src/spdk/dpdk/lib/librte_rib/Makefile new file mode 100644 index 000000000..be80ce5d3 --- /dev/null +++ b/src/spdk/dpdk/lib/librte_rib/Makefile @@ -0,0 +1,22 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2018 Vladimir Medvedkin <medvedkinv@gmail.com> +# Copyright(c) 2019 Intel Corporation + +include $(RTE_SDK)/mk/rte.vars.mk + +# library name +LIB = librte_rib.a + +CFLAGS += -O3 +CFLAGS += $(WERROR_FLAGS) -I$(SRCDIR) +LDLIBS += -lrte_eal -lrte_mempool + +EXPORT_MAP := rte_rib_version.map + +# all source are stored in SRCS-y +SRCS-$(CONFIG_RTE_LIBRTE_RIB) := rte_rib.c rte_rib6.c + +# install this header file +SYMLINK-$(CONFIG_RTE_LIBRTE_RIB)-include := rte_rib.h rte_rib6.h + +include $(RTE_SDK)/mk/rte.lib.mk diff --git a/src/spdk/dpdk/lib/librte_rib/meson.build b/src/spdk/dpdk/lib/librte_rib/meson.build new file mode 100644 index 000000000..0fcdba0da --- /dev/null +++ b/src/spdk/dpdk/lib/librte_rib/meson.build @@ -0,0 +1,9 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2018 Vladimir Medvedkin <medvedkinv@gmail.com> +# Copyright(c) 2019 Intel Corporation + +sources = files('rte_rib.c', 'rte_rib6.c') +headers = files('rte_rib.h', 'rte_rib6.h') +deps += ['mempool'] +build = false +reason = 'not needed by SPDK' diff --git a/src/spdk/dpdk/lib/librte_rib/rte_rib.c b/src/spdk/dpdk/lib/librte_rib/rte_rib.c new file mode 100644 index 000000000..55d612dc2 --- /dev/null +++ b/src/spdk/dpdk/lib/librte_rib/rte_rib.c @@ -0,0 +1,532 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Vladimir Medvedkin <medvedkinv@gmail.com> + * Copyright(c) 2019 Intel Corporation + */ + +#include <stdbool.h> +#include <stdint.h> + +#include <rte_eal.h> +#include <rte_eal_memconfig.h> +#include <rte_errno.h> +#include <rte_malloc.h> +#include <rte_mempool.h> +#include <rte_rwlock.h> +#include <rte_string_fns.h> +#include <rte_tailq.h> + +#include <rte_rib.h> + +TAILQ_HEAD(rte_rib_list, rte_tailq_entry); +static struct rte_tailq_elem rte_rib_tailq = { + .name = "RTE_RIB", +}; +EAL_REGISTER_TAILQ(rte_rib_tailq) + +#define RTE_RIB_VALID_NODE 1 +/* Maximum depth value possible for IPv4 RIB. */ +#define RIB_MAXDEPTH 32 +/* Maximum length of a RIB name. */ +#define RTE_RIB_NAMESIZE 64 + +struct rte_rib_node { + struct rte_rib_node *left; + struct rte_rib_node *right; + struct rte_rib_node *parent; + uint32_t ip; + uint8_t depth; + uint8_t flag; + uint64_t nh; + __extension__ uint64_t ext[0]; +}; + +struct rte_rib { + char name[RTE_RIB_NAMESIZE]; + struct rte_rib_node *tree; + struct rte_mempool *node_pool; + uint32_t cur_nodes; + uint32_t cur_routes; + uint32_t max_nodes; +}; + +static inline bool +is_valid_node(struct rte_rib_node *node) +{ + return (node->flag & RTE_RIB_VALID_NODE) == RTE_RIB_VALID_NODE; +} + +static inline bool +is_right_node(struct rte_rib_node *node) +{ + return node->parent->right == node; +} + +/* + * Check if ip1 is covered by ip2/depth prefix + */ +static inline bool +is_covered(uint32_t ip1, uint32_t ip2, uint8_t depth) +{ + return ((ip1 ^ ip2) & rte_rib_depth_to_mask(depth)) == 0; +} + +static inline struct rte_rib_node * +get_nxt_node(struct rte_rib_node *node, uint32_t ip) +{ + return (ip & (1 << (31 - node->depth))) ? node->right : node->left; +} + +static struct rte_rib_node * +node_alloc(struct rte_rib *rib) +{ + struct rte_rib_node *ent; + int ret; + + ret = rte_mempool_get(rib->node_pool, (void *)&ent); + if (unlikely(ret != 0)) + return NULL; + ++rib->cur_nodes; + return ent; +} + +static void +node_free(struct rte_rib *rib, struct rte_rib_node *ent) +{ + --rib->cur_nodes; + rte_mempool_put(rib->node_pool, ent); +} + +struct rte_rib_node * +rte_rib_lookup(struct rte_rib *rib, uint32_t ip) +{ + struct rte_rib_node *cur, *prev = NULL; + + if (rib == NULL) { + rte_errno = EINVAL; + return NULL; + } + + cur = rib->tree; + while ((cur != NULL) && is_covered(ip, cur->ip, cur->depth)) { + if (is_valid_node(cur)) + prev = cur; + cur = get_nxt_node(cur, ip); + } + return prev; +} + +struct rte_rib_node * +rte_rib_lookup_parent(struct rte_rib_node *ent) +{ + struct rte_rib_node *tmp; + + if (ent == NULL) + return NULL; + tmp = ent->parent; + while ((tmp != NULL) && !is_valid_node(tmp)) + tmp = tmp->parent; + return tmp; +} + +static struct rte_rib_node * +__rib_lookup_exact(struct rte_rib *rib, uint32_t ip, uint8_t depth) +{ + struct rte_rib_node *cur; + + cur = rib->tree; + while (cur != NULL) { + if ((cur->ip == ip) && (cur->depth == depth) && + is_valid_node(cur)) + return cur; + if ((cur->depth > depth) || + !is_covered(ip, cur->ip, cur->depth)) + break; + cur = get_nxt_node(cur, ip); + } + return NULL; +} + +struct rte_rib_node * +rte_rib_lookup_exact(struct rte_rib *rib, uint32_t ip, uint8_t depth) +{ + if ((rib == NULL) || (depth > RIB_MAXDEPTH)) { + rte_errno = EINVAL; + return NULL; + } + ip &= rte_rib_depth_to_mask(depth); + + return __rib_lookup_exact(rib, ip, depth); +} + +/* + * Traverses on subtree and retrieves more specific routes + * for a given in args ip/depth prefix + * last = NULL means the first invocation + */ +struct rte_rib_node * +rte_rib_get_nxt(struct rte_rib *rib, uint32_t ip, + uint8_t depth, struct rte_rib_node *last, int flag) +{ + struct rte_rib_node *tmp, *prev = NULL; + + if ((rib == NULL) || (depth > RIB_MAXDEPTH)) { + rte_errno = EINVAL; + return NULL; + } + + if (last == NULL) { + tmp = rib->tree; + while ((tmp) && (tmp->depth < depth)) + tmp = get_nxt_node(tmp, ip); + } else { + tmp = last; + while ((tmp->parent != NULL) && (is_right_node(tmp) || + (tmp->parent->right == NULL))) { + tmp = tmp->parent; + if (is_valid_node(tmp) && + (is_covered(tmp->ip, ip, depth) && + (tmp->depth > depth))) + return tmp; + } + tmp = (tmp->parent) ? tmp->parent->right : NULL; + } + while (tmp) { + if (is_valid_node(tmp) && + (is_covered(tmp->ip, ip, depth) && + (tmp->depth > depth))) { + prev = tmp; + if (flag == RTE_RIB_GET_NXT_COVER) + return prev; + } + tmp = (tmp->left) ? tmp->left : tmp->right; + } + return prev; +} + +void +rte_rib_remove(struct rte_rib *rib, uint32_t ip, uint8_t depth) +{ + struct rte_rib_node *cur, *prev, *child; + + cur = rte_rib_lookup_exact(rib, ip, depth); + if (cur == NULL) + return; + + --rib->cur_routes; + cur->flag &= ~RTE_RIB_VALID_NODE; + while (!is_valid_node(cur)) { + if ((cur->left != NULL) && (cur->right != NULL)) + return; + child = (cur->left == NULL) ? cur->right : cur->left; + if (child != NULL) + child->parent = cur->parent; + if (cur->parent == NULL) { + rib->tree = child; + node_free(rib, cur); + return; + } + if (cur->parent->left == cur) + cur->parent->left = child; + else + cur->parent->right = child; + prev = cur; + cur = cur->parent; + node_free(rib, prev); + } +} + +struct rte_rib_node * +rte_rib_insert(struct rte_rib *rib, uint32_t ip, uint8_t depth) +{ + struct rte_rib_node **tmp; + struct rte_rib_node *prev = NULL; + struct rte_rib_node *new_node = NULL; + struct rte_rib_node *common_node = NULL; + int d = 0; + uint32_t common_prefix; + uint8_t common_depth; + + if ((rib == NULL) || (depth > RIB_MAXDEPTH)) { + rte_errno = EINVAL; + return NULL; + } + + tmp = &rib->tree; + ip &= rte_rib_depth_to_mask(depth); + new_node = __rib_lookup_exact(rib, ip, depth); + if (new_node != NULL) { + rte_errno = EEXIST; + return NULL; + } + + new_node = node_alloc(rib); + if (new_node == NULL) { + rte_errno = ENOMEM; + return NULL; + } + new_node->left = NULL; + new_node->right = NULL; + new_node->parent = NULL; + new_node->ip = ip; + new_node->depth = depth; + new_node->flag = RTE_RIB_VALID_NODE; + + /* traverse down the tree to find matching node or closest matching */ + while (1) { + /* insert as the last node in the branch */ + if (*tmp == NULL) { + *tmp = new_node; + new_node->parent = prev; + ++rib->cur_routes; + return *tmp; + } + /* + * Intermediate node found. + * Previous rte_rib_lookup_exact() returned NULL + * but node with proper search criteria is found. + * Validate intermediate node and return. + */ + if ((ip == (*tmp)->ip) && (depth == (*tmp)->depth)) { + node_free(rib, new_node); + (*tmp)->flag |= RTE_RIB_VALID_NODE; + ++rib->cur_routes; + return *tmp; + } + d = (*tmp)->depth; + if ((d >= depth) || !is_covered(ip, (*tmp)->ip, d)) + break; + prev = *tmp; + tmp = (ip & (1 << (31 - d))) ? &(*tmp)->right : &(*tmp)->left; + } + /* closest node found, new_node should be inserted in the middle */ + common_depth = RTE_MIN(depth, (*tmp)->depth); + common_prefix = ip ^ (*tmp)->ip; + d = __builtin_clz(common_prefix); + + common_depth = RTE_MIN(d, common_depth); + common_prefix = ip & rte_rib_depth_to_mask(common_depth); + if ((common_prefix == ip) && (common_depth == depth)) { + /* insert as a parent */ + if ((*tmp)->ip & (1 << (31 - depth))) + new_node->right = *tmp; + else + new_node->left = *tmp; + new_node->parent = (*tmp)->parent; + (*tmp)->parent = new_node; + *tmp = new_node; + } else { + /* create intermediate node */ + common_node = node_alloc(rib); + if (common_node == NULL) { + node_free(rib, new_node); + rte_errno = ENOMEM; + return NULL; + } + common_node->ip = common_prefix; + common_node->depth = common_depth; + common_node->flag = 0; + common_node->parent = (*tmp)->parent; + new_node->parent = common_node; + (*tmp)->parent = common_node; + if ((new_node->ip & (1 << (31 - common_depth))) == 0) { + common_node->left = new_node; + common_node->right = *tmp; + } else { + common_node->left = *tmp; + common_node->right = new_node; + } + *tmp = common_node; + } + ++rib->cur_routes; + return new_node; +} + +int +rte_rib_get_ip(struct rte_rib_node *node, uint32_t *ip) +{ + if ((node == NULL) || (ip == NULL)) { + rte_errno = EINVAL; + return -1; + } + *ip = node->ip; + return 0; +} + +int +rte_rib_get_depth(struct rte_rib_node *node, uint8_t *depth) +{ + if ((node == NULL) || (depth == NULL)) { + rte_errno = EINVAL; + return -1; + } + *depth = node->depth; + return 0; +} + +void * +rte_rib_get_ext(struct rte_rib_node *node) +{ + return (node == NULL) ? NULL : &node->ext[0]; +} + +int +rte_rib_get_nh(struct rte_rib_node *node, uint64_t *nh) +{ + if ((node == NULL) || (nh == NULL)) { + rte_errno = EINVAL; + return -1; + } + *nh = node->nh; + return 0; +} + +int +rte_rib_set_nh(struct rte_rib_node *node, uint64_t nh) +{ + if (node == NULL) { + rte_errno = EINVAL; + return -1; + } + node->nh = nh; + return 0; +} + +struct rte_rib * +rte_rib_create(const char *name, int socket_id, struct rte_rib_conf *conf) +{ + char mem_name[RTE_RIB_NAMESIZE]; + struct rte_rib *rib = NULL; + struct rte_tailq_entry *te; + struct rte_rib_list *rib_list; + struct rte_mempool *node_pool; + + /* Check user arguments. */ + if ((name == NULL) || (conf == NULL) || + (conf->max_nodes == 0)) { + rte_errno = EINVAL; + return NULL; + } + + snprintf(mem_name, sizeof(mem_name), "MP_%s", name); + node_pool = rte_mempool_create(mem_name, conf->max_nodes, + sizeof(struct rte_rib_node) + conf->ext_sz, 0, 0, + NULL, NULL, NULL, NULL, socket_id, 0); + + if (node_pool == NULL) { + RTE_LOG(ERR, LPM, + "Can not allocate mempool for RIB %s\n", name); + return NULL; + } + + snprintf(mem_name, sizeof(mem_name), "RIB_%s", name); + rib_list = RTE_TAILQ_CAST(rte_rib_tailq.head, rte_rib_list); + + rte_mcfg_tailq_write_lock(); + + /* guarantee there's no existing */ + TAILQ_FOREACH(te, rib_list, next) { + rib = (struct rte_rib *)te->data; + if (strncmp(name, rib->name, RTE_RIB_NAMESIZE) == 0) + break; + } + rib = NULL; + if (te != NULL) { + rte_errno = EEXIST; + goto exit; + } + + /* allocate tailq entry */ + te = rte_zmalloc("RIB_TAILQ_ENTRY", sizeof(*te), 0); + if (te == NULL) { + RTE_LOG(ERR, LPM, + "Can not allocate tailq entry for RIB %s\n", name); + rte_errno = ENOMEM; + goto exit; + } + + /* Allocate memory to store the RIB data structures. */ + rib = rte_zmalloc_socket(mem_name, + sizeof(struct rte_rib), RTE_CACHE_LINE_SIZE, socket_id); + if (rib == NULL) { + RTE_LOG(ERR, LPM, "RIB %s memory allocation failed\n", name); + rte_errno = ENOMEM; + goto free_te; + } + + rte_strlcpy(rib->name, name, sizeof(rib->name)); + rib->tree = NULL; + rib->max_nodes = conf->max_nodes; + rib->node_pool = node_pool; + te->data = (void *)rib; + TAILQ_INSERT_TAIL(rib_list, te, next); + + rte_mcfg_tailq_write_unlock(); + + return rib; + +free_te: + rte_free(te); +exit: + rte_mcfg_tailq_write_unlock(); + rte_mempool_free(node_pool); + + return NULL; +} + +struct rte_rib * +rte_rib_find_existing(const char *name) +{ + struct rte_rib *rib = NULL; + struct rte_tailq_entry *te; + struct rte_rib_list *rib_list; + + rib_list = RTE_TAILQ_CAST(rte_rib_tailq.head, rte_rib_list); + + rte_mcfg_tailq_read_lock(); + TAILQ_FOREACH(te, rib_list, next) { + rib = (struct rte_rib *) te->data; + if (strncmp(name, rib->name, RTE_RIB_NAMESIZE) == 0) + break; + } + rte_mcfg_tailq_read_unlock(); + + if (te == NULL) { + rte_errno = ENOENT; + return NULL; + } + + return rib; +} + +void +rte_rib_free(struct rte_rib *rib) +{ + struct rte_tailq_entry *te; + struct rte_rib_list *rib_list; + struct rte_rib_node *tmp = NULL; + + if (rib == NULL) + return; + + rib_list = RTE_TAILQ_CAST(rte_rib_tailq.head, rte_rib_list); + + rte_mcfg_tailq_write_lock(); + + /* find our tailq entry */ + TAILQ_FOREACH(te, rib_list, next) { + if (te->data == (void *)rib) + break; + } + if (te != NULL) + TAILQ_REMOVE(rib_list, te, next); + + rte_mcfg_tailq_write_unlock(); + + while ((tmp = rte_rib_get_nxt(rib, 0, 0, tmp, + RTE_RIB_GET_NXT_ALL)) != NULL) + rte_rib_remove(rib, tmp->ip, tmp->depth); + + rte_mempool_free(rib->node_pool); + rte_free(rib); + rte_free(te); +} diff --git a/src/spdk/dpdk/lib/librte_rib/rte_rib.h b/src/spdk/dpdk/lib/librte_rib/rte_rib.h new file mode 100644 index 000000000..6b70de980 --- /dev/null +++ b/src/spdk/dpdk/lib/librte_rib/rte_rib.h @@ -0,0 +1,277 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Vladimir Medvedkin <medvedkinv@gmail.com> + * Copyright(c) 2019 Intel Corporation + */ + +#ifndef _RTE_RIB_H_ +#define _RTE_RIB_H_ + +/** + * @file + * Level compressed tree implementation for IPv4 Longest Prefix Match + */ + +#include <rte_compat.h> + +/** + * rte_rib_get_nxt() flags + */ +enum { + /** flag to get all subroutes in a RIB tree */ + RTE_RIB_GET_NXT_ALL, + /** flag to get first matched subroutes in a RIB tree */ + RTE_RIB_GET_NXT_COVER +}; + +struct rte_rib; +struct rte_rib_node; + +/** RIB configuration structure */ +struct rte_rib_conf { + /** + * Size of extension block inside rte_rib_node. + * This space could be used to store additional user + * defined data. + */ + size_t ext_sz; + /* size of rte_rib_node's pool */ + int max_nodes; +}; + +/** + * Get an IPv4 mask from prefix length + * It is caller responsibility to make sure depth is not bigger than 32 + * + * @param depth + * prefix length + * @return + * IPv4 mask + */ +static inline uint32_t +rte_rib_depth_to_mask(uint8_t depth) +{ + return (uint32_t)(UINT64_MAX << (32 - depth)); +} + +/** + * Lookup an IP into the RIB structure + * + * @param rib + * RIB object handle + * @param ip + * IP to be looked up in the RIB + * @return + * pointer to struct rte_rib_node on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib_node * +rte_rib_lookup(struct rte_rib *rib, uint32_t ip); + +/** + * Lookup less specific route into the RIB structure + * + * @param ent + * Pointer to struct rte_rib_node that represents target route + * @return + * pointer to struct rte_rib_node that represents + * less specific route on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib_node * +rte_rib_lookup_parent(struct rte_rib_node *ent); + +/** + * Lookup prefix into the RIB structure + * + * @param rib + * RIB object handle + * @param ip + * net to be looked up in the RIB + * @param depth + * prefix length + * @return + * pointer to struct rte_rib_node on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib_node * +rte_rib_lookup_exact(struct rte_rib *rib, uint32_t ip, uint8_t depth); + +/** + * Retrieve next more specific prefix from the RIB + * that is covered by ip/depth supernet in an ascending order + * + * @param rib + * RIB object handle + * @param ip + * net address of supernet prefix that covers returned more specific prefixes + * @param depth + * supernet prefix length + * @param last + * pointer to the last returned prefix to get next prefix + * or + * NULL to get first more specific prefix + * @param flag + * -RTE_RIB_GET_NXT_ALL + * get all prefixes from subtrie + * -RTE_RIB_GET_NXT_COVER + * get only first more specific prefix even if it have more specifics + * @return + * pointer to the next more specific prefix + * NULL if there is no prefixes left + */ +__rte_experimental +struct rte_rib_node * +rte_rib_get_nxt(struct rte_rib *rib, uint32_t ip, uint8_t depth, + struct rte_rib_node *last, int flag); + +/** + * Remove prefix from the RIB + * + * @param rib + * RIB object handle + * @param ip + * net to be removed from the RIB + * @param depth + * prefix length + */ +__rte_experimental +void +rte_rib_remove(struct rte_rib *rib, uint32_t ip, uint8_t depth); + +/** + * Insert prefix into the RIB + * + * @param rib + * RIB object handle + * @param ip + * net to be inserted to the RIB + * @param depth + * prefix length + * @return + * pointer to new rte_rib_node on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib_node * +rte_rib_insert(struct rte_rib *rib, uint32_t ip, uint8_t depth); + +/** + * Get an ip from rte_rib_node + * + * @param node + * pointer to the rib node + * @param ip + * pointer to the ip to save + * @return + * 0 on success. + * -1 on failure with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib_get_ip(struct rte_rib_node *node, uint32_t *ip); + +/** + * Get a depth from rte_rib_node + * + * @param node + * pointer to the rib node + * @param depth + * pointer to the depth to save + * @return + * 0 on success. + * -1 on failure with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib_get_depth(struct rte_rib_node *node, uint8_t *depth); + +/** + * Get ext field from the rib node + * It is caller responsibility to make sure there are necessary space + * for the ext field inside rib node. + * + * @param node + * pointer to the rib node + * @return + * pointer to the ext + */ +__rte_experimental +void * +rte_rib_get_ext(struct rte_rib_node *node); + +/** + * Get nexthop from the rib node + * + * @param node + * pointer to the rib node + * @param nh + * pointer to the nexthop to save + * @return + * 0 on success. + * -1 on failure with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib_get_nh(struct rte_rib_node *node, uint64_t *nh); + +/** + * Set nexthop into the rib node + * + * @param node + * pointer to the rib node + * @param nh + * nexthop value to set to the rib node + * @return + * 0 on success. + * -1 on failure with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib_set_nh(struct rte_rib_node *node, uint64_t nh); + +/** + * Create RIB + * + * @param name + * RIB name + * @param socket_id + * NUMA socket ID for RIB table memory allocation + * @param conf + * Structure containing the configuration + * @return + * Handle to RIB object on success + * NULL otherwise with rte_errno indicating reason for failure. + */ +__rte_experimental +struct rte_rib * +rte_rib_create(const char *name, int socket_id, struct rte_rib_conf *conf); + +/** + * Find an existing RIB object and return a pointer to it. + * + * @param name + * Name of the rib object as passed to rte_rib_create() + * @return + * Pointer to RIB object on success + * NULL otherwise with rte_errno indicating reason for failure. + */ +__rte_experimental +struct rte_rib * +rte_rib_find_existing(const char *name); + +/** + * Free an RIB object. + * + * @param rib + * RIB object handle + * @return + * None + */ +__rte_experimental +void +rte_rib_free(struct rte_rib *rib); + +#endif /* _RTE_RIB_H_ */ diff --git a/src/spdk/dpdk/lib/librte_rib/rte_rib6.c b/src/spdk/dpdk/lib/librte_rib/rte_rib6.c new file mode 100644 index 000000000..78b8dcfd9 --- /dev/null +++ b/src/spdk/dpdk/lib/librte_rib/rte_rib6.c @@ -0,0 +1,598 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Vladimir Medvedkin <medvedkinv@gmail.com> + * Copyright(c) 2019 Intel Corporation + */ + +#include <stdbool.h> +#include <stdint.h> + +#include <rte_eal.h> +#include <rte_eal_memconfig.h> +#include <rte_errno.h> +#include <rte_malloc.h> +#include <rte_mempool.h> +#include <rte_rwlock.h> +#include <rte_string_fns.h> +#include <rte_tailq.h> + +#include <rte_rib6.h> + +#define RTE_RIB_VALID_NODE 1 +#define RIB6_MAXDEPTH 128 +/* Maximum length of a RIB6 name. */ +#define RTE_RIB6_NAMESIZE 64 + +TAILQ_HEAD(rte_rib6_list, rte_tailq_entry); +static struct rte_tailq_elem rte_rib6_tailq = { + .name = "RTE_RIB6", +}; +EAL_REGISTER_TAILQ(rte_rib6_tailq) + +struct rte_rib6_node { + struct rte_rib6_node *left; + struct rte_rib6_node *right; + struct rte_rib6_node *parent; + uint64_t nh; + uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE]; + uint8_t depth; + uint8_t flag; + __extension__ uint64_t ext[0]; +}; + +struct rte_rib6 { + char name[RTE_RIB6_NAMESIZE]; + struct rte_rib6_node *tree; + struct rte_mempool *node_pool; + uint32_t cur_nodes; + uint32_t cur_routes; + int max_nodes; +}; + +static inline bool +is_valid_node(struct rte_rib6_node *node) +{ + return (node->flag & RTE_RIB_VALID_NODE) == RTE_RIB_VALID_NODE; +} + +static inline bool +is_right_node(struct rte_rib6_node *node) +{ + return node->parent->right == node; +} + +/* + * Check if ip1 is covered by ip2/depth prefix + */ +static inline bool +is_covered(const uint8_t ip1[RTE_RIB6_IPV6_ADDR_SIZE], + const uint8_t ip2[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth) +{ + int i; + + for (i = 0; i < RTE_RIB6_IPV6_ADDR_SIZE; i++) + if ((ip1[i] ^ ip2[i]) & get_msk_part(depth, i)) + return false; + + return true; +} + +static inline int +get_dir(const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth) +{ + int i = 0; + uint8_t p_depth, msk; + + for (p_depth = depth; p_depth >= 8; p_depth -= 8) + i++; + + msk = 1 << (7 - p_depth); + return (ip[i] & msk) != 0; +} + +static inline struct rte_rib6_node * +get_nxt_node(struct rte_rib6_node *node, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE]) +{ + return (get_dir(ip, node->depth)) ? node->right : node->left; +} + +static struct rte_rib6_node * +node_alloc(struct rte_rib6 *rib) +{ + struct rte_rib6_node *ent; + int ret; + + ret = rte_mempool_get(rib->node_pool, (void *)&ent); + if (unlikely(ret != 0)) + return NULL; + ++rib->cur_nodes; + return ent; +} + +static void +node_free(struct rte_rib6 *rib, struct rte_rib6_node *ent) +{ + --rib->cur_nodes; + rte_mempool_put(rib->node_pool, ent); +} + +struct rte_rib6_node * +rte_rib6_lookup(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE]) +{ + struct rte_rib6_node *cur; + struct rte_rib6_node *prev = NULL; + + if (unlikely(rib == NULL)) { + rte_errno = EINVAL; + return NULL; + } + cur = rib->tree; + + while ((cur != NULL) && is_covered(ip, cur->ip, cur->depth)) { + if (is_valid_node(cur)) + prev = cur; + cur = get_nxt_node(cur, ip); + } + return prev; +} + +struct rte_rib6_node * +rte_rib6_lookup_parent(struct rte_rib6_node *ent) +{ + struct rte_rib6_node *tmp; + + if (ent == NULL) + return NULL; + + tmp = ent->parent; + while ((tmp != NULL) && (!is_valid_node(tmp))) + tmp = tmp->parent; + + return tmp; +} + +struct rte_rib6_node * +rte_rib6_lookup_exact(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth) +{ + struct rte_rib6_node *cur; + uint8_t tmp_ip[RTE_RIB6_IPV6_ADDR_SIZE]; + int i; + + if ((rib == NULL) || (ip == NULL) || (depth > RIB6_MAXDEPTH)) { + rte_errno = EINVAL; + return NULL; + } + cur = rib->tree; + + for (i = 0; i < RTE_RIB6_IPV6_ADDR_SIZE; i++) + tmp_ip[i] = ip[i] & get_msk_part(depth, i); + + while (cur != NULL) { + if (rte_rib6_is_equal(cur->ip, tmp_ip) && + (cur->depth == depth) && + is_valid_node(cur)) + return cur; + + if (!(is_covered(tmp_ip, cur->ip, cur->depth)) || + (cur->depth >= depth)) + break; + + cur = get_nxt_node(cur, tmp_ip); + } + + return NULL; +} + +/* + * Traverses on subtree and retreeves more specific routes + * for a given in args ip/depth prefix + * last = NULL means the first invocation + */ +struct rte_rib6_node * +rte_rib6_get_nxt(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], + uint8_t depth, struct rte_rib6_node *last, int flag) +{ + struct rte_rib6_node *tmp, *prev = NULL; + uint8_t tmp_ip[RTE_RIB6_IPV6_ADDR_SIZE]; + int i; + + if ((rib == NULL) || (ip == NULL) || (depth > RIB6_MAXDEPTH)) { + rte_errno = EINVAL; + return NULL; + } + + for (i = 0; i < RTE_RIB6_IPV6_ADDR_SIZE; i++) + tmp_ip[i] = ip[i] & get_msk_part(depth, i); + + if (last == NULL) { + tmp = rib->tree; + while ((tmp) && (tmp->depth < depth)) + tmp = get_nxt_node(tmp, tmp_ip); + } else { + tmp = last; + while ((tmp->parent != NULL) && (is_right_node(tmp) || + (tmp->parent->right == NULL))) { + tmp = tmp->parent; + if (is_valid_node(tmp) && + (is_covered(tmp->ip, tmp_ip, depth) && + (tmp->depth > depth))) + return tmp; + } + tmp = (tmp->parent != NULL) ? tmp->parent->right : NULL; + } + while (tmp) { + if (is_valid_node(tmp) && + (is_covered(tmp->ip, tmp_ip, depth) && + (tmp->depth > depth))) { + prev = tmp; + if (flag == RTE_RIB6_GET_NXT_COVER) + return prev; + } + tmp = (tmp->left != NULL) ? tmp->left : tmp->right; + } + return prev; +} + +void +rte_rib6_remove(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth) +{ + struct rte_rib6_node *cur, *prev, *child; + + cur = rte_rib6_lookup_exact(rib, ip, depth); + if (cur == NULL) + return; + + --rib->cur_routes; + cur->flag &= ~RTE_RIB_VALID_NODE; + while (!is_valid_node(cur)) { + if ((cur->left != NULL) && (cur->right != NULL)) + return; + child = (cur->left == NULL) ? cur->right : cur->left; + if (child != NULL) + child->parent = cur->parent; + if (cur->parent == NULL) { + rib->tree = child; + node_free(rib, cur); + return; + } + if (cur->parent->left == cur) + cur->parent->left = child; + else + cur->parent->right = child; + prev = cur; + cur = cur->parent; + node_free(rib, prev); + } +} + +struct rte_rib6_node * +rte_rib6_insert(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth) +{ + struct rte_rib6_node **tmp; + struct rte_rib6_node *prev = NULL; + struct rte_rib6_node *new_node = NULL; + struct rte_rib6_node *common_node = NULL; + uint8_t common_prefix[RTE_RIB6_IPV6_ADDR_SIZE]; + uint8_t tmp_ip[RTE_RIB6_IPV6_ADDR_SIZE]; + int i, d; + uint8_t common_depth, ip_xor; + + if (unlikely((rib == NULL) || (ip == NULL) || + (depth > RIB6_MAXDEPTH))) { + rte_errno = EINVAL; + return NULL; + } + + tmp = &rib->tree; + + for (i = 0; i < RTE_RIB6_IPV6_ADDR_SIZE; i++) + tmp_ip[i] = ip[i] & get_msk_part(depth, i); + + new_node = rte_rib6_lookup_exact(rib, tmp_ip, depth); + if (new_node != NULL) { + rte_errno = EEXIST; + return NULL; + } + + new_node = node_alloc(rib); + if (new_node == NULL) { + rte_errno = ENOMEM; + return NULL; + } + new_node->left = NULL; + new_node->right = NULL; + new_node->parent = NULL; + rte_rib6_copy_addr(new_node->ip, tmp_ip); + new_node->depth = depth; + new_node->flag = RTE_RIB_VALID_NODE; + + /* traverse down the tree to find matching node or closest matching */ + while (1) { + /* insert as the last node in the branch */ + if (*tmp == NULL) { + *tmp = new_node; + new_node->parent = prev; + ++rib->cur_routes; + return *tmp; + } + /* + * Intermediate node found. + * Previous rte_rib6_lookup_exact() returned NULL + * but node with proper search criteria is found. + * Validate intermediate node and return. + */ + if (rte_rib6_is_equal(tmp_ip, (*tmp)->ip) && + (depth == (*tmp)->depth)) { + node_free(rib, new_node); + (*tmp)->flag |= RTE_RIB_VALID_NODE; + ++rib->cur_routes; + return *tmp; + } + + if (!is_covered(tmp_ip, (*tmp)->ip, (*tmp)->depth) || + ((*tmp)->depth >= depth)) { + break; + } + prev = *tmp; + + tmp = (get_dir(tmp_ip, (*tmp)->depth)) ? &(*tmp)->right : + &(*tmp)->left; + } + + /* closest node found, new_node should be inserted in the middle */ + common_depth = RTE_MIN(depth, (*tmp)->depth); + for (i = 0, d = 0; i < RTE_RIB6_IPV6_ADDR_SIZE; i++) { + ip_xor = tmp_ip[i] ^ (*tmp)->ip[i]; + if (ip_xor == 0) + d += 8; + else { + d += __builtin_clz(ip_xor << 24); + break; + } + } + + common_depth = RTE_MIN(d, common_depth); + + for (i = 0; i < RTE_RIB6_IPV6_ADDR_SIZE; i++) + common_prefix[i] = tmp_ip[i] & get_msk_part(common_depth, i); + + if (rte_rib6_is_equal(common_prefix, tmp_ip) && + (common_depth == depth)) { + /* insert as a parent */ + if (get_dir((*tmp)->ip, depth)) + new_node->right = *tmp; + else + new_node->left = *tmp; + new_node->parent = (*tmp)->parent; + (*tmp)->parent = new_node; + *tmp = new_node; + } else { + /* create intermediate node */ + common_node = node_alloc(rib); + if (common_node == NULL) { + node_free(rib, new_node); + rte_errno = ENOMEM; + return NULL; + } + rte_rib6_copy_addr(common_node->ip, common_prefix); + common_node->depth = common_depth; + common_node->flag = 0; + common_node->parent = (*tmp)->parent; + new_node->parent = common_node; + (*tmp)->parent = common_node; + if (get_dir((*tmp)->ip, common_depth) == 1) { + common_node->left = new_node; + common_node->right = *tmp; + } else { + common_node->left = *tmp; + common_node->right = new_node; + } + *tmp = common_node; + } + ++rib->cur_routes; + return new_node; +} + +int +rte_rib6_get_ip(struct rte_rib6_node *node, uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE]) +{ + if ((node == NULL) || (ip == NULL)) { + rte_errno = EINVAL; + return -1; + } + rte_rib6_copy_addr(ip, node->ip); + return 0; +} + +int +rte_rib6_get_depth(struct rte_rib6_node *node, uint8_t *depth) +{ + if ((node == NULL) || (depth == NULL)) { + rte_errno = EINVAL; + return -1; + } + *depth = node->depth; + return 0; +} + +void * +rte_rib6_get_ext(struct rte_rib6_node *node) +{ + return (node == NULL) ? NULL : &node->ext[0]; +} + +int +rte_rib6_get_nh(struct rte_rib6_node *node, uint64_t *nh) +{ + if ((node == NULL) || (nh == NULL)) { + rte_errno = EINVAL; + return -1; + } + *nh = node->nh; + return 0; +} + +int +rte_rib6_set_nh(struct rte_rib6_node *node, uint64_t nh) +{ + if (node == NULL) { + rte_errno = EINVAL; + return -1; + } + node->nh = nh; + return 0; +} + +struct rte_rib6 * +rte_rib6_create(const char *name, int socket_id, struct rte_rib6_conf *conf) +{ + char mem_name[RTE_RIB6_NAMESIZE]; + struct rte_rib6 *rib = NULL; + struct rte_tailq_entry *te; + struct rte_rib6_list *rib6_list; + struct rte_mempool *node_pool; + + /* Check user arguments. */ + if ((name == NULL) || (conf == NULL) || + (conf->max_nodes == 0)) { + rte_errno = EINVAL; + return NULL; + } + + snprintf(mem_name, sizeof(mem_name), "MP_%s", name); + node_pool = rte_mempool_create(mem_name, conf->max_nodes, + sizeof(struct rte_rib6_node) + conf->ext_sz, 0, 0, + NULL, NULL, NULL, NULL, socket_id, 0); + + if (node_pool == NULL) { + RTE_LOG(ERR, LPM, + "Can not allocate mempool for RIB6 %s\n", name); + return NULL; + } + + snprintf(mem_name, sizeof(mem_name), "RIB6_%s", name); + rib6_list = RTE_TAILQ_CAST(rte_rib6_tailq.head, rte_rib6_list); + + rte_mcfg_tailq_write_lock(); + + /* guarantee there's no existing */ + TAILQ_FOREACH(te, rib6_list, next) { + rib = (struct rte_rib6 *)te->data; + if (strncmp(name, rib->name, RTE_RIB6_NAMESIZE) == 0) + break; + } + rib = NULL; + if (te != NULL) { + rte_errno = EEXIST; + goto exit; + } + + /* allocate tailq entry */ + te = rte_zmalloc("RIB6_TAILQ_ENTRY", sizeof(*te), 0); + if (te == NULL) { + RTE_LOG(ERR, LPM, + "Can not allocate tailq entry for RIB6 %s\n", name); + rte_errno = ENOMEM; + goto exit; + } + + /* Allocate memory to store the RIB6 data structures. */ + rib = rte_zmalloc_socket(mem_name, + sizeof(struct rte_rib6), RTE_CACHE_LINE_SIZE, socket_id); + if (rib == NULL) { + RTE_LOG(ERR, LPM, "RIB6 %s memory allocation failed\n", name); + rte_errno = ENOMEM; + goto free_te; + } + + rte_strlcpy(rib->name, name, sizeof(rib->name)); + rib->tree = NULL; + rib->max_nodes = conf->max_nodes; + rib->node_pool = node_pool; + + te->data = (void *)rib; + TAILQ_INSERT_TAIL(rib6_list, te, next); + + rte_mcfg_tailq_write_unlock(); + + return rib; + +free_te: + rte_free(te); +exit: + rte_mcfg_tailq_write_unlock(); + rte_mempool_free(node_pool); + + return NULL; +} + +struct rte_rib6 * +rte_rib6_find_existing(const char *name) +{ + struct rte_rib6 *rib = NULL; + struct rte_tailq_entry *te; + struct rte_rib6_list *rib6_list; + + if (unlikely(name == NULL)) { + rte_errno = EINVAL; + return NULL; + } + + rib6_list = RTE_TAILQ_CAST(rte_rib6_tailq.head, rte_rib6_list); + + rte_mcfg_tailq_read_lock(); + TAILQ_FOREACH(te, rib6_list, next) { + rib = (struct rte_rib6 *) te->data; + if (strncmp(name, rib->name, RTE_RIB6_NAMESIZE) == 0) + break; + } + rte_mcfg_tailq_read_unlock(); + + if (te == NULL) { + rte_errno = ENOENT; + return NULL; + } + + return rib; +} + +void +rte_rib6_free(struct rte_rib6 *rib) +{ + struct rte_tailq_entry *te; + struct rte_rib6_list *rib6_list; + struct rte_rib6_node *tmp = NULL; + + if (unlikely(rib == NULL)) { + rte_errno = EINVAL; + return; + } + + rib6_list = RTE_TAILQ_CAST(rte_rib6_tailq.head, rte_rib6_list); + + rte_mcfg_tailq_write_lock(); + + /* find our tailq entry */ + TAILQ_FOREACH(te, rib6_list, next) { + if (te->data == (void *)rib) + break; + } + if (te != NULL) + TAILQ_REMOVE(rib6_list, te, next); + + rte_mcfg_tailq_write_unlock(); + + while ((tmp = rte_rib6_get_nxt(rib, 0, 0, tmp, + RTE_RIB6_GET_NXT_ALL)) != NULL) + rte_rib6_remove(rib, tmp->ip, tmp->depth); + + rte_mempool_free(rib->node_pool); + + rte_free(rib); + rte_free(te); +} diff --git a/src/spdk/dpdk/lib/librte_rib/rte_rib6.h b/src/spdk/dpdk/lib/librte_rib/rte_rib6.h new file mode 100644 index 000000000..871457138 --- /dev/null +++ b/src/spdk/dpdk/lib/librte_rib/rte_rib6.h @@ -0,0 +1,334 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Vladimir Medvedkin <medvedkinv@gmail.com> + * Copyright(c) 2019 Intel Corporation + */ + +#ifndef _RTE_RIB6_H_ +#define _RTE_RIB6_H_ + +/** + * @file + * Level compressed tree implementation for IPv6 Longest Prefix Match + */ + +#include <rte_memcpy.h> +#include <rte_compat.h> + +#define RTE_RIB6_IPV6_ADDR_SIZE 16 + +/** + * rte_rib6_get_nxt() flags + */ +enum { + /** flag to get all subroutes in a RIB tree */ + RTE_RIB6_GET_NXT_ALL, + /** flag to get first matched subroutes in a RIB tree */ + RTE_RIB6_GET_NXT_COVER +}; + +struct rte_rib6; +struct rte_rib6_node; + +/** RIB configuration structure */ +struct rte_rib6_conf { + /** + * Size of extension block inside rte_rib_node. + * This space could be used to store additional user + * defined data. + */ + size_t ext_sz; + /* size of rte_rib_node's pool */ + int max_nodes; +}; + +/** + * Copy IPv6 address from one location to another + * + * @param dst + * pointer to the place to copy + * @param src + * pointer from where to copy + */ +static inline void +rte_rib6_copy_addr(uint8_t *dst, const uint8_t *src) +{ + if ((dst == NULL) || (src == NULL)) + return; + rte_memcpy(dst, src, RTE_RIB6_IPV6_ADDR_SIZE); +} + +/** + * Compare two IPv6 addresses + * + * @param ip1 + * pointer to the first ipv6 address + * @param ip2 + * pointer to the second ipv6 address + * + * @return + * 1 if equal + * 0 otherwise + */ +static inline int +rte_rib6_is_equal(uint8_t *ip1, uint8_t *ip2) { + int i; + + if ((ip1 == NULL) || (ip2 == NULL)) + return 0; + for (i = 0; i < RTE_RIB6_IPV6_ADDR_SIZE; i++) { + if (ip1[i] != ip2[i]) + return 0; + } + return 1; +} + +/** + * Get 8-bit part of 128-bit IPv6 mask + * + * @param depth + * ipv6 prefix length + * @param byte + * position of a 8-bit chunk in the 128-bit mask + * + * @return + * 8-bit chunk of the 128-bit IPv6 mask + */ +static inline uint8_t +get_msk_part(uint8_t depth, int byte) { + uint8_t part; + + byte &= 0xf; + depth = RTE_MIN(depth, 128); + part = RTE_MAX((int16_t)depth - (byte * 8), 0); + part = (part > 8) ? 8 : part; + return (uint16_t)(~UINT8_MAX) >> part; +} + +/** + * Lookup an IP into the RIB structure + * + * @param rib + * RIB object handle + * @param ip + * IP to be looked up in the RIB + * @return + * pointer to struct rte_rib6_node on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib6_node * +rte_rib6_lookup(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE]); + +/** + * Lookup less specific route into the RIB structure + * + * @param ent + * Pointer to struct rte_rib6_node that represents target route + * @return + * pointer to struct rte_rib6_node that represents + * less specific route on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib6_node * +rte_rib6_lookup_parent(struct rte_rib6_node *ent); + +/** + * Provides exact mach lookup of the prefix into the RIB structure + * + * @param rib + * RIB object handle + * @param ip + * net to be looked up in the RIB + * @param depth + * prefix length + * @return + * pointer to struct rte_rib6_node on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib6_node * +rte_rib6_lookup_exact(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth); + +/** + * Retrieve next more specific prefix from the RIB + * that is covered by ip/depth supernet in an ascending order + * + * @param rib + * RIB object handle + * @param ip + * net address of supernet prefix that covers returned more specific prefixes + * @param depth + * supernet prefix length + * @param last + * pointer to the last returned prefix to get next prefix + * or + * NULL to get first more specific prefix + * @param flag + * -RTE_RIB6_GET_NXT_ALL + * get all prefixes from subtrie + * -RTE_RIB6_GET_NXT_COVER + * get only first more specific prefix even if it have more specifics + * @return + * pointer to the next more specific prefix + * NULL if there is no prefixes left + */ +__rte_experimental +struct rte_rib6_node * +rte_rib6_get_nxt(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], + uint8_t depth, struct rte_rib6_node *last, int flag); + +/** + * Remove prefix from the RIB + * + * @param rib + * RIB object handle + * @param ip + * net to be removed from the RIB + * @param depth + * prefix length + */ +__rte_experimental +void +rte_rib6_remove(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth); + +/** + * Insert prefix into the RIB + * + * @param rib + * RIB object handle + * @param ip + * net to be inserted to the RIB + * @param depth + * prefix length + * @return + * pointer to new rte_rib6_node on success + * NULL otherwise + */ +__rte_experimental +struct rte_rib6_node * +rte_rib6_insert(struct rte_rib6 *rib, + const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth); + +/** + * Get an ip from rte_rib6_node + * + * @param node + * pointer to the rib6 node + * @param ip + * pointer to the ipv6 to save + * @return + * 0 on success + * -1 on failure with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib6_get_ip(struct rte_rib6_node *node, + uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE]); + +/** + * Get a depth from rte_rib6_node + * + * @param node + * pointer to the rib6 node + * @param depth + * pointer to the depth to save + * @return + * 0 on success + * -1 on failure with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib6_get_depth(struct rte_rib6_node *node, uint8_t *depth); + +/** + * Get ext field from the rte_rib6_node + * It is caller responsibility to make sure there are necessary space + * for the ext field inside rib6 node. + * + * @param node + * pointer to the rte_rib6_node + * @return + * pointer to the ext + */ +__rte_experimental +void * +rte_rib6_get_ext(struct rte_rib6_node *node); + +/** + * Get nexthop from the rte_rib6_node + * + * @param node + * pointer to the rib6 node + * @param nh + * pointer to the nexthop to save + * @return + * 0 on success + * -1 on failure, with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib6_get_nh(struct rte_rib6_node *node, uint64_t *nh); + +/** + * Set nexthop into the rte_rib6_node + * + * @param node + * pointer to the rib6 node + * @param nh + * nexthop value to set to the rib6 node + * @return + * 0 on success + * -1 on failure, with rte_errno indicating reason for failure. + */ +__rte_experimental +int +rte_rib6_set_nh(struct rte_rib6_node *node, uint64_t nh); + +/** + * Create RIB + * + * @param name + * RIB name + * @param socket_id + * NUMA socket ID for RIB table memory allocation + * @param conf + * Structure containing the configuration + * @return + * Pointer to RIB object on success + * NULL otherwise with rte_errno indicating reason for failure. + */ +__rte_experimental +struct rte_rib6 * +rte_rib6_create(const char *name, int socket_id, struct rte_rib6_conf *conf); + +/** + * Find an existing RIB object and return a pointer to it. + * + * @param name + * Name of the rib object as passed to rte_rib_create() + * @return + * Pointer to RIB object on success + * NULL otherwise with rte_errno indicating reason for failure. + */ +__rte_experimental +struct rte_rib6 * +rte_rib6_find_existing(const char *name); + +/** + * Free an RIB object. + * + * @param rib + * RIB object handle + * @return + * None + */ +__rte_experimental +void +rte_rib6_free(struct rte_rib6 *rib); + +#endif /* _RTE_RIB_H_ */ diff --git a/src/spdk/dpdk/lib/librte_rib/rte_rib_version.map b/src/spdk/dpdk/lib/librte_rib/rte_rib_version.map new file mode 100644 index 000000000..9b6161a09 --- /dev/null +++ b/src/spdk/dpdk/lib/librte_rib/rte_rib_version.map @@ -0,0 +1,35 @@ +EXPERIMENTAL { + global: + + rte_rib_create; + rte_rib_find_existing; + rte_rib_free; + rte_rib_get_depth; + rte_rib_get_ext; + rte_rib_get_ip; + rte_rib_get_nh; + rte_rib_get_nxt; + rte_rib_insert; + rte_rib_lookup; + rte_rib_lookup_parent; + rte_rib_lookup_exact; + rte_rib_set_nh; + rte_rib_remove; + + rte_rib6_create; + rte_rib6_find_existing; + rte_rib6_free; + rte_rib6_get_depth; + rte_rib6_get_ext; + rte_rib6_get_ip; + rte_rib6_get_nh; + rte_rib6_get_nxt; + rte_rib6_insert; + rte_rib6_lookup; + rte_rib6_lookup_parent; + rte_rib6_lookup_exact; + rte_rib6_set_nh; + rte_rib6_remove; + + local: *; +}; |