diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-07-23 09:41:28 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-07-23 11:00:43 +0000 |
commit | d0b1bae8c5c70c5d06f3dcecc450a75e7f7cb5af (patch) | |
tree | 7ea7c5e622a5d7c9c989057a1eca8954c4d7fefb /src/schema_compile.c | |
parent | Initial commit. (diff) | |
download | libyang3-d0b1bae8c5c70c5d06f3dcecc450a75e7f7cb5af.tar.xz libyang3-d0b1bae8c5c70c5d06f3dcecc450a75e7f7cb5af.zip |
Adding upstream version 3.1.0+dfsg.upstream/3.1.0+dfsg
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/schema_compile.c')
-rw-r--r-- | src/schema_compile.c | 1936 |
1 files changed, 1936 insertions, 0 deletions
diff --git a/src/schema_compile.c b/src/schema_compile.c new file mode 100644 index 0000000..aa9a3d0 --- /dev/null +++ b/src/schema_compile.c @@ -0,0 +1,1936 @@ +/** + * @file schema_compile.c + * @author Radek Krejci <rkrejci@cesnet.cz> + * @author Michal Vasko <mvasko@cesnet.cz> + * @brief Schema compilation. + * + * Copyright (c) 2015 - 2022 CESNET, z.s.p.o. + * + * This source code is licensed under BSD 3-Clause License (the "License"). + * You may not use this file except in compliance with the License. + * You may obtain a copy of the License at + * + * https://opensource.org/licenses/BSD-3-Clause + */ + +#define _GNU_SOURCE + +#include "schema_compile.h" + +#include <assert.h> +#include <stddef.h> +#include <stdint.h> +#include <stdio.h> +#include <stdlib.h> +#include <string.h> + +#include "compat.h" +#include "context.h" +#include "dict.h" +#include "in.h" +#include "log.h" +#include "ly_common.h" +#include "parser_schema.h" +#include "path.h" +#include "plugins.h" +#include "plugins_exts.h" +#include "plugins_internal.h" +#include "plugins_types.h" +#include "schema_compile_amend.h" +#include "schema_compile_node.h" +#include "schema_features.h" +#include "set.h" +#include "tree.h" +#include "tree_data.h" +#include "tree_schema.h" +#include "tree_schema_free.h" +#include "tree_schema_internal.h" +#include "xpath.h" + +void +lysc_update_path(struct lysc_ctx *ctx, const struct lys_module *parent_module, const char *name) +{ + int len; + uint8_t nextlevel = 0; /* 0 - no starttag, 1 - '/' starttag, 2 - '=' starttag + '}' endtag */ + + if (!name) { + /* removing last path segment */ + if (ctx->path[ctx->path_len - 1] == '}') { + for ( ; ctx->path[ctx->path_len] != '=' && ctx->path[ctx->path_len] != '{'; --ctx->path_len) {} + if (ctx->path[ctx->path_len] == '=') { + ctx->path[ctx->path_len++] = '}'; + } else { + /* not a top-level special tag, remove also preceiding '/' */ + goto remove_nodelevel; + } + } else { +remove_nodelevel: + for ( ; ctx->path[ctx->path_len] != '/'; --ctx->path_len) {} + if (ctx->path_len == 0) { + /* top-level (last segment) */ + ctx->path_len = 1; + } + } + /* set new terminating NULL-byte */ + ctx->path[ctx->path_len] = '\0'; + } else { + if (ctx->path_len > 1) { + if (!parent_module && (ctx->path[ctx->path_len - 1] == '}') && (ctx->path[ctx->path_len - 2] != '\'')) { + /* extension of the special tag */ + nextlevel = 2; + --ctx->path_len; + } else { + /* there is already some path, so add next level */ + nextlevel = 1; + } + } /* else the path is just initiated with '/', so do not add additional slash in case of top-level nodes */ + + if (nextlevel != 2) { + if ((parent_module && (parent_module == ctx->cur_mod)) || (!parent_module && (ctx->path_len > 1) && (name[0] == '{'))) { + /* module not changed, print the name unprefixed */ + len = snprintf(&ctx->path[ctx->path_len], LYSC_CTX_BUFSIZE - ctx->path_len, "%s%s", + nextlevel ? "/" : "", name); + } else { + len = snprintf(&ctx->path[ctx->path_len], LYSC_CTX_BUFSIZE - ctx->path_len, "%s%s:%s", + nextlevel ? "/" : "", ctx->cur_mod->name, name); + } + } else { + len = snprintf(&ctx->path[ctx->path_len], LYSC_CTX_BUFSIZE - ctx->path_len, "='%s'}", name); + } + if (len >= LYSC_CTX_BUFSIZE - (int)ctx->path_len) { + /* output truncated */ + ctx->path_len = LYSC_CTX_BUFSIZE - 1; + } else { + ctx->path_len += len; + } + } + + ly_log_location_revert(0, 0, 1, 0); + ly_log_location(NULL, NULL, ctx->path, NULL); +} + +/** + * @brief Fill in the prepared compiled extensions definition structure according to the parsed extension definition. + * + * @param[in] ctx Compile context. + * @param[in] extp Parsed extension instance. + * @param[out] ext Compiled extension definition. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_extension(struct lysc_ctx *ctx, struct lysp_ext_instance *extp, struct lysc_ext **ext) +{ + LY_ERR ret = LY_SUCCESS; + struct lysp_ext *ep = extp->def; + + if (!ep->compiled) { + lysc_update_path(ctx, NULL, "{extension}"); + lysc_update_path(ctx, NULL, ep->name); + + /* compile the extension definition */ + *ext = ep->compiled = calloc(1, sizeof **ext); + DUP_STRING_GOTO(ctx->ctx, ep->name, (*ext)->name, ret, cleanup); + DUP_STRING_GOTO(ctx->ctx, ep->argname, (*ext)->argname, ret, cleanup); + LY_CHECK_GOTO(ret = lysp_ext_find_definition(ctx->ctx, extp, (const struct lys_module **)&(*ext)->module, NULL), + cleanup); + + /* compile nested extensions */ + COMPILE_EXTS_GOTO(ctx, ep->exts, (*ext)->exts, *ext, ret, cleanup); + + lysc_update_path(ctx, NULL, NULL); + lysc_update_path(ctx, NULL, NULL); + + /* find extension definition plugin */ + (*ext)->plugin = extp->record ? (struct lyplg_ext *)&extp->record->plugin : NULL; + } + + *ext = ep->compiled; + +cleanup: + if (ret) { + lysc_update_path(ctx, NULL, NULL); + lysc_update_path(ctx, NULL, NULL); + } + return ret; +} + +LY_ERR +lys_compile_ext(struct lysc_ctx *ctx, struct lysp_ext_instance *extp, struct lysc_ext_instance *ext, void *parent) +{ + LY_ERR ret = LY_SUCCESS; + + DUP_STRING_GOTO(ctx->ctx, extp->argument, ext->argument, ret, cleanup); + ext->module = ctx->cur_mod; + ext->parent = parent; + ext->parent_stmt = extp->parent_stmt; + ext->parent_stmt_index = extp->parent_stmt_index; + + lysc_update_path(ctx, (ext->parent_stmt & LY_STMT_NODE_MASK) ? ((struct lysc_node *)ext->parent)->module : NULL, + "{extension}"); + lysc_update_path(ctx, NULL, extp->name); + + /* compile extension if not already */ + LY_CHECK_GOTO(ret = lys_compile_extension(ctx, extp, &ext->def), cleanup); + + /* compile */ + if (ext->def->plugin && ext->def->plugin->compile) { + if (ext->argument) { + lysc_update_path(ctx, ext->module, ext->argument); + } + ret = ext->def->plugin->compile(ctx, extp, ext); + if (ret == LY_ENOT) { + lysc_ext_instance_free(&ctx->free_ctx, ext); + } + if (ext->argument) { + lysc_update_path(ctx, NULL, NULL); + } + LY_CHECK_GOTO(ret, cleanup); + } + +cleanup: + lysc_update_path(ctx, NULL, NULL); + lysc_update_path(ctx, NULL, NULL); + return ret; +} + +static void +lysc_unres_must_free(struct lysc_unres_must *m) +{ + LY_ARRAY_FREE(m->local_mods); + free(m); +} + +static void +lysc_unres_dflt_free(const struct ly_ctx *ctx, struct lysc_unres_dflt *r) +{ + assert(!r->dflt || !r->dflts); + if (r->dflt) { + lysp_qname_free((struct ly_ctx *)ctx, r->dflt); + free(r->dflt); + } else { + FREE_ARRAY((struct ly_ctx *)ctx, r->dflts, lysp_qname_free); + } + free(r); +} + +LY_ERR +lys_identity_precompile(struct lysc_ctx *ctx_sc, struct ly_ctx *ctx, struct lysp_module *parsed_mod, + const struct lysp_ident *identities_p, struct lysc_ident **identities) +{ + LY_ARRAY_COUNT_TYPE u; + struct lysc_ctx cctx; + struct lysc_ident *ident; + LY_ERR ret = LY_SUCCESS; + + assert(ctx_sc || ctx); + + if (!ctx_sc) { + if (parsed_mod) { + LYSC_CTX_INIT_PMOD(cctx, parsed_mod, NULL); + } else { + LYSC_CTX_INIT_CTX(cctx, ctx); + } + ctx_sc = &cctx; + } + + if (!identities_p) { + return LY_SUCCESS; + } + + lysc_update_path(ctx_sc, NULL, "{identity}"); + LY_ARRAY_FOR(identities_p, u) { + lysc_update_path(ctx_sc, NULL, identities_p[u].name); + + /* add new compiled identity */ + LY_ARRAY_NEW_GOTO(ctx_sc->ctx, *identities, ident, ret, done); + + DUP_STRING_GOTO(ctx_sc->ctx, identities_p[u].name, ident->name, ret, done); + DUP_STRING_GOTO(ctx_sc->ctx, identities_p[u].dsc, ident->dsc, ret, done); + DUP_STRING_GOTO(ctx_sc->ctx, identities_p[u].ref, ident->ref, ret, done); + ident->module = ctx_sc->cur_mod; + /* backlinks (derived) can be added no sooner than when all the identities in the current module are present */ + COMPILE_EXTS_GOTO(ctx_sc, identities_p[u].exts, ident->exts, ident, ret, done); + ident->flags = identities_p[u].flags; + + lysc_update_path(ctx_sc, NULL, NULL); + } + lysc_update_path(ctx_sc, NULL, NULL); + +done: + if (ret) { + lysc_update_path(ctx_sc, NULL, NULL); + lysc_update_path(ctx_sc, NULL, NULL); + } + return ret; +} + +/** + * @brief Check circular dependency of identities - identity MUST NOT reference itself (via their base statement). + * + * The function works in the same way as lys_compile_feature_circular_check() with different structures and error messages. + * + * @param[in] ctx Compile context for logging. + * @param[in] ident The base identity (its derived list is being extended by the identity being currently processed). + * @param[in] derived The list of derived identities of the identity being currently processed (not the one provided as @p ident) + * @return LY_SUCCESS if everything is ok. + * @return LY_EVALID if the identity is derived from itself. + */ +static LY_ERR +lys_compile_identity_circular_check(struct lysc_ctx *ctx, struct lysc_ident *ident, struct lysc_ident **derived) +{ + LY_ERR ret = LY_SUCCESS; + LY_ARRAY_COUNT_TYPE u, v; + struct ly_set recursion = {0}; + struct lysc_ident *drv; + + if (!derived) { + return LY_SUCCESS; + } + + for (u = 0; u < LY_ARRAY_COUNT(derived); ++u) { + if (ident == derived[u]) { + LOGVAL(ctx->ctx, LYVE_REFERENCE, + "Identity \"%s\" is indirectly derived from itself.", ident->name); + ret = LY_EVALID; + goto cleanup; + } + ret = ly_set_add(&recursion, derived[u], 0, NULL); + LY_CHECK_GOTO(ret, cleanup); + } + + for (v = 0; v < recursion.count; ++v) { + drv = recursion.objs[v]; + for (u = 0; u < LY_ARRAY_COUNT(drv->derived); ++u) { + if (ident == drv->derived[u]) { + LOGVAL(ctx->ctx, LYVE_REFERENCE, + "Identity \"%s\" is indirectly derived from itself.", ident->name); + ret = LY_EVALID; + goto cleanup; + } + ret = ly_set_add(&recursion, drv->derived[u], 0, NULL); + LY_CHECK_GOTO(ret, cleanup); + } + } + +cleanup: + ly_set_erase(&recursion, NULL); + return ret; +} + +LY_ERR +lys_compile_identity_bases(struct lysc_ctx *ctx, const struct lysp_module *base_pmod, const char **bases_p, + struct lysc_ident *ident, struct lysc_ident ***bases) +{ + LY_ARRAY_COUNT_TYPE u, v; + const char *s, *name; + const struct lys_module *mod; + struct lysc_ident **idref; + + assert(ident || bases); + + if ((LY_ARRAY_COUNT(bases_p) > 1) && (ctx->pmod->version < LYS_VERSION_1_1)) { + LOGVAL(ctx->ctx, LYVE_SYNTAX_YANG, + "Multiple bases in %s are allowed only in YANG 1.1 modules.", ident ? "identity" : "identityref type"); + return LY_EVALID; + } + + LY_ARRAY_FOR(bases_p, u) { + s = strchr(bases_p[u], ':'); + if (s) { + /* prefixed identity */ + name = &s[1]; + mod = ly_resolve_prefix(ctx->ctx, bases_p[u], s - bases_p[u], LY_VALUE_SCHEMA, (void *)base_pmod); + } else { + name = bases_p[u]; + mod = base_pmod->mod; + } + if (!mod) { + if (ident) { + LOGVAL(ctx->ctx, LYVE_SYNTAX_YANG, + "Invalid prefix used for base (%s) of identity \"%s\".", bases_p[u], ident->name); + } else { + LOGVAL(ctx->ctx, LYVE_SYNTAX_YANG, + "Invalid prefix used for base (%s) of identityref.", bases_p[u]); + } + return LY_EVALID; + } + + idref = NULL; + LY_ARRAY_FOR(mod->identities, v) { + if (!strcmp(name, mod->identities[v].name)) { + if (ident) { + if (ident == &mod->identities[v]) { + LOGVAL(ctx->ctx, LYVE_REFERENCE, + "Identity \"%s\" is derived from itself.", ident->name); + return LY_EVALID; + } + LY_CHECK_RET(lys_compile_identity_circular_check(ctx, &mod->identities[v], ident->derived)); + /* we have match! store the backlink */ + LY_ARRAY_NEW_RET(ctx->ctx, mod->identities[v].derived, idref, LY_EMEM); + *idref = ident; + } else { + /* we have match! store the found identity */ + LY_ARRAY_NEW_RET(ctx->ctx, *bases, idref, LY_EMEM); + *idref = &mod->identities[v]; + } + break; + } + } + if (!idref) { + if (ident) { + LOGVAL(ctx->ctx, LYVE_SYNTAX_YANG, + "Unable to find base (%s) of identity \"%s\".", bases_p[u], ident->name); + } else { + LOGVAL(ctx->ctx, LYVE_SYNTAX_YANG, + "Unable to find base (%s) of identityref.", bases_p[u]); + } + return LY_EVALID; + } + } + + return LY_SUCCESS; +} + +/** + * @brief For the given array of identities, set the backlinks from all their base identities. + * + * @param[in] ctx Compile context, not only for logging but also to get the current module to resolve prefixes. + * @param[in] idents_p Array of identities definitions from the parsed schema structure. + * @param[in,out] idents Array of referencing identities to which the backlinks are supposed to be set. + * @return LY_ERR value - LY_SUCCESS or LY_EVALID. + */ +static LY_ERR +lys_compile_identities_derived(struct lysc_ctx *ctx, struct lysp_ident *idents_p, struct lysc_ident **idents) +{ + LY_ARRAY_COUNT_TYPE u, v; + + lysc_update_path(ctx, NULL, "{identity}"); + + for (u = 0; u < LY_ARRAY_COUNT(*idents); ++u) { + /* find matching parsed identity */ + for (v = 0; v < LY_ARRAY_COUNT(idents_p); ++v) { + if (idents_p[v].name == (*idents)[u].name) { + break; + } + } + + if ((v == LY_ARRAY_COUNT(idents_p)) || !idents_p[v].bases) { + /* identity not found (it may be from a submodule) or identity without bases */ + continue; + } + + lysc_update_path(ctx, NULL, (*idents)[u].name); + LY_CHECK_RET(lys_compile_identity_bases(ctx, ctx->pmod, idents_p[v].bases, &(*idents)[u], NULL)); + lysc_update_path(ctx, NULL, NULL); + } + + lysc_update_path(ctx, NULL, NULL); + return LY_SUCCESS; +} + +LY_ERR +lys_compile_expr_implement(const struct ly_ctx *ctx, const struct lyxp_expr *expr, LY_VALUE_FORMAT format, + void *prefix_data, ly_bool implement, struct lys_glob_unres *unres, const struct lys_module **mod_p) +{ + uint32_t i; + const char *ptr, *start, **imp_f, *all_f[] = {"*", NULL}; + const struct lys_module *mod; + + assert(implement || mod_p); + + if (mod_p) { + *mod_p = NULL; + } + + for (i = 0; i < expr->used; ++i) { + if ((expr->tokens[i] != LYXP_TOKEN_NAMETEST) && (expr->tokens[i] != LYXP_TOKEN_LITERAL)) { + /* token cannot have a prefix */ + continue; + } + + start = expr->expr + expr->tok_pos[i]; + if (!(ptr = ly_strnchr(start, ':', expr->tok_len[i]))) { + /* token without a prefix */ + continue; + } + + if (!(mod = ly_resolve_prefix(ctx, start, ptr - start, format, prefix_data))) { + /* unknown prefix, do not care right now */ + continue; + } + + /* unimplemented module found */ + if (!mod->implemented && !implement) { + /* should not be implemented now */ + *mod_p = mod; + break; + } + + if (!mod->implemented) { + /* implement if not implemented */ + imp_f = (ctx->flags & LY_CTX_ENABLE_IMP_FEATURES) ? all_f : NULL; + LY_CHECK_RET(lys_implement((struct lys_module *)mod, imp_f, unres)); + } + if (!mod->compiled) { + /* compile if not implemented before or only marked for compilation */ + LY_CHECK_RET(lys_compile((struct lys_module *)mod, &unres->ds_unres)); + } + } + + return LY_SUCCESS; +} + +/** + * @brief Check when for cyclic dependencies. + * + * @param[in] set Set with all the referenced nodes. + * @param[in] node Node whose "when" referenced nodes are in @p set. + * @return LY_ERR value + */ +static LY_ERR +lys_compile_unres_when_cyclic(struct lyxp_set *set, const struct lysc_node *node) +{ + struct lyxp_set tmp_set; + struct lyxp_set_scnode *xp_scnode; + uint32_t i, j, idx; + LY_ARRAY_COUNT_TYPE u; + LY_ERR ret = LY_SUCCESS; + + memset(&tmp_set, 0, sizeof tmp_set); + + /* prepare in_ctx of the set */ + for (i = 0; i < set->used; ++i) { + xp_scnode = &set->val.scnodes[i]; + + if (xp_scnode->in_ctx != LYXP_SET_SCNODE_START_USED) { + /* check node when, skip the context node (it was just checked) */ + xp_scnode->in_ctx = LYXP_SET_SCNODE_ATOM_CTX; + } + } + + for (i = 0; i < set->used; ++i) { + xp_scnode = &set->val.scnodes[i]; + if (xp_scnode->in_ctx != LYXP_SET_SCNODE_ATOM_CTX) { + /* already checked */ + continue; + } + + if ((xp_scnode->type != LYXP_NODE_ELEM) || !lysc_node_when(xp_scnode->scnode)) { + /* no when to check */ + xp_scnode->in_ctx = LYXP_SET_SCNODE_ATOM_NODE; + continue; + } + + node = xp_scnode->scnode; + do { + struct lysc_when **when_list, *when; + + LOG_LOCSET(node, NULL); + when_list = lysc_node_when(node); + LY_ARRAY_FOR(when_list, u) { + when = when_list[u]; + ret = lyxp_atomize(set->ctx, when->cond, node->module, LY_VALUE_SCHEMA_RESOLVED, when->prefixes, + when->context, when->context, &tmp_set, LYXP_SCNODE_SCHEMA); + if (ret != LY_SUCCESS) { + LOGVAL(set->ctx, LYVE_SEMANTICS, "Invalid when condition \"%s\".", when->cond->expr); + LOG_LOCBACK(1, 0); + goto cleanup; + } + + for (j = 0; j < tmp_set.used; ++j) { + if (tmp_set.val.scnodes[j].type != LYXP_NODE_ELEM) { + /* skip roots'n'stuff, no when, nothing to check */ + tmp_set.val.scnodes[j].in_ctx = LYXP_SET_SCNODE_ATOM_NODE; + continue; + } + + /* try to find this node in our set */ + if (lyxp_set_scnode_contains(set, tmp_set.val.scnodes[j].scnode, LYXP_NODE_ELEM, -1, &idx) && + (set->val.scnodes[idx].in_ctx == LYXP_SET_SCNODE_START_USED)) { + LOGVAL(set->ctx, LYVE_SEMANTICS, "When condition cyclic dependency on the node \"%s\".", + tmp_set.val.scnodes[j].scnode->name); + ret = LY_EVALID; + LOG_LOCBACK(1, 0); + goto cleanup; + } + + /* needs to be checked, if in both sets, will be ignored */ + tmp_set.val.scnodes[j].in_ctx = LYXP_SET_SCNODE_ATOM_CTX; + } + + if (when->context != node) { + /* node actually depends on this "when", not the context node */ + assert(tmp_set.val.scnodes[0].scnode == when->context); + if (tmp_set.val.scnodes[0].in_ctx == LYXP_SET_SCNODE_START_USED) { + /* replace the non-traversed context node with the dependent node */ + tmp_set.val.scnodes[0].scnode = (struct lysc_node *)node; + } else { + /* context node was traversed, so just add the dependent node */ + ret = lyxp_set_scnode_insert_node(&tmp_set, node, LYXP_SET_SCNODE_START_USED, LYXP_AXIS_CHILD, NULL); + LY_CHECK_ERR_GOTO(ret, LOG_LOCBACK(1, 0), cleanup); + } + } + + /* merge this set into the global when set */ + lyxp_set_scnode_merge(set, &tmp_set); + } + LOG_LOCBACK(1, 0); + + /* check when of non-data parents as well */ + node = node->parent; + } while (node && (node->nodetype & (LYS_CASE | LYS_CHOICE))); + + /* this node when was checked (xp_scnode could have been reallocd) */ + set->val.scnodes[i].in_ctx = LYXP_SET_SCNODE_ATOM_NODE; + } + +cleanup: + lyxp_set_free_content(&tmp_set); + return ret; +} + +LY_ERR +lysc_check_status(struct lysc_ctx *ctx, uint16_t flags1, void *mod1, const char *name1, uint16_t flags2, void *mod2, + const char *name2) +{ + uint16_t flg1, flg2; + + flg1 = (flags1 & LYS_STATUS_MASK) ? (flags1 & LYS_STATUS_MASK) : LYS_STATUS_CURR; + flg2 = (flags2 & LYS_STATUS_MASK) ? (flags2 & LYS_STATUS_MASK) : LYS_STATUS_CURR; + + if ((flg1 < flg2) && (mod1 == mod2)) { + if (ctx) { + LOGVAL(ctx->ctx, LYVE_REFERENCE, "A %s definition \"%s\" is not allowed to reference %s definition \"%s\".", + flg1 == LYS_STATUS_CURR ? "current" : "deprecated", name1, + flg2 == LYS_STATUS_OBSLT ? "obsolete" : "deprecated", name2); + } + return LY_EVALID; + } + + return LY_SUCCESS; +} + +/** + * @brief Check when expressions of a node on a complete compiled schema tree. + * + * @param[in] ctx Compile context. + * @param[in] when When to check. + * @param[in] node Node with @p when. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_when(struct lysc_ctx *ctx, const struct lysc_when *when, const struct lysc_node *node) +{ + struct lyxp_set tmp_set = {0}; + uint32_t i, opts; + struct lysc_node *schema; + LY_ERR ret = LY_SUCCESS; + + opts = LYXP_SCNODE_SCHEMA | ((node->flags & LYS_IS_OUTPUT) ? LYXP_SCNODE_OUTPUT : 0); + + /* check "when" */ + ret = lyxp_atomize(ctx->ctx, when->cond, node->module, LY_VALUE_SCHEMA_RESOLVED, when->prefixes, when->context, + when->context, &tmp_set, opts); + if (ret) { + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "Invalid when condition \"%s\".", when->cond->expr); + goto cleanup; + } + + ctx->path[0] = '\0'; + lysc_path(node, LYSC_PATH_LOG, ctx->path, LYSC_CTX_BUFSIZE); + for (i = 0; i < tmp_set.used; ++i) { + if (tmp_set.val.scnodes[i].type != LYXP_NODE_ELEM) { + /* skip roots'n'stuff */ + continue; + } else if (tmp_set.val.scnodes[i].in_ctx == LYXP_SET_SCNODE_START_USED) { + /* context node not actually traversed */ + continue; + } + + schema = tmp_set.val.scnodes[i].scnode; + + /* XPath expression cannot reference "lower" status than the node that has the definition */ + if (lysc_check_status(NULL, when->flags, node->module, node->name, schema->flags, schema->module, + schema->name)) { + LOGWRN(ctx->ctx, "When condition \"%s\" may be referencing %s node \"%s\".", when->cond->expr, + (schema->flags == LYS_STATUS_OBSLT) ? "obsolete" : "deprecated", schema->name); + } + + /* check dummy node children/value accessing */ + if (lysc_data_parent(schema) == node) { + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "When condition is accessing its own conditional node children."); + ret = LY_EVALID; + goto cleanup; + } else if ((schema == node) && (tmp_set.val.scnodes[i].in_ctx == LYXP_SET_SCNODE_ATOM_VAL)) { + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "When condition is accessing its own conditional node value."); + ret = LY_EVALID; + goto cleanup; + } + } + + if (when->context != node) { + /* node actually depends on this "when", not the context node */ + assert(tmp_set.val.scnodes[0].scnode == when->context); + if (tmp_set.val.scnodes[0].in_ctx == LYXP_SET_SCNODE_START_USED) { + /* replace the non-traversed context node with the dependent node */ + tmp_set.val.scnodes[0].scnode = (struct lysc_node *)node; + } else { + /* context node was traversed, so just add the dependent node */ + ret = lyxp_set_scnode_insert_node(&tmp_set, node, LYXP_SET_SCNODE_START_USED, LYXP_AXIS_CHILD, NULL); + LY_CHECK_GOTO(ret, cleanup); + } + } + + /* check cyclic dependencies */ + ret = lys_compile_unres_when_cyclic(&tmp_set, node); + LY_CHECK_GOTO(ret, cleanup); + +cleanup: + lyxp_set_free_content(&tmp_set); + return ret; +} + +/** + * @brief Check must expressions of a node on a complete compiled schema tree. + * + * @param[in] ctx Compile context. + * @param[in] node Node to check. + * @param[in] local_mods Sized array of local modules for musts of @p node at the same index. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_must(struct lysc_ctx *ctx, const struct lysc_node *node, const struct lysp_module **local_mods) +{ + struct lyxp_set tmp_set; + uint32_t i, opts; + LY_ARRAY_COUNT_TYPE u; + struct lysc_must *musts; + LY_ERR ret = LY_SUCCESS; + uint16_t flg; + + LOG_LOCSET(node, NULL); + + memset(&tmp_set, 0, sizeof tmp_set); + opts = LYXP_SCNODE_SCHEMA | ((node->flags & LYS_IS_OUTPUT) ? LYXP_SCNODE_OUTPUT : 0); + + musts = lysc_node_musts(node); + LY_ARRAY_FOR(musts, u) { + /* check "must" */ + ret = lyxp_atomize(ctx->ctx, musts[u].cond, node->module, LY_VALUE_SCHEMA_RESOLVED, musts[u].prefixes, node, + node, &tmp_set, opts); + if (ret) { + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "Invalid must condition \"%s\".", musts[u].cond->expr); + goto cleanup; + } + + ctx->path[0] = '\0'; + lysc_path(node, LYSC_PATH_LOG, ctx->path, LYSC_CTX_BUFSIZE); + for (i = 0; i < tmp_set.used; ++i) { + /* skip roots'n'stuff */ + if (tmp_set.val.scnodes[i].type == LYXP_NODE_ELEM) { + struct lysc_node *schema = tmp_set.val.scnodes[i].scnode; + + /* XPath expression cannot reference "lower" status than the node that has the definition */ + if (local_mods[u]->mod == node->module) { + /* use flags of the context node since the definition is local */ + flg = node->flags; + } else { + /* definition is foreign (deviation, refine), always current */ + flg = LYS_STATUS_CURR; + } + if (lysc_check_status(NULL, flg, local_mods[u]->mod, node->name, schema->flags, schema->module, + schema->name)) { + LOGWRN(ctx->ctx, "Must condition \"%s\" may be referencing %s node \"%s\".", musts[u].cond->expr, + (schema->flags == LYS_STATUS_OBSLT) ? "obsolete" : "deprecated", schema->name); + break; + } + } + } + + lyxp_set_free_content(&tmp_set); + } + +cleanup: + lyxp_set_free_content(&tmp_set); + LOG_LOCBACK(1, 0); + return ret; +} + +/** + * @brief Remove all disabled bits/enums from a sized array. + * + * @param[in] ctx Context with the dictionary. + * @param[in] items Sized array of bits/enums. + */ +static void +lys_compile_unres_disabled_bitenum_remove(struct lysf_ctx *ctx, struct lysc_type_bitenum_item *items) +{ + LY_ARRAY_COUNT_TYPE u = 0, last_u; + + while (u < LY_ARRAY_COUNT(items)) { + if (items[u].flags & LYS_DISABLED) { + /* free the disabled item */ + lysc_enum_item_free(ctx, &items[u]); + + /* replace it with the following items */ + last_u = LY_ARRAY_COUNT(items) - 1; + if (u < last_u) { + memmove(items + u, items + u + 1, (last_u - u) * sizeof *items); + } + + /* one item less */ + LY_ARRAY_DECREMENT(items); + continue; + } + + ++u; + } +} + +/** + * @brief Find and remove all disabled bits/enums in a leaf/leaf-list type. + * + * @param[in] ctx Compile context. + * @param[in] leaf Leaf/leaf-list to check. + * @return LY_ERR value + */ +static LY_ERR +lys_compile_unres_disabled_bitenum(struct lysc_ctx *ctx, struct lysc_node_leaf *leaf) +{ + struct lysc_type **t; + LY_ARRAY_COUNT_TYPE u, count; + struct lysc_type_enum *ent; + ly_bool has_value = 0; + + if (leaf->type->basetype == LY_TYPE_UNION) { + t = ((struct lysc_type_union *)leaf->type)->types; + count = LY_ARRAY_COUNT(t); + } else { + t = &leaf->type; + count = 1; + } + for (u = 0; u < count; ++u) { + if ((t[u]->basetype == LY_TYPE_BITS) || (t[u]->basetype == LY_TYPE_ENUM)) { + /* remove all disabled items */ + ent = (struct lysc_type_enum *)(t[u]); + lys_compile_unres_disabled_bitenum_remove(&ctx->free_ctx, ent->enums); + + if (LY_ARRAY_COUNT(ent->enums)) { + has_value = 1; + } + } else { + has_value = 1; + } + } + + if (!has_value) { + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "Node \"%s\" without any (or all disabled) valid values.", leaf->name); + return LY_EVALID; + } + + return LY_SUCCESS; +} + +/** + * @brief Check leafref for its target existence on a complete compiled schema tree. + * + * @param[in] ctx Compile context. + * @param[in] node Context node for the leafref. + * @param[in] lref Leafref to check/resolve. + * @param[in] local_mod Local module for the leafref type. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_leafref(struct lysc_ctx *ctx, const struct lysc_node *node, struct lysc_type_leafref *lref, + const struct lysp_module *local_mod) +{ + const struct lysc_node *target = NULL; + struct ly_path *p; + struct lysc_type *type; + uint16_t flg; + + assert(node->nodetype & (LYS_LEAF | LYS_LEAFLIST)); + + if (lref->realtype) { + /* already resolved, may happen (shared union typedef with a leafref) */ + return LY_SUCCESS; + } + + /* try to find the target, current module is that of the context node (RFC 7950 6.4.1 second bullet) */ + LY_CHECK_RET(ly_path_compile_leafref(ctx->ctx, node, ctx->ext, lref->path, + (node->flags & LYS_IS_OUTPUT) ? LY_PATH_OPER_OUTPUT : LY_PATH_OPER_INPUT, LY_PATH_TARGET_MANY, + LY_VALUE_SCHEMA_RESOLVED, lref->prefixes, &p)); + + /* get the target node */ + target = p[LY_ARRAY_COUNT(p) - 1].node; + ly_path_free(node->module->ctx, p); + + if (!(target->nodetype & (LYS_LEAF | LYS_LEAFLIST))) { + LOGVAL(ctx->ctx, LYVE_REFERENCE, "Invalid leafref path \"%s\" - target node is %s instead of leaf or leaf-list.", + lref->path->expr, lys_nodetype2str(target->nodetype)); + return LY_EVALID; + } + + /* check status */ + ctx->path[0] = '\0'; + lysc_path(node, LYSC_PATH_LOG, ctx->path, LYSC_CTX_BUFSIZE); + ctx->path_len = strlen(ctx->path); + if (node->module == local_mod->mod) { + /* use flags of the context node since the definition is local */ + flg = node->flags; + } else { + /* definition is foreign (deviation), always current */ + flg = LYS_STATUS_CURR; + } + if (lysc_check_status(ctx, flg, local_mod->mod, node->name, target->flags, target->module, target->name)) { + return LY_EVALID; + } + ctx->path_len = 1; + ctx->path[1] = '\0'; + + /* check config */ + if (lref->require_instance) { + if ((node->flags & LYS_CONFIG_W) && (target->flags & LYS_CONFIG_R)) { + LOGVAL(ctx->ctx, LYVE_REFERENCE, "Invalid leafref path \"%s\" - target is supposed" + " to represent configuration data (as the leafref does), but it does not.", lref->path->expr); + return LY_EVALID; + } + } + + /* check for circular chain of leafrefs */ + for (type = ((struct lysc_node_leaf *)target)->type; + type && (type->basetype == LY_TYPE_LEAFREF); + type = ((struct lysc_type_leafref *)type)->realtype) { + if (type == (struct lysc_type *)lref) { + /* circular chain detected */ + LOGVAL(ctx->ctx, LYVE_REFERENCE, "Invalid leafref path \"%s\" - circular chain of leafrefs detected.", + lref->path->expr); + return LY_EVALID; + } + } + + /* store the type */ + lref->realtype = ((struct lysc_node_leaf *)target)->type; + ++lref->realtype->refcount; + return LY_SUCCESS; +} + +/** + * @brief Compile default value(s) for leaf or leaf-list expecting a complete compiled schema tree. + * + * @param[in] ctx Compile context. + * @param[in] node Leaf or leaf-list to compile the default value(s) for. + * @param[in] type Type of the default value. + * @param[in] dflt Default value. + * @param[in] dflt_pmod Parsed module of the @p dflt to resolve possible prefixes. + * @param[in,out] storage Storage for the compiled default value. + * @param[in,out] unres Global unres structure for newly implemented modules. + * @return LY_ERECOMPILE if the whole dep set needs to be recompiled for the value to be checked. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_dflt(struct lysc_ctx *ctx, struct lysc_node *node, struct lysc_type *type, const char *dflt, + const struct lysp_module *dflt_pmod, struct lyd_value *storage, struct lys_glob_unres *unres) +{ + LY_ERR ret; + uint32_t options; + struct ly_err_item *err = NULL; + + options = (ctx->ctx->flags & LY_CTX_REF_IMPLEMENTED) ? LYPLG_TYPE_STORE_IMPLEMENT : 0; + ret = type->plugin->store(ctx->ctx, type, dflt, strlen(dflt), options, LY_VALUE_SCHEMA, (void *)dflt_pmod, + LYD_HINT_SCHEMA, node, storage, unres, &err); + if (ret == LY_ERECOMPILE) { + /* fine, but we need to recompile */ + return LY_ERECOMPILE; + } else if (ret == LY_EINCOMPLETE) { + /* we have no data so we will not be resolving it */ + ret = LY_SUCCESS; + } + + if (ret) { + LOG_LOCSET(node, NULL); + if (err) { + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "Invalid default - value does not fit the type (%s).", err->msg); + ly_err_free(err); + } else { + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "Invalid default - value does not fit the type."); + } + LOG_LOCBACK(1, 0); + return ret; + } + + LY_ATOMIC_INC_BARRIER(((struct lysc_type *)storage->realtype)->refcount); + return LY_SUCCESS; +} + +/** + * @brief Compile default value of a leaf expecting a complete compiled schema tree. + * + * @param[in] ctx Compile context. + * @param[in] leaf Leaf that the default value is for. + * @param[in] dflt Default value to compile. + * @param[in,out] unres Global unres structure for newly implemented modules. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_leaf_dlft(struct lysc_ctx *ctx, struct lysc_node_leaf *leaf, struct lysp_qname *dflt, + struct lys_glob_unres *unres) +{ + LY_ERR ret; + + assert(!leaf->dflt); + + if (leaf->flags & (LYS_MAND_TRUE | LYS_KEY)) { + /* ignore default values for keys and mandatory leaves */ + return LY_SUCCESS; + } + + /* allocate the default value */ + leaf->dflt = calloc(1, sizeof *leaf->dflt); + LY_CHECK_ERR_RET(!leaf->dflt, LOGMEM(ctx->ctx), LY_EMEM); + + /* store the default value */ + ret = lys_compile_unres_dflt(ctx, &leaf->node, leaf->type, dflt->str, dflt->mod, leaf->dflt, unres); + if (ret) { + free(leaf->dflt); + leaf->dflt = NULL; + } + + return ret; +} + +/** + * @brief Compile default values of a leaf-list expecting a complete compiled schema tree. + * + * @param[in] ctx Compile context. + * @param[in] llist Leaf-list that the default value(s) are for. + * @param[in] dflt Default value to compile, in case of a single value. + * @param[in] dflts Sized array of default values, in case of more values. + * @param[in,out] unres Global unres structure for newly implemented modules. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_llist_dflts(struct lysc_ctx *ctx, struct lysc_node_leaflist *llist, struct lysp_qname *dflt, + struct lysp_qname *dflts, struct lys_glob_unres *unres) +{ + LY_ERR ret; + LY_ARRAY_COUNT_TYPE orig_count, u, v; + + assert(dflt || dflts); + + /* in case there were already some defaults and we are adding new by deviations */ + orig_count = LY_ARRAY_COUNT(llist->dflts); + + /* allocate new items */ + LY_ARRAY_CREATE_RET(ctx->ctx, llist->dflts, orig_count + (dflts ? LY_ARRAY_COUNT(dflts) : 1), LY_EMEM); + + /* fill each new default value */ + if (dflts) { + LY_ARRAY_FOR(dflts, u) { + llist->dflts[orig_count + u] = calloc(1, sizeof **llist->dflts); + ret = lys_compile_unres_dflt(ctx, &llist->node, llist->type, dflts[u].str, dflts[u].mod, + llist->dflts[orig_count + u], unres); + LY_CHECK_ERR_RET(ret, free(llist->dflts[orig_count + u]), ret); + LY_ARRAY_INCREMENT(llist->dflts); + } + } else { + llist->dflts[orig_count] = calloc(1, sizeof **llist->dflts); + ret = lys_compile_unres_dflt(ctx, &llist->node, llist->type, dflt->str, dflt->mod, + llist->dflts[orig_count], unres); + LY_CHECK_ERR_RET(ret, free(llist->dflts[orig_count]), ret); + LY_ARRAY_INCREMENT(llist->dflts); + } + + /* check default value uniqueness */ + if (llist->flags & LYS_CONFIG_W) { + /* configuration data values must be unique - so check the default values */ + for (u = orig_count; u < LY_ARRAY_COUNT(llist->dflts); ++u) { + for (v = 0; v < u; ++v) { + if (!llist->dflts[u]->realtype->plugin->compare(ctx->ctx, llist->dflts[u], llist->dflts[v])) { + lysc_update_path(ctx, llist->parent ? llist->parent->module : NULL, llist->name); + LOGVAL(ctx->ctx, LYVE_SEMANTICS, "Configuration leaf-list has multiple defaults of the same value \"%s\".", + (char *)llist->dflts[u]->realtype->plugin->print(ctx->ctx, llist->dflts[u], LY_VALUE_CANON, + NULL, NULL, NULL)); + lysc_update_path(ctx, NULL, NULL); + return LY_EVALID; + } + } + } + } + + return LY_SUCCESS; +} + +/** + * @brief Iteratively get all leafrefs from @p node + * if the node is of type union, otherwise just return the leafref. + * + * @param[in] node Node that may contain the leafref. + * @param[in,out] index Value that is passed between function calls. + * For each new node, initialize value of the @p index to 0, otherwise + * do not modify the value between calls. + * @return Pointer to the leafref or next leafref, otherwise NULL. + */ +static struct lysc_type_leafref * +lys_type_leafref_next(const struct lysc_node *node, LY_ARRAY_COUNT_TYPE *index) +{ + struct lysc_type_leafref *ret = NULL; + struct lysc_type_union *uni; + struct lysc_type *leaf_type; + + assert(node->nodetype & LYD_NODE_TERM); + + leaf_type = ((struct lysc_node_leaf *)node)->type; + if (leaf_type->basetype == LY_TYPE_UNION) { + uni = (struct lysc_type_union *)leaf_type; + + /* find next union leafref */ + while (*index < LY_ARRAY_COUNT(uni->types)) { + if (uni->types[*index]->basetype == LY_TYPE_LEAFREF) { + ret = (struct lysc_type_leafref *)uni->types[*index]; + ++(*index); + break; + } + + ++(*index); + } + } else { + /* return just the single leafref */ + if (*index == 0) { + ++(*index); + assert(leaf_type->basetype == LY_TYPE_LEAFREF); + ret = (struct lysc_type_leafref *)leaf_type; + } + } + + return ret; +} + +/** + * @brief Implement all referenced modules by leafrefs, when and must conditions. + * + * @param[in] ctx libyang context. + * @param[in] unres Global unres structure with the sets to resolve. + * @return LY_SUCCESS on success. + * @return LY_ERECOMPILE if the whole dep set needs to be recompiled with the new implemented modules. + * @return LY_ERR value on error. + */ +static LY_ERR +lys_compile_unres_depset_implement(struct ly_ctx *ctx, struct lys_glob_unres *unres) +{ + struct lys_depset_unres *ds_unres = &unres->ds_unres; + struct lysc_type_leafref *lref; + const struct lys_module *mod; + LY_ARRAY_COUNT_TYPE u; + struct lysc_unres_leafref *l; + struct lysc_unres_when *w; + struct lysc_unres_must *m; + struct lysc_must *musts; + ly_bool not_implemented; + uint32_t di = 0, li = 0, wi = 0, mi = 0; + +implement_all: + /* disabled leafrefs - even those because we need to check their target exists */ + while (di < ds_unres->disabled_leafrefs.count) { + l = ds_unres->disabled_leafrefs.objs[di]; + + u = 0; + while ((lref = lys_type_leafref_next(l->node, &u))) { + LY_CHECK_RET(lys_compile_expr_implement(ctx, lref->path, LY_VALUE_SCHEMA_RESOLVED, lref->prefixes, 1, unres, NULL)); + } + + ++di; + } + + /* leafrefs */ + while (li < ds_unres->leafrefs.count) { + l = ds_unres->leafrefs.objs[li]; + + u = 0; + while ((lref = lys_type_leafref_next(l->node, &u))) { + LY_CHECK_RET(lys_compile_expr_implement(ctx, lref->path, LY_VALUE_SCHEMA_RESOLVED, lref->prefixes, 1, unres, NULL)); + } + + ++li; + } + + /* when conditions */ + while (wi < ds_unres->whens.count) { + w = ds_unres->whens.objs[wi]; + + LY_CHECK_RET(lys_compile_expr_implement(ctx, w->when->cond, LY_VALUE_SCHEMA_RESOLVED, w->when->prefixes, + ctx->flags & LY_CTX_REF_IMPLEMENTED, unres, &mod)); + if (mod) { + LOGWRN(ctx, "When condition \"%s\" check skipped because referenced module \"%s\" is not implemented.", + w->when->cond->expr, mod->name); + + /* remove from the set to skip the check */ + ly_set_rm_index(&ds_unres->whens, wi, free); + continue; + } + + ++wi; + } + + /* must conditions */ + while (mi < ds_unres->musts.count) { + m = ds_unres->musts.objs[mi]; + + not_implemented = 0; + musts = lysc_node_musts(m->node); + LY_ARRAY_FOR(musts, u) { + LY_CHECK_RET(lys_compile_expr_implement(ctx, musts[u].cond, LY_VALUE_SCHEMA_RESOLVED, musts[u].prefixes, + ctx->flags & LY_CTX_REF_IMPLEMENTED, unres, &mod)); + if (mod) { + LOGWRN(ctx, "Must condition \"%s\" check skipped because referenced module \"%s\" is not implemented.", + musts[u].cond->expr, mod->name); + + /* need to implement modules from all the expressions */ + not_implemented = 1; + } + } + + if (not_implemented) { + /* remove from the set to skip the check */ + lysc_unres_must_free(m); + ly_set_rm_index(&ds_unres->musts, mi, NULL); + continue; + } + + ++mi; + } + + if ((di < ds_unres->disabled_leafrefs.count) || (li < ds_unres->leafrefs.count) || (wi < ds_unres->whens.count)) { + /* new items in the sets */ + goto implement_all; + } + + return LY_SUCCESS; +} + +/** + * @brief Check that a disabled node (to be freed) can be freed and is not referenced. + * + * @param[in] node Disabled node to check. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_check_disabled(const struct lysc_node *node) +{ + const struct lysc_node *parent; + struct lysc_node_list *slist; + LY_ARRAY_COUNT_TYPE u, v; + int found; + + if (node->flags & LYS_KEY) { + LOG_LOCSET(node, NULL); + LOGVAL(node->module->ctx, LYVE_REFERENCE, "Key \"%s\" is disabled.", node->name); + LOG_LOCBACK(1, 0); + return LY_EVALID; + } + + for (parent = node->parent; parent; parent = parent->parent) { + if (parent->nodetype != LYS_LIST) { + continue; + } + + /* check and fix list uniques */ + slist = (struct lysc_node_list *)parent; + found = 0; + LY_ARRAY_FOR(slist->uniques, u) { + LY_ARRAY_FOR(slist->uniques[u], v) { + if (slist->uniques[u][v] == (struct lysc_node_leaf *)node) { + found = 1; + break; + } + } + + if (found) { + break; + } + } + + if (found) { + if (LY_ARRAY_COUNT(slist->uniques[u]) > 1) { + /* remove the item */ + if (v < LY_ARRAY_COUNT(slist->uniques[u]) - 1) { + memmove(&slist->uniques[u][v], &slist->uniques[u][v + 1], + (LY_ARRAY_COUNT(slist->uniques[u]) - v - 1) * sizeof slist->uniques[u][v]); + } + LY_ARRAY_DECREMENT(slist->uniques[u]); + } else { + /* remove the whole unique array */ + LY_ARRAY_FREE(slist->uniques[u]); + if (LY_ARRAY_COUNT(slist->uniques) > 1) { + if (u < LY_ARRAY_COUNT(slist->uniques) - 1) { + memmove(&slist->uniques[u], &slist->uniques[u + 1], + (LY_ARRAY_COUNT(slist->uniques) - u - 1) * sizeof slist->uniques[u]); + } + LY_ARRAY_DECREMENT(slist->uniques); + } else { + LY_ARRAY_FREE(slist->uniques); + slist->uniques = NULL; + } + } + } + } + + return LY_SUCCESS; +} + +/** + * @brief Finish dependency set compilation by resolving all the unres sets. + * + * @param[in] ctx libyang context. + * @param[in] unres Global unres structure with the sets to resolve. + * @return LY_SUCCESS on success. + * @return LY_ERECOMPILE if the dep set needs to be recompiled. + * @return LY_ERR value on error. + */ +static LY_ERR +lys_compile_unres_depset(struct ly_ctx *ctx, struct lys_glob_unres *unres) +{ + LY_ERR ret = LY_SUCCESS; + struct lysc_node *node; + struct lysc_type *typeiter; + struct lysc_type_leafref *lref; + struct lysc_ctx cctx = {0}; + struct lys_depset_unres *ds_unres = &unres->ds_unres; + struct ly_path *path; + LY_ARRAY_COUNT_TYPE v; + struct lysc_unres_leafref *l; + struct lysc_unres_when *w; + struct lysc_unres_must *m; + struct lysc_unres_dflt *d; + uint32_t i, processed_leafrefs = 0; + +resolve_all: + /* implement all referenced modules to get final ds_unres set */ + if ((ret = lys_compile_unres_depset_implement(ctx, unres))) { + goto cleanup; + } + + /* check disabled leafrefs */ + while (ds_unres->disabled_leafrefs.count) { + /* remember index, it can change before we get to free this item */ + i = ds_unres->disabled_leafrefs.count - 1; + l = ds_unres->disabled_leafrefs.objs[i]; + LYSC_CTX_INIT_PMOD(cctx, l->node->module->parsed, l->ext); + + LOG_LOCSET(l->node, NULL); + v = 0; + while ((ret == LY_SUCCESS) && (lref = lys_type_leafref_next(l->node, &v))) { + ret = lys_compile_unres_leafref(&cctx, l->node, lref, l->local_mod); + } + LOG_LOCBACK(1, 0); + LY_CHECK_GOTO(ret, cleanup); + + ly_set_rm_index(&ds_unres->disabled_leafrefs, i, free); + } + + /* for leafref, we need 2 rounds - first detects circular chain by storing the first referred type (which + * can be also leafref, in case it is already resolved, go through the chain and check that it does not + * point to the starting leafref type). The second round stores the first non-leafref type for later data validation. + * Also do the same check for set of the disabled leafrefs, but without the second round. */ + for (i = processed_leafrefs; i < ds_unres->leafrefs.count; ++i) { + l = ds_unres->leafrefs.objs[i]; + LYSC_CTX_INIT_PMOD(cctx, l->node->module->parsed, l->ext); + + LOG_LOCSET(l->node, NULL); + v = 0; + while ((ret == LY_SUCCESS) && (lref = lys_type_leafref_next(l->node, &v))) { + ret = lys_compile_unres_leafref(&cctx, l->node, lref, l->local_mod); + } + LOG_LOCBACK(1, 0); + LY_CHECK_GOTO(ret, cleanup); + } + for (i = processed_leafrefs; i < ds_unres->leafrefs.count; ++i) { + l = ds_unres->leafrefs.objs[i]; + + /* store pointer to the real type */ + v = 0; + while ((lref = lys_type_leafref_next(l->node, &v))) { + for (typeiter = lref->realtype; + typeiter->basetype == LY_TYPE_LEAFREF; + typeiter = ((struct lysc_type_leafref *)typeiter)->realtype) {} + + lysc_type_free(&cctx.free_ctx, lref->realtype); + lref->realtype = typeiter; + ++lref->realtype->refcount; + } + + /* if 'goto' will be used on the 'resolve_all' label, then the current leafref will not be processed again */ + processed_leafrefs++; + } + + /* check when, the referenced modules must be implemented now */ + while (ds_unres->whens.count) { + i = ds_unres->whens.count - 1; + w = ds_unres->whens.objs[i]; + LYSC_CTX_INIT_PMOD(cctx, w->node->module->parsed, NULL); + + LOG_LOCSET(w->node, NULL); + ret = lys_compile_unres_when(&cctx, w->when, w->node); + LOG_LOCBACK(1, 0); + LY_CHECK_GOTO(ret, cleanup); + + free(w); + ly_set_rm_index(&ds_unres->whens, i, NULL); + } + + /* check must */ + while (ds_unres->musts.count) { + i = ds_unres->musts.count - 1; + m = ds_unres->musts.objs[i]; + LYSC_CTX_INIT_PMOD(cctx, m->node->module->parsed, m->ext); + + LOG_LOCSET(m->node, NULL); + ret = lys_compile_unres_must(&cctx, m->node, m->local_mods); + LOG_LOCBACK(1, 0); + LY_CHECK_GOTO(ret, cleanup); + + lysc_unres_must_free(m); + ly_set_rm_index(&ds_unres->musts, i, NULL); + } + + /* remove disabled enums/bits */ + while (ds_unres->disabled_bitenums.count) { + i = ds_unres->disabled_bitenums.count - 1; + node = ds_unres->disabled_bitenums.objs[i]; + LYSC_CTX_INIT_PMOD(cctx, node->module->parsed, NULL); + + LOG_LOCSET(node, NULL); + ret = lys_compile_unres_disabled_bitenum(&cctx, (struct lysc_node_leaf *)node); + LOG_LOCBACK(1, 0); + LY_CHECK_GOTO(ret, cleanup); + + ly_set_rm_index(&ds_unres->disabled_bitenums, i, NULL); + } + + /* finish incomplete default values compilation */ + while (ds_unres->dflts.count) { + i = ds_unres->dflts.count - 1; + d = ds_unres->dflts.objs[i]; + LYSC_CTX_INIT_PMOD(cctx, d->leaf->module->parsed, NULL); + + LOG_LOCSET(&d->leaf->node, NULL); + if (d->leaf->nodetype == LYS_LEAF) { + ret = lys_compile_unres_leaf_dlft(&cctx, d->leaf, d->dflt, unres); + } else { + ret = lys_compile_unres_llist_dflts(&cctx, d->llist, d->dflt, d->dflts, unres); + } + LOG_LOCBACK(1, 0); + LY_CHECK_GOTO(ret, cleanup); + + lysc_unres_dflt_free(ctx, d); + ly_set_rm_index(&ds_unres->dflts, i, NULL); + } + + /* some unres items may have been added by the default values */ + if ((processed_leafrefs != ds_unres->leafrefs.count) || ds_unres->disabled_leafrefs.count || + ds_unres->whens.count || ds_unres->musts.count || ds_unres->dflts.count) { + goto resolve_all; + } + + /* finally, remove all disabled nodes */ + for (i = 0; i < ds_unres->disabled.count; ++i) { + node = ds_unres->disabled.snodes[i]; + ret = lys_compile_unres_check_disabled(node); + LY_CHECK_GOTO(ret, cleanup); + + LYSC_CTX_INIT_PMOD(cctx, node->module->parsed, NULL); + + lysc_node_free(&cctx.free_ctx, node, 1); + } + + /* also check if the leafref target has not been disabled */ + for (i = 0; i < ds_unres->leafrefs.count; ++i) { + l = ds_unres->leafrefs.objs[i]; + LYSC_CTX_INIT_PMOD(cctx, l->node->module->parsed, l->ext); + + v = 0; + while ((lref = lys_type_leafref_next(l->node, &v))) { + ret = ly_path_compile_leafref(cctx.ctx, l->node, cctx.ext, lref->path, + (l->node->flags & LYS_IS_OUTPUT) ? LY_PATH_OPER_OUTPUT : LY_PATH_OPER_INPUT, LY_PATH_TARGET_MANY, + LY_VALUE_SCHEMA_RESOLVED, lref->prefixes, &path); + ly_path_free(l->node->module->ctx, path); + + assert(ret != LY_ERECOMPILE); + if (ret) { + LOG_LOCSET(l->node, NULL); + LOGVAL(ctx, LYVE_REFERENCE, "Target of leafref \"%s\" cannot be referenced because it is disabled.", + l->node->name); + LOG_LOCBACK(1, 0); + ret = LY_EVALID; + goto cleanup; + } + } + } + +cleanup: + lysf_ctx_erase(&cctx.free_ctx); + return ret; +} + +/** + * @brief Erase dep set unres. + * + * @param[in] ctx libyang context. + * @param[in] unres Global unres structure with the sets to resolve. + */ +static void +lys_compile_unres_depset_erase(const struct ly_ctx *ctx, struct lys_glob_unres *unres) +{ + uint32_t i; + + ly_set_erase(&unres->ds_unres.whens, free); + for (i = 0; i < unres->ds_unres.musts.count; ++i) { + lysc_unres_must_free(unres->ds_unres.musts.objs[i]); + } + ly_set_erase(&unres->ds_unres.musts, NULL); + ly_set_erase(&unres->ds_unres.leafrefs, free); + for (i = 0; i < unres->ds_unres.dflts.count; ++i) { + lysc_unres_dflt_free(ctx, unres->ds_unres.dflts.objs[i]); + } + ly_set_erase(&unres->ds_unres.dflts, NULL); + ly_set_erase(&unres->ds_unres.disabled, NULL); + ly_set_erase(&unres->ds_unres.disabled_leafrefs, free); + ly_set_erase(&unres->ds_unres.disabled_bitenums, NULL); +} + +/** + * @brief Compile all flagged modules in a dependency set, recursively if recompilation is needed. + * + * @param[in] ctx libyang context. + * @param[in] dep_set Dependency set to compile. + * @param[in,out] unres Global unres to use. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_depset_r(struct ly_ctx *ctx, struct ly_set *dep_set, struct lys_glob_unres *unres) +{ + LY_ERR ret = LY_SUCCESS; + struct lysf_ctx fctx = {.ctx = ctx}; + struct lys_module *mod; + uint32_t i; + + for (i = 0; i < dep_set->count; ++i) { + mod = dep_set->objs[i]; + if (!mod->to_compile) { + /* skip */ + continue; + } + assert(mod->implemented); + + /* free the compiled module, if any */ + lysc_module_free(&fctx, mod->compiled); + mod->compiled = NULL; + + /* (re)compile the module */ + LY_CHECK_GOTO(ret = lys_compile(mod, &unres->ds_unres), cleanup); + } + + /* resolve dep set unres */ + ret = lys_compile_unres_depset(ctx, unres); + if (ret == LY_ERECOMPILE) { + /* new module is implemented, discard current dep set unres and recompile the whole dep set */ + lys_compile_unres_depset_erase(ctx, unres); + return lys_compile_depset_r(ctx, dep_set, unres); + } else if (ret) { + /* error */ + goto cleanup; + } + + /* success, unset the flags of all the modules in the dep set */ + for (i = 0; i < dep_set->count; ++i) { + mod = dep_set->objs[i]; + mod->to_compile = 0; + } + +cleanup: + lys_compile_unres_depset_erase(ctx, unres); + lysf_ctx_erase(&fctx); + return ret; +} + +/** + * @brief Check if-feature of all features of all modules in a dep set. + * + * @param[in] dep_set Dep set to check. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_depset_check_features(struct ly_set *dep_set) +{ + struct lys_module *mod; + uint32_t i; + + for (i = 0; i < dep_set->count; ++i) { + mod = dep_set->objs[i]; + if (!mod->to_compile) { + /* skip */ + continue; + } + + /* check features of this module */ + LY_CHECK_RET(lys_check_features(mod->parsed)); + } + + return LY_SUCCESS; +} + +LY_ERR +lys_compile_depset_all(struct ly_ctx *ctx, struct lys_glob_unres *unres) +{ + uint32_t i; + + for (i = 0; i < unres->dep_sets.count; ++i) { + LY_CHECK_RET(lys_compile_depset_check_features(unres->dep_sets.objs[i])); + LY_CHECK_RET(lys_compile_depset_r(ctx, unres->dep_sets.objs[i], unres)); + } + + return LY_SUCCESS; +} + +/** + * @brief Finish compilation of all the module unres sets in a compile context. + * + * @param[in] ctx Compile context with unres sets. + * @return LY_ERR value. + */ +static LY_ERR +lys_compile_unres_mod(struct lysc_ctx *ctx) +{ + struct lysc_augment *aug; + struct lysc_deviation *dev; + struct lys_module *orig_mod = ctx->cur_mod; + uint32_t i; + + /* check that all augments were applied */ + for (i = 0; i < ctx->augs.count; ++i) { + aug = ctx->augs.objs[i]; + ctx->cur_mod = aug->aug_pmod->mod; + if (aug->ext) { + lysc_update_path(ctx, NULL, "{extension}"); + lysc_update_path(ctx, NULL, aug->ext->name); + } + lysc_update_path(ctx, NULL, "{augment}"); + lysc_update_path(ctx, NULL, aug->nodeid->str); + LOGVAL(ctx->ctx, LYVE_REFERENCE, "Augment%s target node \"%s\" from module \"%s\" was not found.", + aug->ext ? " extension" : "", aug->nodeid->str, LYSP_MODULE_NAME(aug->aug_pmod)); + ctx->cur_mod = orig_mod; + lysc_update_path(ctx, NULL, NULL); + lysc_update_path(ctx, NULL, NULL); + if (aug->ext) { + lysc_update_path(ctx, NULL, NULL); + lysc_update_path(ctx, NULL, NULL); + } + } + if (ctx->augs.count) { + return LY_ENOTFOUND; + } + + /* check that all deviations were applied */ + for (i = 0; i < ctx->devs.count; ++i) { + dev = ctx->devs.objs[i]; + lysc_update_path(ctx, NULL, "{deviation}"); + lysc_update_path(ctx, NULL, dev->nodeid->str); + LOGVAL(ctx->ctx, LYVE_REFERENCE, "Deviation(s) target node \"%s\" from module \"%s\" was not found.", + dev->nodeid->str, LYSP_MODULE_NAME(dev->dev_pmods[0])); + lysc_update_path(ctx, NULL, NULL); + lysc_update_path(ctx, NULL, NULL); + } + if (ctx->devs.count) { + return LY_ENOTFOUND; + } + + return LY_SUCCESS; +} + +/** + * @brief Erase all the module unres sets in a compile context. + * + * @param[in] ctx Compile context with unres sets. + * @param[in] error Whether the compilation finished with an error or not. + */ +static void +lys_compile_unres_mod_erase(struct lysc_ctx *ctx, ly_bool error) +{ + uint32_t i; + + ly_set_erase(&ctx->groupings, NULL); + ly_set_erase(&ctx->tpdf_chain, NULL); + + if (!error) { + /* there can be no leftover deviations or augments */ + LY_CHECK_ERR_RET(ctx->augs.count, LOGINT(ctx->ctx), ); + LY_CHECK_ERR_RET(ctx->devs.count, LOGINT(ctx->ctx), ); + + ly_set_erase(&ctx->augs, NULL); + ly_set_erase(&ctx->devs, NULL); + ly_set_erase(&ctx->uses_augs, NULL); + ly_set_erase(&ctx->uses_rfns, NULL); + } else { + for (i = 0; i < ctx->augs.count; ++i) { + lysc_augment_free(ctx->ctx, ctx->augs.objs[i]); + } + ly_set_erase(&ctx->augs, NULL); + for (i = 0; i < ctx->devs.count; ++i) { + lysc_deviation_free(ctx->ctx, ctx->devs.objs[i]); + } + ly_set_erase(&ctx->devs, NULL); + for (i = 0; i < ctx->uses_augs.count; ++i) { + lysc_augment_free(ctx->ctx, ctx->uses_augs.objs[i]); + } + ly_set_erase(&ctx->uses_augs, NULL); + for (i = 0; i < ctx->uses_rfns.count; ++i) { + lysc_refine_free(ctx->ctx, ctx->uses_rfns.objs[i]); + } + ly_set_erase(&ctx->uses_rfns, NULL); + } +} + +LY_ERR +lys_compile(struct lys_module *mod, struct lys_depset_unres *unres) +{ + struct lysc_ctx ctx; + struct lysc_module *mod_c = NULL; + struct lysp_module *sp; + struct lysp_submodule *submod; + struct lysp_node *pnode; + struct lysp_node_grp *grp; + LY_ARRAY_COUNT_TYPE u; + LY_ERR ret = LY_SUCCESS; + + LY_CHECK_ARG_RET(NULL, mod, mod->parsed, !mod->compiled, mod->ctx, LY_EINVAL); + + assert(mod->implemented && mod->to_compile); + + sp = mod->parsed; + LYSC_CTX_INIT_PMOD(ctx, sp, NULL); + ctx.unres = unres; + + ++mod->ctx->change_count; + mod->compiled = mod_c = calloc(1, sizeof *mod_c); + LY_CHECK_ERR_RET(!mod_c, LOGMEM(mod->ctx), LY_EMEM); + mod_c->mod = mod; + + /* compile augments and deviations of our module from other modules so they can be applied during compilation */ + LY_CHECK_GOTO(ret = lys_precompile_own_augments(&ctx), cleanup); + LY_CHECK_GOTO(ret = lys_precompile_own_deviations(&ctx), cleanup); + + /* data nodes */ + LY_LIST_FOR(sp->data, pnode) { + LY_CHECK_GOTO(ret = lys_compile_node(&ctx, pnode, NULL, 0, NULL), cleanup); + } + + /* top-level RPCs */ + LY_LIST_FOR((struct lysp_node *)sp->rpcs, pnode) { + LY_CHECK_GOTO(ret = lys_compile_node(&ctx, pnode, NULL, 0, NULL), cleanup); + } + + /* top-level notifications */ + LY_LIST_FOR((struct lysp_node *)sp->notifs, pnode) { + LY_CHECK_GOTO(ret = lys_compile_node(&ctx, pnode, NULL, 0, NULL), cleanup); + } + + /* module extension instances */ + COMPILE_EXTS_GOTO(&ctx, sp->exts, mod_c->exts, mod_c, ret, cleanup); + + /* the same for submodules */ + LY_ARRAY_FOR(sp->includes, u) { + submod = sp->includes[u].submodule; + ctx.pmod = (struct lysp_module *)submod; + + LY_LIST_FOR(submod->data, pnode) { + ret = lys_compile_node(&ctx, pnode, NULL, 0, NULL); + LY_CHECK_GOTO(ret, cleanup); + } + + LY_LIST_FOR((struct lysp_node *)submod->rpcs, pnode) { + ret = lys_compile_node(&ctx, pnode, NULL, 0, NULL); + LY_CHECK_GOTO(ret, cleanup); + } + + LY_LIST_FOR((struct lysp_node *)submod->notifs, pnode) { + ret = lys_compile_node(&ctx, pnode, NULL, 0, NULL); + LY_CHECK_GOTO(ret, cleanup); + } + + COMPILE_EXTS_GOTO(&ctx, submod->exts, mod_c->exts, mod_c, ret, cleanup); + } + ctx.pmod = sp; + + /* validate non-instantiated groupings from the parsed schema, + * without it we would accept even the schemas with invalid grouping specification */ + ctx.compile_opts |= LYS_COMPILE_GROUPING; + LY_LIST_FOR(sp->groupings, grp) { + if (!(grp->flags & LYS_USED_GRP)) { + LY_CHECK_GOTO(ret = lys_compile_grouping(&ctx, NULL, grp), cleanup); + } + } + LY_LIST_FOR(sp->data, pnode) { + LY_LIST_FOR((struct lysp_node_grp *)lysp_node_groupings(pnode), grp) { + if (!(grp->flags & LYS_USED_GRP)) { + LY_CHECK_GOTO(ret = lys_compile_grouping(&ctx, pnode, grp), cleanup); + } + } + } + LY_ARRAY_FOR(sp->includes, u) { + submod = sp->includes[u].submodule; + ctx.pmod = (struct lysp_module *)submod; + + LY_LIST_FOR(submod->groupings, grp) { + if (!(grp->flags & LYS_USED_GRP)) { + LY_CHECK_GOTO(ret = lys_compile_grouping(&ctx, NULL, grp), cleanup); + } + } + LY_LIST_FOR(submod->data, pnode) { + LY_LIST_FOR((struct lysp_node_grp *)lysp_node_groupings(pnode), grp) { + if (!(grp->flags & LYS_USED_GRP)) { + LY_CHECK_GOTO(ret = lys_compile_grouping(&ctx, pnode, grp), cleanup); + } + } + } + } + ctx.pmod = sp; + + ly_log_location_revert(0, 0, 1, 0); + + /* finish compilation for all unresolved module items in the context */ + LY_CHECK_GOTO(ret = lys_compile_unres_mod(&ctx), cleanup); + +cleanup: + ly_log_location_revert(0, 0, 1, 0); + lys_compile_unres_mod_erase(&ctx, ret); + if (ret) { + lysc_module_free(&ctx.free_ctx, mod_c); + mod->compiled = NULL; + } + return ret; +} + +LY_ERR +lys_compile_identities(struct lys_module *mod) +{ + LY_ERR rc = LY_SUCCESS; + struct lysc_ctx ctx; + struct lysp_submodule *submod; + LY_ARRAY_COUNT_TYPE u; + + /* pre-compile identities of the module and any submodules */ + rc = lys_identity_precompile(NULL, mod->ctx, mod->parsed, mod->parsed->identities, &mod->identities); + LY_CHECK_GOTO(rc, cleanup); + LY_ARRAY_FOR(mod->parsed->includes, u) { + submod = mod->parsed->includes[u].submodule; + rc = lys_identity_precompile(NULL, mod->ctx, (struct lysp_module *)submod, submod->identities, &mod->identities); + LY_CHECK_GOTO(rc, cleanup); + } + + /* prepare context */ + LYSC_CTX_INIT_PMOD(ctx, mod->parsed, NULL); + + if (mod->parsed->identities) { + rc = lys_compile_identities_derived(&ctx, mod->parsed->identities, &mod->identities); + LY_CHECK_GOTO(rc, cleanup); + } + lysc_update_path(&ctx, NULL, "{submodule}"); + LY_ARRAY_FOR(mod->parsed->includes, u) { + submod = mod->parsed->includes[u].submodule; + if (submod->identities) { + ctx.pmod = (struct lysp_module *)submod; + lysc_update_path(&ctx, NULL, submod->name); + rc = lys_compile_identities_derived(&ctx, submod->identities, &mod->identities); + lysc_update_path(&ctx, NULL, NULL); + } + + if (rc) { + break; + } + } + lysc_update_path(&ctx, NULL, NULL); + +cleanup: + /* always needed when using lysc_update_path() */ + ly_log_location_revert(0, 0, 1, 0); + return rc; +} + +/** + * @brief Check whether a module does not have any (recursive) compiled import. + * + * @param[in] mod Module to examine. + * @return LY_SUCCESS on success. + * @return LY_ERECOMPILE on required recompilation of the dep set. + * @return LY_ERR on error. + */ +static LY_ERR +lys_has_compiled_import_r(struct lys_module *mod) +{ + LY_ARRAY_COUNT_TYPE u; + struct lys_module *m; + + LY_ARRAY_FOR(mod->parsed->imports, u) { + m = mod->parsed->imports[u].module; + if (!m->implemented) { + continue; + } + + if (!m->to_compile) { + /* module was not/will not be compiled in this compilation (so disabled nodes are not present) */ + m->to_compile = 1; + return LY_ERECOMPILE; + } + + /* recursive */ + LY_CHECK_RET(lys_has_compiled_import_r(m)); + } + + return LY_SUCCESS; +} + +LY_ERR +lys_implement(struct lys_module *mod, const char **features, struct lys_glob_unres *unres) +{ + LY_ERR r; + struct lys_module *m; + + assert(!mod->implemented); + + /* check collision with other implemented revision */ + m = ly_ctx_get_module_implemented(mod->ctx, mod->name); + if (m) { + assert(m != mod); + LOGERR(mod->ctx, LY_EDENIED, "Module \"%s@%s\" is already implemented in revision \"%s\".", + mod->name, mod->revision ? mod->revision : "<none>", m->revision ? m->revision : "<none>"); + return LY_EDENIED; + } + + /* set features */ + r = lys_set_features(mod->parsed, features); + if (r && (r != LY_EEXIST)) { + return r; + } + + /* + * mark the module implemented, which means + * 1) to (re)compile it only ::lys_compile() call is needed + * 2) its compilation will never cause new modules to be implemented (::lys_compile() does not return ::LY_ERECOMPILE) + * but there can be some unres items added that do + */ + mod->implemented = 1; + + /* this module is compiled in this compilation */ + mod->to_compile = 1; + + /* add the module into newly implemented module set */ + LY_CHECK_RET(ly_set_add(&unres->implementing, mod, 1, NULL)); + + /* mark target modules with our augments and deviations */ + LY_CHECK_RET(lys_precompile_augments_deviations(mod, unres)); + + /* check whether this module may reference any modules compiled previously */ + LY_CHECK_RET(lys_has_compiled_import_r(mod)); + + return LY_SUCCESS; +} |