summaryrefslogtreecommitdiffstats
path: root/backends/backends.c
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2022-04-14 18:12:14 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2022-04-14 18:12:14 +0000
commitbb50acdcb8073654ea667b8c0272e335bd43f844 (patch)
tree1e00c8a29871426f8182658928dcb62e42d57ce8 /backends/backends.c
parentReleasing debian version 1.33.1-1. (diff)
downloadnetdata-bb50acdcb8073654ea667b8c0272e335bd43f844.tar.xz
netdata-bb50acdcb8073654ea667b8c0272e335bd43f844.zip
Merging upstream version 1.34.0.
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'backends/backends.c')
-rw-r--r--backends/backends.c1247
1 files changed, 0 insertions, 1247 deletions
diff --git a/backends/backends.c b/backends/backends.c
deleted file mode 100644
index dca21ef1c..000000000
--- a/backends/backends.c
+++ /dev/null
@@ -1,1247 +0,0 @@
-// SPDX-License-Identifier: GPL-3.0-or-later
-
-#define BACKENDS_INTERNALS
-#include "backends.h"
-
-// ----------------------------------------------------------------------------
-// How backends work in netdata:
-//
-// 1. There is an independent thread that runs at the required interval
-// (for example, once every 10 seconds)
-//
-// 2. Every time it wakes, it calls the backend formatting functions to build
-// a buffer of data. This is a very fast, memory only operation.
-//
-// 3. If the buffer already includes data, the new data are appended.
-// If the buffer becomes too big, because the data cannot be sent, a
-// log is written and the buffer is discarded.
-//
-// 4. Then it tries to send all the data. It blocks until all the data are sent
-// or the socket returns an error.
-// If the time required for this is above the interval, it starts skipping
-// intervals, but the calculated values include the entire database, without
-// gaps (it remembers the timestamps and continues from where it stopped).
-//
-// 5. repeats the above forever.
-//
-
-const char *global_backend_prefix = "netdata";
-const char *global_backend_send_charts_matching = "*";
-int global_backend_update_every = 10;
-BACKEND_OPTIONS global_backend_options = BACKEND_SOURCE_DATA_AVERAGE | BACKEND_OPTION_SEND_NAMES;
-const char *global_backend_source = NULL;
-
-// ----------------------------------------------------------------------------
-// helper functions for backends
-
-size_t backend_name_copy(char *d, const char *s, size_t usable) {
- size_t n;
-
- for(n = 0; *s && n < usable ; d++, s++, n++) {
- char c = *s;
-
- if(c != '.' && !isalnum(c)) *d = '_';
- else *d = c;
- }
- *d = '\0';
-
- return n;
-}
-
-// calculate the SUM or AVERAGE of a dimension, for any timeframe
-// may return NAN if the database does not have any value in the give timeframe
-
-calculated_number backend_calculate_value_from_stored_data(
- RRDSET *st // the chart
- , RRDDIM *rd // the dimension
- , time_t after // the start timestamp
- , time_t before // the end timestamp
- , BACKEND_OPTIONS backend_options // BACKEND_SOURCE_* bitmap
- , time_t *first_timestamp // the first point of the database used in this response
- , time_t *last_timestamp // the timestamp that should be reported to backend
-) {
- RRDHOST *host = st->rrdhost;
- (void)host;
-
- // find the edges of the rrd database for this chart
- time_t first_t = rd->state->query_ops.oldest_time(rd);
- time_t last_t = rd->state->query_ops.latest_time(rd);
- time_t update_every = st->update_every;
- struct rrddim_query_handle handle;
- storage_number n;
-
- // step back a little, to make sure we have complete data collection
- // for all metrics
- after -= update_every * 2;
- before -= update_every * 2;
-
- // align the time-frame
- after = after - (after % update_every);
- before = before - (before % update_every);
-
- // for before, loose another iteration
- // the latest point will be reported the next time
- before -= update_every;
-
- if(unlikely(after > before))
- // this can happen when update_every > before - after
- after = before;
-
- if(unlikely(after < first_t))
- after = first_t;
-
- if(unlikely(before > last_t))
- before = last_t;
-
- if(unlikely(before < first_t || after > last_t)) {
- // the chart has not been updated in the wanted timeframe
- debug(D_BACKEND, "BACKEND: %s.%s.%s: aligned timeframe %lu to %lu is outside the chart's database range %lu to %lu",
- host->hostname, st->id, rd->id,
- (unsigned long)after, (unsigned long)before,
- (unsigned long)first_t, (unsigned long)last_t
- );
- return NAN;
- }
-
- *first_timestamp = after;
- *last_timestamp = before;
-
- size_t counter = 0;
- calculated_number sum = 0;
-
-/*
- long start_at_slot = rrdset_time2slot(st, before),
- stop_at_slot = rrdset_time2slot(st, after),
- slot, stop_now = 0;
-
- for(slot = start_at_slot; !stop_now ; slot--) {
-
- if(unlikely(slot < 0)) slot = st->entries - 1;
- if(unlikely(slot == stop_at_slot)) stop_now = 1;
-
- storage_number n = rd->values[slot];
-
- if(unlikely(!does_storage_number_exist(n))) {
- // not collected
- continue;
- }
-
- calculated_number value = unpack_storage_number(n);
- sum += value;
-
- counter++;
- }
-*/
- for(rd->state->query_ops.init(rd, &handle, after, before) ; !rd->state->query_ops.is_finished(&handle) ; ) {
- time_t curr_t;
- n = rd->state->query_ops.next_metric(&handle, &curr_t);
-
- if(unlikely(!does_storage_number_exist(n))) {
- // not collected
- continue;
- }
-
- calculated_number value = unpack_storage_number(n);
- sum += value;
-
- counter++;
- }
- rd->state->query_ops.finalize(&handle);
- if(unlikely(!counter)) {
- debug(D_BACKEND, "BACKEND: %s.%s.%s: no values stored in database for range %lu to %lu",
- host->hostname, st->id, rd->id,
- (unsigned long)after, (unsigned long)before
- );
- return NAN;
- }
-
- if(unlikely(BACKEND_OPTIONS_DATA_SOURCE(backend_options) == BACKEND_SOURCE_DATA_SUM))
- return sum;
-
- return sum / (calculated_number)counter;
-}
-
-
-// discard a response received by a backend
-// after logging a simple of it to error.log
-
-int discard_response(BUFFER *b, const char *backend) {
- char sample[1024];
- const char *s = buffer_tostring(b);
- char *d = sample, *e = &sample[sizeof(sample) - 1];
-
- for(; *s && d < e ;s++) {
- char c = *s;
- if(unlikely(!isprint(c))) c = ' ';
- *d++ = c;
- }
- *d = '\0';
-
- info("BACKEND: received %zu bytes from %s backend. Ignoring them. Sample: '%s'", buffer_strlen(b), backend, sample);
- buffer_flush(b);
- return 0;
-}
-
-
-// ----------------------------------------------------------------------------
-// the backend thread
-
-static SIMPLE_PATTERN *charts_pattern = NULL;
-static SIMPLE_PATTERN *hosts_pattern = NULL;
-
-inline int backends_can_send_rrdset(BACKEND_OPTIONS backend_options, RRDSET *st) {
- RRDHOST *host = st->rrdhost;
- (void)host;
-
- if(unlikely(rrdset_flag_check(st, RRDSET_FLAG_BACKEND_IGNORE)))
- return 0;
-
- if(unlikely(!rrdset_flag_check(st, RRDSET_FLAG_BACKEND_SEND))) {
- // we have not checked this chart
- if(simple_pattern_matches(charts_pattern, st->id) || simple_pattern_matches(charts_pattern, st->name))
- rrdset_flag_set(st, RRDSET_FLAG_BACKEND_SEND);
- else {
- rrdset_flag_set(st, RRDSET_FLAG_BACKEND_IGNORE);
- debug(D_BACKEND, "BACKEND: not sending chart '%s' of host '%s', because it is disabled for backends.", st->id, host->hostname);
- return 0;
- }
- }
-
- if(unlikely(!rrdset_is_available_for_exporting_and_alarms(st))) {
- debug(D_BACKEND, "BACKEND: not sending chart '%s' of host '%s', because it is not available for backends.", st->id, host->hostname);
- return 0;
- }
-
- if(unlikely(st->rrd_memory_mode == RRD_MEMORY_MODE_NONE && !(BACKEND_OPTIONS_DATA_SOURCE(backend_options) == BACKEND_SOURCE_DATA_AS_COLLECTED))) {
- debug(D_BACKEND, "BACKEND: not sending chart '%s' of host '%s' because its memory mode is '%s' and the backend requires database access.", st->id, host->hostname, rrd_memory_mode_name(host->rrd_memory_mode));
- return 0;
- }
-
- return 1;
-}
-
-inline BACKEND_OPTIONS backend_parse_data_source(const char *source, BACKEND_OPTIONS backend_options) {
- if(!strcmp(source, "raw") || !strcmp(source, "as collected") || !strcmp(source, "as-collected") || !strcmp(source, "as_collected") || !strcmp(source, "ascollected")) {
- backend_options |= BACKEND_SOURCE_DATA_AS_COLLECTED;
- backend_options &= ~(BACKEND_OPTIONS_SOURCE_BITS ^ BACKEND_SOURCE_DATA_AS_COLLECTED);
- }
- else if(!strcmp(source, "average")) {
- backend_options |= BACKEND_SOURCE_DATA_AVERAGE;
- backend_options &= ~(BACKEND_OPTIONS_SOURCE_BITS ^ BACKEND_SOURCE_DATA_AVERAGE);
- }
- else if(!strcmp(source, "sum") || !strcmp(source, "volume")) {
- backend_options |= BACKEND_SOURCE_DATA_SUM;
- backend_options &= ~(BACKEND_OPTIONS_SOURCE_BITS ^ BACKEND_SOURCE_DATA_SUM);
- }
- else {
- error("BACKEND: invalid data source method '%s'.", source);
- }
-
- return backend_options;
-}
-
-static void backends_main_cleanup(void *ptr) {
- struct netdata_static_thread *static_thread = (struct netdata_static_thread *)ptr;
- static_thread->enabled = NETDATA_MAIN_THREAD_EXITING;
-
- info("cleaning up...");
-
- static_thread->enabled = NETDATA_MAIN_THREAD_EXITED;
-}
-
-/**
- * Set Kinesis variables
- *
- * Set the variables necessary to work with this specific backend.
- *
- * @param default_port the default port of the backend
- * @param brc function called to check the result.
- * @param brf function called to format the message to the backend
- */
-void backend_set_kinesis_variables(int *default_port,
- backend_response_checker_t brc,
- backend_request_formatter_t brf)
-{
- (void)default_port;
-#ifndef HAVE_KINESIS
- (void)brc;
- (void)brf;
-#endif
-
-#if HAVE_KINESIS
- *brc = process_json_response;
- if (BACKEND_OPTIONS_DATA_SOURCE(global_backend_options) == BACKEND_SOURCE_DATA_AS_COLLECTED)
- *brf = backends_format_dimension_collected_json_plaintext;
- else
- *brf = backends_format_dimension_stored_json_plaintext;
-#endif
-}
-
-/**
- * Set Prometheus variables
- *
- * Set the variables necessary to work with this specific backend.
- *
- * @param default_port the default port of the backend
- * @param brc function called to check the result.
- * @param brf function called to format the message to the backend
- */
-void backend_set_prometheus_variables(int *default_port,
- backend_response_checker_t brc,
- backend_request_formatter_t brf)
-{
- (void)default_port;
- (void)brf;
-#ifndef ENABLE_PROMETHEUS_REMOTE_WRITE
- (void)brc;
-#endif
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- *brc = backends_process_prometheus_remote_write_response;
-#endif /* ENABLE_PROMETHEUS_REMOTE_WRITE */
-}
-
-/**
- * Set MongoDB variables
- *
- * Set the variables necessary to work with this specific backend.
- *
- * @param default_port the default port of the backend
- * @param brc function called to check the result.
- * @param brf function called to format the message to the backend
- */
-void backend_set_mongodb_variables(int *default_port,
- backend_response_checker_t brc,
- backend_request_formatter_t brf)
-{
- (void)default_port;
-#ifndef HAVE_MONGOC
- (void)brc;
- (void)brf;
-#endif
-
-#if HAVE_MONGOC
- *brc = process_json_response;
- if(BACKEND_OPTIONS_DATA_SOURCE(global_backend_options) == BACKEND_SOURCE_DATA_AS_COLLECTED)
- *brf = backends_format_dimension_collected_json_plaintext;
- else
- *brf = backends_format_dimension_stored_json_plaintext;
-#endif
-}
-
-/**
- * Set JSON variables
- *
- * Set the variables necessary to work with this specific backend.
- *
- * @param default_port the default port of the backend
- * @param brc function called to check the result.
- * @param brf function called to format the message to the backend
- */
-void backend_set_json_variables(int *default_port,
- backend_response_checker_t brc,
- backend_request_formatter_t brf)
-{
- *default_port = 5448;
- *brc = process_json_response;
-
- if (BACKEND_OPTIONS_DATA_SOURCE(global_backend_options) == BACKEND_SOURCE_DATA_AS_COLLECTED)
- *brf = backends_format_dimension_collected_json_plaintext;
- else
- *brf = backends_format_dimension_stored_json_plaintext;
-}
-
-/**
- * Set OpenTSDB HTTP variables
- *
- * Set the variables necessary to work with this specific backend.
- *
- * @param default_port the default port of the backend
- * @param brc function called to check the result.
- * @param brf function called to format the message to the backend
- */
-void backend_set_opentsdb_http_variables(int *default_port,
- backend_response_checker_t brc,
- backend_request_formatter_t brf)
-{
- *default_port = 4242;
- *brc = process_opentsdb_response;
-
- if(BACKEND_OPTIONS_DATA_SOURCE(global_backend_options) == BACKEND_SOURCE_DATA_AS_COLLECTED)
- *brf = backends_format_dimension_collected_opentsdb_http;
- else
- *brf = backends_format_dimension_stored_opentsdb_http;
-
-}
-
-/**
- * Set OpenTSDB Telnet variables
- *
- * Set the variables necessary to work with this specific backend.
- *
- * @param default_port the default port of the backend
- * @param brc function called to check the result.
- * @param brf function called to format the message to the backend
- */
-void backend_set_opentsdb_telnet_variables(int *default_port,
- backend_response_checker_t brc,
- backend_request_formatter_t brf)
-{
- *default_port = 4242;
- *brc = process_opentsdb_response;
-
- if(BACKEND_OPTIONS_DATA_SOURCE(global_backend_options) == BACKEND_SOURCE_DATA_AS_COLLECTED)
- *brf = backends_format_dimension_collected_opentsdb_telnet;
- else
- *brf = backends_format_dimension_stored_opentsdb_telnet;
-}
-
-/**
- * Set Graphite variables
- *
- * Set the variables necessary to work with this specific backend.
- *
- * @param default_port the default port of the backend
- * @param brc function called to check the result.
- * @param brf function called to format the message to the backend
- */
-void backend_set_graphite_variables(int *default_port,
- backend_response_checker_t brc,
- backend_request_formatter_t brf)
-{
- *default_port = 2003;
- *brc = process_graphite_response;
-
- if(BACKEND_OPTIONS_DATA_SOURCE(global_backend_options) == BACKEND_SOURCE_DATA_AS_COLLECTED)
- *brf = backends_format_dimension_collected_graphite_plaintext;
- else
- *brf = backends_format_dimension_stored_graphite_plaintext;
-}
-
-/**
- * Select Type
- *
- * Select the backend type based in the user input
- *
- * @param type is the string that defines the backend type
- *
- * @return It returns the backend id.
- */
-BACKEND_TYPE backend_select_type(const char *type) {
- if(!strcmp(type, "graphite") || !strcmp(type, "graphite:plaintext")) {
- return BACKEND_TYPE_GRAPHITE;
- }
- else if(!strcmp(type, "opentsdb") || !strcmp(type, "opentsdb:telnet")) {
- return BACKEND_TYPE_OPENTSDB_USING_TELNET;
- }
- else if(!strcmp(type, "opentsdb:http") || !strcmp(type, "opentsdb:https")) {
- return BACKEND_TYPE_OPENTSDB_USING_HTTP;
- }
- else if (!strcmp(type, "json") || !strcmp(type, "json:plaintext")) {
- return BACKEND_TYPE_JSON;
- }
- else if (!strcmp(type, "prometheus_remote_write")) {
- return BACKEND_TYPE_PROMETHEUS_REMOTE_WRITE;
- }
- else if (!strcmp(type, "kinesis") || !strcmp(type, "kinesis:plaintext")) {
- return BACKEND_TYPE_KINESIS;
- }
- else if (!strcmp(type, "mongodb") || !strcmp(type, "mongodb:plaintext")) {
- return BACKEND_TYPE_MONGODB;
- }
-
- return BACKEND_TYPE_UNKNOWN;
-}
-
-/**
- * Backend main
- *
- * The main thread used to control the backends.
- *
- * @param ptr a pointer to netdata_static_structure.
- *
- * @return It always return NULL.
- */
-void *backends_main(void *ptr) {
- netdata_thread_cleanup_push(backends_main_cleanup, ptr);
-
- int default_port = 0;
- int sock = -1;
- BUFFER *b = buffer_create(1), *response = buffer_create(1);
- int (*backend_request_formatter)(BUFFER *, const char *, RRDHOST *, const char *, RRDSET *, RRDDIM *, time_t, time_t, BACKEND_OPTIONS) = NULL;
- int (*backend_response_checker)(BUFFER *) = NULL;
-
-#if HAVE_KINESIS
- int do_kinesis = 0;
- char *kinesis_auth_key_id = NULL, *kinesis_secure_key = NULL, *kinesis_stream_name = NULL;
-#endif
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- int do_prometheus_remote_write = 0;
- BUFFER *http_request_header = NULL;
-#endif
-
-#if HAVE_MONGOC
- int do_mongodb = 0;
- char *mongodb_uri = NULL;
- char *mongodb_database = NULL;
- char *mongodb_collection = NULL;
-
- // set the default socket timeout in ms
- int32_t mongodb_default_socket_timeout = (int32_t)(global_backend_update_every >= 2)?(global_backend_update_every * MSEC_PER_SEC - 500):1000;
-
-#endif
-
-#ifdef ENABLE_HTTPS
- struct netdata_ssl opentsdb_ssl = {NULL , NETDATA_SSL_START};
-#endif
-
- // ------------------------------------------------------------------------
- // collect configuration options
-
- struct timeval timeout = {
- .tv_sec = 0,
- .tv_usec = 0
- };
- int enabled = config_get_boolean(CONFIG_SECTION_BACKEND, "enabled", 0);
- const char *source = config_get(CONFIG_SECTION_BACKEND, "data source", "average");
- const char *type = config_get(CONFIG_SECTION_BACKEND, "type", "graphite");
- const char *destination = config_get(CONFIG_SECTION_BACKEND, "destination", "localhost");
- global_backend_prefix = config_get(CONFIG_SECTION_BACKEND, "prefix", "netdata");
- const char *hostname = config_get(CONFIG_SECTION_BACKEND, "hostname", localhost->hostname);
- global_backend_update_every = (int)config_get_number(CONFIG_SECTION_BACKEND, "update every", global_backend_update_every);
- int buffer_on_failures = (int)config_get_number(CONFIG_SECTION_BACKEND, "buffer on failures", 10);
- long timeoutms = config_get_number(CONFIG_SECTION_BACKEND, "timeout ms", global_backend_update_every * 2 * 1000);
-
- if(config_get_boolean(CONFIG_SECTION_BACKEND, "send names instead of ids", (global_backend_options & BACKEND_OPTION_SEND_NAMES)))
- global_backend_options |= BACKEND_OPTION_SEND_NAMES;
- else
- global_backend_options &= ~BACKEND_OPTION_SEND_NAMES;
-
- charts_pattern = simple_pattern_create(
- global_backend_send_charts_matching = config_get(CONFIG_SECTION_BACKEND, "send charts matching", "*"),
- NULL,
- SIMPLE_PATTERN_EXACT);
- hosts_pattern = simple_pattern_create(config_get(CONFIG_SECTION_BACKEND, "send hosts matching", "localhost *"), NULL, SIMPLE_PATTERN_EXACT);
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- const char *remote_write_path = config_get(CONFIG_SECTION_BACKEND, "remote write URL path", "/receive");
-#endif
-
- // ------------------------------------------------------------------------
- // validate configuration options
- // and prepare for sending data to our backend
-
- global_backend_options = backend_parse_data_source(source, global_backend_options);
- global_backend_source = source;
-
- if(timeoutms < 1) {
- error("BACKEND: invalid timeout %ld ms given. Assuming %d ms.", timeoutms, global_backend_update_every * 2 * 1000);
- timeoutms = global_backend_update_every * 2 * 1000;
- }
- timeout.tv_sec = (timeoutms * 1000) / 1000000;
- timeout.tv_usec = (timeoutms * 1000) % 1000000;
-
- if(!enabled || global_backend_update_every < 1)
- goto cleanup;
-
- // ------------------------------------------------------------------------
- // select the backend type
- BACKEND_TYPE work_type = backend_select_type(type);
- if (work_type == BACKEND_TYPE_UNKNOWN) {
- error("BACKEND: Unknown backend type '%s'", type);
- goto cleanup;
- }
-
- switch (work_type) {
- case BACKEND_TYPE_OPENTSDB_USING_HTTP: {
-#ifdef ENABLE_HTTPS
- if (!strcmp(type, "opentsdb:https")) {
- security_start_ssl(NETDATA_SSL_CONTEXT_EXPORTING);
- }
-#endif
- backend_set_opentsdb_http_variables(&default_port,&backend_response_checker,&backend_request_formatter);
- break;
- }
- case BACKEND_TYPE_PROMETHEUS_REMOTE_WRITE: {
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- do_prometheus_remote_write = 1;
-
- http_request_header = buffer_create(1);
- backends_init_write_request();
-#else
- error("BACKEND: Prometheus remote write support isn't compiled");
-#endif // ENABLE_PROMETHEUS_REMOTE_WRITE
- backend_set_prometheus_variables(&default_port,&backend_response_checker,&backend_request_formatter);
- break;
- }
- case BACKEND_TYPE_KINESIS: {
-#if HAVE_KINESIS
- do_kinesis = 1;
-
- if(unlikely(read_kinesis_conf(netdata_configured_user_config_dir, &kinesis_auth_key_id, &kinesis_secure_key, &kinesis_stream_name))) {
- error("BACKEND: kinesis backend type is set but cannot read its configuration from %s/aws_kinesis.conf", netdata_configured_user_config_dir);
- goto cleanup;
- }
-
- backends_kinesis_init(destination, kinesis_auth_key_id, kinesis_secure_key, timeout.tv_sec * 1000 + timeout.tv_usec / 1000);
-#else
- error("BACKEND: AWS Kinesis support isn't compiled");
-#endif // HAVE_KINESIS
- backend_set_kinesis_variables(&default_port,&backend_response_checker,&backend_request_formatter);
- break;
- }
- case BACKEND_TYPE_MONGODB: {
-#if HAVE_MONGOC
- if(unlikely(read_mongodb_conf(netdata_configured_user_config_dir,
- &mongodb_uri,
- &mongodb_database,
- &mongodb_collection))) {
- error("BACKEND: mongodb backend type is set but cannot read its configuration from %s/mongodb.conf",
- netdata_configured_user_config_dir);
- goto cleanup;
- }
-
- if(likely(!backends_mongodb_init(mongodb_uri, mongodb_database, mongodb_collection, mongodb_default_socket_timeout))) {
- backend_set_mongodb_variables(&default_port, &backend_response_checker, &backend_request_formatter);
- do_mongodb = 1;
- }
- else {
- error("BACKEND: cannot initialize MongoDB backend");
- goto cleanup;
- }
-#else
- error("BACKEND: MongoDB support isn't compiled");
-#endif // HAVE_MONGOC
- break;
- }
- case BACKEND_TYPE_GRAPHITE: {
- backend_set_graphite_variables(&default_port,&backend_response_checker,&backend_request_formatter);
- break;
- }
- case BACKEND_TYPE_OPENTSDB_USING_TELNET: {
- backend_set_opentsdb_telnet_variables(&default_port,&backend_response_checker,&backend_request_formatter);
- break;
- }
- case BACKEND_TYPE_JSON: {
- backend_set_json_variables(&default_port,&backend_response_checker,&backend_request_formatter);
- break;
- }
- case BACKEND_TYPE_UNKNOWN: {
- break;
- }
- default: {
- break;
- }
- }
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- if((backend_request_formatter == NULL && !do_prometheus_remote_write) || backend_response_checker == NULL) {
-#else
- if(backend_request_formatter == NULL || backend_response_checker == NULL) {
-#endif
- error("BACKEND: backend is misconfigured - disabling it.");
- goto cleanup;
- }
-
-
-// ------------------------------------------------------------------------
-// prepare the charts for monitoring the backend operation
-
- struct rusage thread;
-
- collected_number
- chart_buffered_metrics = 0,
- chart_lost_metrics = 0,
- chart_sent_metrics = 0,
- chart_buffered_bytes = 0,
- chart_received_bytes = 0,
- chart_sent_bytes = 0,
- chart_receptions = 0,
- chart_transmission_successes = 0,
- chart_transmission_failures = 0,
- chart_data_lost_events = 0,
- chart_lost_bytes = 0,
- chart_backend_reconnects = 0;
- // chart_backend_latency = 0;
-
- RRDSET *chart_metrics = rrdset_create_localhost("netdata", "backend_metrics", NULL, "backend", NULL, "Netdata Buffered Metrics", "metrics", "backends", NULL, 130600, global_backend_update_every, RRDSET_TYPE_LINE);
- rrddim_add(chart_metrics, "buffered", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_metrics, "lost", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_metrics, "sent", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
-
- RRDSET *chart_bytes = rrdset_create_localhost("netdata", "backend_bytes", NULL, "backend", NULL, "Netdata Backend Data Size", "KiB", "backends", NULL, 130610, global_backend_update_every, RRDSET_TYPE_AREA);
- rrddim_add(chart_bytes, "buffered", NULL, 1, 1024, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_bytes, "lost", NULL, 1, 1024, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_bytes, "sent", NULL, 1, 1024, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_bytes, "received", NULL, 1, 1024, RRD_ALGORITHM_ABSOLUTE);
-
- RRDSET *chart_ops = rrdset_create_localhost("netdata", "backend_ops", NULL, "backend", NULL, "Netdata Backend Operations", "operations", "backends", NULL, 130630, global_backend_update_every, RRDSET_TYPE_LINE);
- rrddim_add(chart_ops, "write", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_ops, "discard", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_ops, "reconnect", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_ops, "failure", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
- rrddim_add(chart_ops, "read", NULL, 1, 1, RRD_ALGORITHM_ABSOLUTE);
-
- /*
- * this is misleading - we can only measure the time we need to send data
- * this time is not related to the time required for the data to travel to
- * the backend database and the time that server needed to process them
- *
- * issue #1432 and https://www.softlab.ntua.gr/facilities/documentation/unix/unix-socket-faq/unix-socket-faq-2.html
- *
- RRDSET *chart_latency = rrdset_create_localhost("netdata", "backend_latency", NULL, "backend", NULL, "Netdata Backend Latency", "ms", "backends", NULL, 130620, global_backend_update_every, RRDSET_TYPE_AREA);
- rrddim_add(chart_latency, "latency", NULL, 1, 1000, RRD_ALGORITHM_ABSOLUTE);
- */
-
- RRDSET *chart_rusage = rrdset_create_localhost("netdata", "backend_thread_cpu", NULL, "backend", NULL, "Netdata Backend Thread CPU usage", "milliseconds/s", "backends", NULL, 130630, global_backend_update_every, RRDSET_TYPE_STACKED);
- rrddim_add(chart_rusage, "user", NULL, 1, 1000, RRD_ALGORITHM_INCREMENTAL);
- rrddim_add(chart_rusage, "system", NULL, 1, 1000, RRD_ALGORITHM_INCREMENTAL);
-
-
- // ------------------------------------------------------------------------
- // prepare the backend main loop
-
- info("BACKEND: configured ('%s' on '%s' sending '%s' data, every %d seconds, as host '%s', with prefix '%s')", type, destination, source, global_backend_update_every, hostname, global_backend_prefix);
- send_statistics("BACKEND_START", "OK", type);
-
- usec_t step_ut = global_backend_update_every * USEC_PER_SEC;
- time_t after = now_realtime_sec();
- int failures = 0;
- heartbeat_t hb;
- heartbeat_init(&hb);
-
- while(!netdata_exit) {
-
- // ------------------------------------------------------------------------
- // Wait for the next iteration point.
-
- heartbeat_next(&hb, step_ut);
- time_t before = now_realtime_sec();
- debug(D_BACKEND, "BACKEND: preparing buffer for timeframe %lu to %lu", (unsigned long)after, (unsigned long)before);
-
- // ------------------------------------------------------------------------
- // add to the buffer the data we need to send to the backend
-
- netdata_thread_disable_cancelability();
-
- size_t count_hosts = 0;
- size_t count_charts_total = 0;
- size_t count_dims_total = 0;
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- if(do_prometheus_remote_write)
- backends_clear_write_request();
-#endif
- rrd_rdlock();
- RRDHOST *host;
- rrdhost_foreach_read(host) {
- if(unlikely(!rrdhost_flag_check(host, RRDHOST_FLAG_BACKEND_SEND|RRDHOST_FLAG_BACKEND_DONT_SEND))) {
- char *name = (host == localhost)?"localhost":host->hostname;
- if (!hosts_pattern || simple_pattern_matches(hosts_pattern, name)) {
- rrdhost_flag_set(host, RRDHOST_FLAG_BACKEND_SEND);
- info("enabled backend for host '%s'", name);
- }
- else {
- rrdhost_flag_set(host, RRDHOST_FLAG_BACKEND_DONT_SEND);
- info("disabled backend for host '%s'", name);
- }
- }
-
- if(unlikely(!rrdhost_flag_check(host, RRDHOST_FLAG_BACKEND_SEND)))
- continue;
-
- rrdhost_rdlock(host);
-
- count_hosts++;
- size_t count_charts = 0;
- size_t count_dims = 0;
- size_t count_dims_skipped = 0;
-
- const char *__hostname = (host == localhost)?hostname:host->hostname;
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- if(do_prometheus_remote_write) {
- backends_rrd_stats_remote_write_allmetrics_prometheus(
- host
- , __hostname
- , global_backend_prefix
- , global_backend_options
- , after
- , before
- , &count_charts
- , &count_dims
- , &count_dims_skipped
- );
- chart_buffered_metrics += count_dims;
- }
- else
-#endif
- {
- RRDSET *st;
- rrdset_foreach_read(st, host) {
- if(likely(backends_can_send_rrdset(global_backend_options, st))) {
- rrdset_rdlock(st);
-
- count_charts++;
-
- RRDDIM *rd;
- rrddim_foreach_read(rd, st) {
- if (likely(rd->last_collected_time.tv_sec >= after)) {
- chart_buffered_metrics += backend_request_formatter(b, global_backend_prefix, host, __hostname, st, rd, after, before, global_backend_options);
- count_dims++;
- }
- else {
- debug(D_BACKEND, "BACKEND: not sending dimension '%s' of chart '%s' from host '%s', its last data collection (%lu) is not within our timeframe (%lu to %lu)", rd->id, st->id, __hostname, (unsigned long)rd->last_collected_time.tv_sec, (unsigned long)after, (unsigned long)before);
- count_dims_skipped++;
- }
- }
-
- rrdset_unlock(st);
- }
- }
- }
-
- debug(D_BACKEND, "BACKEND: sending host '%s', metrics of %zu dimensions, of %zu charts. Skipped %zu dimensions.", __hostname, count_dims, count_charts, count_dims_skipped);
- count_charts_total += count_charts;
- count_dims_total += count_dims;
-
- rrdhost_unlock(host);
- }
- rrd_unlock();
-
- netdata_thread_enable_cancelability();
-
- debug(D_BACKEND, "BACKEND: buffer has %zu bytes, added metrics for %zu dimensions, of %zu charts, from %zu hosts", buffer_strlen(b), count_dims_total, count_charts_total, count_hosts);
-
- // ------------------------------------------------------------------------
-
- chart_buffered_bytes = (collected_number)buffer_strlen(b);
-
- // reset the monitoring chart counters
- chart_received_bytes =
- chart_sent_bytes =
- chart_sent_metrics =
- chart_lost_metrics =
- chart_receptions =
- chart_transmission_successes =
- chart_transmission_failures =
- chart_data_lost_events =
- chart_lost_bytes =
- chart_backend_reconnects = 0;
- // chart_backend_latency = 0;
-
- if(unlikely(netdata_exit)) break;
-
- //fprintf(stderr, "\nBACKEND BEGIN:\n%s\nBACKEND END\n", buffer_tostring(b));
- //fprintf(stderr, "after = %lu, before = %lu\n", after, before);
-
- // prepare for the next iteration
- // to add incrementally data to buffer
- after = before;
-
-#if HAVE_KINESIS
- if(do_kinesis) {
- unsigned long long partition_key_seq = 0;
-
- size_t buffer_len = buffer_strlen(b);
- size_t sent = 0;
-
- while(sent < buffer_len) {
- char partition_key[KINESIS_PARTITION_KEY_MAX + 1];
- snprintf(partition_key, KINESIS_PARTITION_KEY_MAX, "netdata_%llu", partition_key_seq++);
- size_t partition_key_len = strnlen(partition_key, KINESIS_PARTITION_KEY_MAX);
-
- const char *first_char = buffer_tostring(b) + sent;
-
- size_t record_len = 0;
-
- // split buffer into chunks of maximum allowed size
- if(buffer_len - sent < KINESIS_RECORD_MAX - partition_key_len) {
- record_len = buffer_len - sent;
- }
- else {
- record_len = KINESIS_RECORD_MAX - partition_key_len;
- while(*(first_char + record_len) != '\n' && record_len) record_len--;
- }
-
- char error_message[ERROR_LINE_MAX + 1] = "";
-
- debug(D_BACKEND, "BACKEND: backends_kinesis_put_record(): dest = %s, id = %s, key = %s, stream = %s, partition_key = %s, \
- buffer = %zu, record = %zu", destination, kinesis_auth_key_id, kinesis_secure_key, kinesis_stream_name,
- partition_key, buffer_len, record_len);
-
- backends_kinesis_put_record(kinesis_stream_name, partition_key, first_char, record_len);
-
- sent += record_len;
- chart_transmission_successes++;
-
- size_t sent_bytes = 0, lost_bytes = 0;
-
- if(unlikely(backends_kinesis_get_result(error_message, &sent_bytes, &lost_bytes))) {
- // oops! we couldn't send (all or some of the) data
- error("BACKEND: %s", error_message);
- error("BACKEND: failed to write data to database backend '%s'. Willing to write %zu bytes, wrote %zu bytes.",
- destination, sent_bytes, sent_bytes - lost_bytes);
-
- chart_transmission_failures++;
- chart_data_lost_events++;
- chart_lost_bytes += lost_bytes;
-
- // estimate the number of lost metrics
- chart_lost_metrics += (collected_number)(chart_buffered_metrics
- * (buffer_len && (lost_bytes > buffer_len) ? (double)lost_bytes / buffer_len : 1));
-
- break;
- }
- else {
- chart_receptions++;
- }
-
- if(unlikely(netdata_exit)) break;
- }
-
- chart_sent_bytes += sent;
- if(likely(sent == buffer_len))
- chart_sent_metrics = chart_buffered_metrics;
-
- buffer_flush(b);
- } else
-#endif /* HAVE_KINESIS */
-
-#if HAVE_MONGOC
- if(do_mongodb) {
- size_t buffer_len = buffer_strlen(b);
- size_t sent = 0;
-
- while(sent < buffer_len) {
- const char *first_char = buffer_tostring(b);
-
- debug(D_BACKEND, "BACKEND: backends_mongodb_insert(): uri = %s, database = %s, collection = %s, \
- buffer = %zu", mongodb_uri, mongodb_database, mongodb_collection, buffer_len);
-
- if(likely(!backends_mongodb_insert((char *)first_char, (size_t)chart_buffered_metrics))) {
- sent += buffer_len;
- chart_transmission_successes++;
- chart_receptions++;
- }
- else {
- // oops! we couldn't send (all or some of the) data
- error("BACKEND: failed to write data to database backend '%s'. Willing to write %zu bytes, wrote %zu bytes.",
- mongodb_uri, buffer_len, 0UL);
-
- chart_transmission_failures++;
- chart_data_lost_events++;
- chart_lost_bytes += buffer_len;
-
- // estimate the number of lost metrics
- chart_lost_metrics += (collected_number)chart_buffered_metrics;
-
- break;
- }
-
- if(unlikely(netdata_exit)) break;
- }
-
- chart_sent_bytes += sent;
- if(likely(sent == buffer_len))
- chart_sent_metrics = chart_buffered_metrics;
-
- buffer_flush(b);
- } else
-#endif /* HAVE_MONGOC */
-
- {
-
- // ------------------------------------------------------------------------
- // if we are connected, receive a response, without blocking
-
- if(likely(sock != -1)) {
- errno = 0;
-
- // loop through to collect all data
- while(sock != -1 && errno != EWOULDBLOCK) {
- buffer_need_bytes(response, 4096);
-
- ssize_t r;
-#ifdef ENABLE_HTTPS
- if(opentsdb_ssl.conn && !opentsdb_ssl.flags) {
- r = SSL_read(opentsdb_ssl.conn, &response->buffer[response->len], response->size - response->len);
- } else {
- r = recv(sock, &response->buffer[response->len], response->size - response->len, MSG_DONTWAIT);
- }
-#else
- r = recv(sock, &response->buffer[response->len], response->size - response->len, MSG_DONTWAIT);
-#endif
- if(likely(r > 0)) {
- // we received some data
- response->len += r;
- chart_received_bytes += r;
- chart_receptions++;
- }
- else if(r == 0) {
- error("BACKEND: '%s' closed the socket", destination);
- close(sock);
- sock = -1;
- }
- else {
- // failed to receive data
- if(errno != EAGAIN && errno != EWOULDBLOCK) {
- error("BACKEND: cannot receive data from backend '%s'.", destination);
- }
- }
- }
-
- // if we received data, process them
- if(buffer_strlen(response))
- backend_response_checker(response);
- }
-
- // ------------------------------------------------------------------------
- // if we are not connected, connect to a backend server
-
- if(unlikely(sock == -1)) {
- // usec_t start_ut = now_monotonic_usec();
- size_t reconnects = 0;
-
- sock = connect_to_one_of(destination, default_port, &timeout, &reconnects, NULL, 0);
-#ifdef ENABLE_HTTPS
- if(sock != -1) {
- if(netdata_exporting_ctx) {
- if(!opentsdb_ssl.conn) {
- opentsdb_ssl.conn = SSL_new(netdata_exporting_ctx);
- if(!opentsdb_ssl.conn) {
- error("Failed to allocate SSL structure %d.", sock);
- opentsdb_ssl.flags = NETDATA_SSL_NO_HANDSHAKE;
- }
- } else {
- SSL_clear(opentsdb_ssl.conn);
- }
- }
-
- if(opentsdb_ssl.conn) {
- if(SSL_set_fd(opentsdb_ssl.conn, sock) != 1) {
- error("Failed to set the socket to the SSL on socket fd %d.", host->rrdpush_sender_socket);
- opentsdb_ssl.flags = NETDATA_SSL_NO_HANDSHAKE;
- } else {
- opentsdb_ssl.flags = NETDATA_SSL_HANDSHAKE_COMPLETE;
- SSL_set_connect_state(opentsdb_ssl.conn);
- int err = SSL_connect(opentsdb_ssl.conn);
- if (err != 1) {
- err = SSL_get_error(opentsdb_ssl.conn, err);
- error("SSL cannot connect with the server: %s ", ERR_error_string((long)SSL_get_error(opentsdb_ssl.conn, err), NULL));
- opentsdb_ssl.flags = NETDATA_SSL_NO_HANDSHAKE;
- } //TODO: check certificate here
- }
- }
- }
-#endif
- chart_backend_reconnects += reconnects;
- // chart_backend_latency += now_monotonic_usec() - start_ut;
- }
-
- if(unlikely(netdata_exit)) break;
-
- // ------------------------------------------------------------------------
- // if we are connected, send our buffer to the backend server
-
- if(likely(sock != -1)) {
- size_t len = buffer_strlen(b);
- // usec_t start_ut = now_monotonic_usec();
- int flags = 0;
- #ifdef MSG_NOSIGNAL
- flags += MSG_NOSIGNAL;
- #endif
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- if(do_prometheus_remote_write) {
- size_t data_size = backends_get_write_request_size();
-
- if(unlikely(!data_size)) {
- error("BACKEND: write request size is out of range");
- continue;
- }
-
- buffer_flush(b);
- buffer_need_bytes(b, data_size);
- if(unlikely(backends_pack_write_request(b->buffer, &data_size))) {
- error("BACKEND: cannot pack write request");
- continue;
- }
- b->len = data_size;
- chart_buffered_bytes = (collected_number)buffer_strlen(b);
-
- buffer_flush(http_request_header);
- buffer_sprintf(http_request_header,
- "POST %s HTTP/1.1\r\n"
- "Host: %s\r\n"
- "Accept: */*\r\n"
- "Content-Length: %zu\r\n"
- "Content-Type: application/x-www-form-urlencoded\r\n\r\n",
- remote_write_path,
- destination,
- data_size
- );
-
- len = buffer_strlen(http_request_header);
- send(sock, buffer_tostring(http_request_header), len, flags);
-
- len = data_size;
- }
-#endif
-
- ssize_t written;
-#ifdef ENABLE_HTTPS
- if(opentsdb_ssl.conn && !opentsdb_ssl.flags) {
- written = SSL_write(opentsdb_ssl.conn, buffer_tostring(b), len);
- } else {
- written = send(sock, buffer_tostring(b), len, flags);
- }
-#else
- written = send(sock, buffer_tostring(b), len, flags);
-#endif
-
- // chart_backend_latency += now_monotonic_usec() - start_ut;
- if(written != -1 && (size_t)written == len) {
- // we sent the data successfully
- chart_transmission_successes++;
- chart_sent_bytes += written;
- chart_sent_metrics = chart_buffered_metrics;
-
- // reset the failures count
- failures = 0;
-
- // empty the buffer
- buffer_flush(b);
- }
- else {
- // oops! we couldn't send (all or some of the) data
- error("BACKEND: failed to write data to database backend '%s'. Willing to write %zu bytes, wrote %zd bytes. Will re-connect.", destination, len, written);
- chart_transmission_failures++;
-
- if(written != -1)
- chart_sent_bytes += written;
-
- // increment the counter we check for data loss
- failures++;
-
- // close the socket - we will re-open it next time
- close(sock);
- sock = -1;
- }
- }
- else {
- error("BACKEND: failed to update database backend '%s'", destination);
- chart_transmission_failures++;
-
- // increment the counter we check for data loss
- failures++;
- }
- }
-
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- if(do_prometheus_remote_write && failures) {
- (void) buffer_on_failures;
- failures = 0;
- chart_lost_bytes = chart_buffered_bytes = backends_get_write_request_size(); // estimated write request size
- chart_data_lost_events++;
- chart_lost_metrics = chart_buffered_metrics;
- } else
-#endif
- if(failures > buffer_on_failures) {
- // too bad! we are going to lose data
- chart_lost_bytes += buffer_strlen(b);
- error("BACKEND: reached %d backend failures. Flushing buffers to protect this host - this results in data loss on back-end server '%s'", failures, destination);
- buffer_flush(b);
- failures = 0;
- chart_data_lost_events++;
- chart_lost_metrics = chart_buffered_metrics;
- }
-
- if(unlikely(netdata_exit)) break;
-
- // ------------------------------------------------------------------------
- // update the monitoring charts
-
- if(likely(chart_ops->counter_done)) rrdset_next(chart_ops);
- rrddim_set(chart_ops, "read", chart_receptions);
- rrddim_set(chart_ops, "write", chart_transmission_successes);
- rrddim_set(chart_ops, "discard", chart_data_lost_events);
- rrddim_set(chart_ops, "failure", chart_transmission_failures);
- rrddim_set(chart_ops, "reconnect", chart_backend_reconnects);
- rrdset_done(chart_ops);
-
- if(likely(chart_metrics->counter_done)) rrdset_next(chart_metrics);
- rrddim_set(chart_metrics, "buffered", chart_buffered_metrics);
- rrddim_set(chart_metrics, "lost", chart_lost_metrics);
- rrddim_set(chart_metrics, "sent", chart_sent_metrics);
- rrdset_done(chart_metrics);
-
- if(likely(chart_bytes->counter_done)) rrdset_next(chart_bytes);
- rrddim_set(chart_bytes, "buffered", chart_buffered_bytes);
- rrddim_set(chart_bytes, "lost", chart_lost_bytes);
- rrddim_set(chart_bytes, "sent", chart_sent_bytes);
- rrddim_set(chart_bytes, "received", chart_received_bytes);
- rrdset_done(chart_bytes);
-
- /*
- if(likely(chart_latency->counter_done)) rrdset_next(chart_latency);
- rrddim_set(chart_latency, "latency", chart_backend_latency);
- rrdset_done(chart_latency);
- */
-
- getrusage(RUSAGE_THREAD, &thread);
- if(likely(chart_rusage->counter_done)) rrdset_next(chart_rusage);
- rrddim_set(chart_rusage, "user", thread.ru_utime.tv_sec * 1000000ULL + thread.ru_utime.tv_usec);
- rrddim_set(chart_rusage, "system", thread.ru_stime.tv_sec * 1000000ULL + thread.ru_stime.tv_usec);
- rrdset_done(chart_rusage);
-
- if(likely(buffer_strlen(b) == 0))
- chart_buffered_metrics = 0;
-
- if(unlikely(netdata_exit)) break;
- }
-
-cleanup:
-#if HAVE_KINESIS
- if(do_kinesis) {
- backends_kinesis_shutdown();
- freez(kinesis_auth_key_id);
- freez(kinesis_secure_key);
- freez(kinesis_stream_name);
- }
-#endif
-
-#if ENABLE_PROMETHEUS_REMOTE_WRITE
- buffer_free(http_request_header);
- if(do_prometheus_remote_write)
- backends_protocol_buffers_shutdown();
-#endif
-
-#if HAVE_MONGOC
- if(do_mongodb) {
- backends_mongodb_cleanup();
- freez(mongodb_uri);
- freez(mongodb_database);
- freez(mongodb_collection);
- }
-#endif
-
- if(sock != -1)
- close(sock);
-
- buffer_free(b);
- buffer_free(response);
-
-#ifdef ENABLE_HTTPS
- if(netdata_exporting_ctx) {
- if(opentsdb_ssl.conn) {
- SSL_free(opentsdb_ssl.conn);
- }
- }
-#endif
-
- netdata_thread_cleanup_pop(1);
- return NULL;
-}